{ "apply_qk_norm": true, "architectures": [ "OpenLMForCausalLM" ], "dim": 224, "model_type": "openlm", "n_heads": 4, "n_layers": 6, "post_embed_norm": false, "seq_len": 2048, "torch_dtype": "float32", "transformers_version": "4.49.0", "vocab_size": 50432, "weight_tying": false }