{
  "architectures": [
    "OpenLMModel"
  ],
  "hidden_dim": 2048,
  "model_type": "openlm",
  "n_heads": 16,
  "n_layers": 24,
  "post_embed_norm": false,
  "seq_len": 2048,
  "torch_dtype": "float32",
  "transformers_version": "4.30.2",
  "vocab_size": 50432,
  "weight_tying": false
}
