Mistral-lb-nllb-200 / config.json
KoseiUemura's picture
Upload entire model folder
ad02c70 verified
raw
history blame contribute delete
324 Bytes
{
"alignments": "linear",
"architectures": [
"LangBridgeModel"
],
"dim_enc": 1024,
"dim_lm": 4096,
"enc": "facebook/nllb-200-distilled-600M",
"freeze_encoder": true,
"freeze_language_model": true,
"lm": "meta-math/MetaMath-Mistral-7B",
"torch_dtype": "bfloat16",
"transformers_version": "4.37.2"
}