gemma-2-9b-it-MNN / llm_config.json
zhaode's picture
Upload folder using huggingface_hub
f0e1182 verified
raw
history blame contribute delete
284 Bytes
{
"hidden_size": 3584,
"layer_nums": 42,
"attention_mask": "float",
"key_value_shape": [
2,
1,
0,
8,
256
],
"prompt_template": "<bos><start_of_turn>user\n%s<end_of_turn>\n<start_of_turn>model\n",
"is_visual": false
}