prefix32 / llama23-1b-prefix-is /head_config.json
jekunz's picture
Model save
f4e244e verified
raw
history blame
279 Bytes
{
"config": null,
"hidden_size": 2048,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"model_class": "LlamaForCausalLM",
"model_name": "meta-llama/Llama-3.2-1B-Instruct",
"model_type": "llama",
"name": null,
"num_labels": 2,
"version": "adapters.1.0.0"
}