davidoj01 commited on
Commit
6c1711e
·
verified ·
1 Parent(s): 99b8f09

Upload MistralForCausalLM

Browse files
adapter_config.json CHANGED
@@ -22,12 +22,12 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
 
 
 
25
  "k_proj",
26
  "up_proj",
27
- "v_proj",
28
- "gate_proj",
29
- "q_proj",
30
- "down_proj"
31
  ],
32
  "task_type": null
33
  }
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "q_proj",
26
+ "v_proj",
27
+ "down_proj",
28
  "k_proj",
29
  "up_proj",
30
+ "gate_proj"
 
 
 
31
  ],
32
  "task_type": null
33
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5aa708771c402fc1220ad0679cd2cb1fd6c75005f163d97d10ef333cdf1f7a7c
3
  size 75548136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64b5232c66cb719c846d433e79b20c26ded48f309391f0ae20a1fa99446ddd66
3
  size 75548136
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.37.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.40.2"
6
  }