Sugyeong commited on
Commit
340fb52
·
verified ·
1 Parent(s): d21bc5d

Upload MistralForCausalLM

Browse files
config.json CHANGED
@@ -8,7 +8,7 @@
8
  "attention_dropout": 0.0,
9
  "auto_map": {
10
  "AutoConfig": "configuration_sparsetral.SparsetralConfig",
11
- "AutoModelForCausalLM": "modeling_sparsetral_sentence_level.MistralForCausalLM"
12
  },
13
  "bos_token_id": 1,
14
  "eos_token_id": 2,
 
8
  "attention_dropout": 0.0,
9
  "auto_map": {
10
  "AutoConfig": "configuration_sparsetral.SparsetralConfig",
11
+ "AutoModelForCausalLM": "modeling_sparsetral_sent_level.MistralForCausalLM"
12
  },
13
  "bos_token_id": 1,
14
  "eos_token_id": 2,
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:896843787b602f5ade378718a2d1428230c49f9d12bc2281f2a7ddbacbe54604
3
  size 4899142136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a63baa0ae4a2cc929f7a089f396c9ca20f591c4d790ca143bb62220cedfa5ec
3
  size 4899142136
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f404bd6f88b9ab4b32a41730ab469fcb3627d2578d2d56a005fb04358c0cd62f
3
  size 4996678776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93560b38f4b1797e481ce0bfd0553dca5e6f12172e891dc94a2b191b5ce4a53c
3
  size 4996678776
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c81d6b0507fa01857bce5f0b8a12a877072967a8a7748e2861c486773a07e3e
3
  size 4822622760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01897130855283dc8b73bdd436f2b48462dc70f97907da922754e10644e56cf7
3
  size 4822622760