OmbeniFaraja commited on
Commit
f5ef85c
·
verified ·
1 Parent(s): 5788dd0

Trained with Unsloth

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "unsloth/phi-3-mini-4k-instruct-bnb-4bit",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -23,7 +23,7 @@
23
  "sliding_window": 2048,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "float16",
26
- "transformers_version": "4.44.0",
27
  "unsloth_version": "2024.8",
28
  "use_cache": true,
29
  "vocab_size": 32064
 
1
  {
2
+ "_name_or_path": "OmbeniFaraja/ph3-finetune-kra-etims",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
23
  "sliding_window": 2048,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "float16",
26
+ "transformers_version": "4.44.2",
27
  "unsloth_version": "2024.8",
28
  "use_cache": true,
29
  "vocab_size": 32064
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  ],
9
  "max_length": 4096,
10
  "pad_token_id": 32009,
11
- "transformers_version": "4.44.0"
12
  }
 
8
  ],
9
  "max_length": 4096,
10
  "pad_token_id": 32009,
11
+ "transformers_version": "4.44.2"
12
  }
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11587a1215d1a5e9d2804b2c55d4a1ace8ccd1390c221e399bf96d91dc423625
3
  size 4991417114
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93a0148a79e1cad1eddf1629119f1537b5a71d4b92e4e9414ba14910a8ac4361
3
  size 4991417114
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:336106319438e2eed0cbc7e67b5c076b9f1adceceb11a7ab30c374c50fee9562
3
  size 2650845472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20ea783555a34eebaea9242aed21d53c172f971fac743e2e4b111a4f207b10e9
3
  size 2650845472