daekeun-ml commited on
Commit
6accd21
·
verified ·
1 Parent(s): 57010c0

Upload Phi4MMForCausalLM

Browse files

Updated weights (4 epochs)

README.md CHANGED
@@ -8,45 +8,45 @@ metrics:
8
  - cer
9
  base_model:
10
  - microsoft/Phi-4-multimodal-instruct
 
 
 
 
 
 
 
 
 
11
  model-index:
12
  - name: Phi-4-mm-inst-zeroth-kor
13
  results:
14
  - task:
15
  type: speech-to-text-translation
16
  dataset:
17
- type: seastar105/fleurs_ko_en_test
18
  name: fleurs (ko-en test intersection)
 
19
  metrics:
20
  - type: bleu
21
- name: ko2en
22
  value: 7.03
 
23
  - type: bleu
24
- name: ko2en-cot
25
  value: 7.04
 
26
  - type: bleu
 
27
  name: en2ko (ko-mecab)
28
- value: 12.50
29
  - type: bleu
30
- name: en2ko-cot (ko-mecab)
31
  value: 9.54
 
32
  - task:
33
  type: automatic-speech-recognition
34
  dataset:
35
- type: kresnik/zeroth_korean
36
  name: zeroth_korean test
 
37
  metrics:
38
  - type: cer
39
- name: test CER
40
  value: 7.02
41
- language:
42
- - ko
43
- license: mit
44
- tags:
45
- - korean
46
- - stt
47
- - custom_code
48
- - phi
49
- - phi-4-multimodal
50
  ---
51
 
52
  # Phi-4-multimodal-finetune-ko-speech
 
8
  - cer
9
  base_model:
10
  - microsoft/Phi-4-multimodal-instruct
11
+ language:
12
+ - ko
13
+ license: mit
14
+ tags:
15
+ - korean
16
+ - stt
17
+ - custom_code
18
+ - phi
19
+ - phi-4-multimodal
20
  model-index:
21
  - name: Phi-4-mm-inst-zeroth-kor
22
  results:
23
  - task:
24
  type: speech-to-text-translation
25
  dataset:
 
26
  name: fleurs (ko-en test intersection)
27
+ type: seastar105/fleurs_ko_en_test
28
  metrics:
29
  - type: bleu
 
30
  value: 7.03
31
+ name: ko2en
32
  - type: bleu
 
33
  value: 7.04
34
+ name: ko2en-cot
35
  - type: bleu
36
+ value: 12.5
37
  name: en2ko (ko-mecab)
 
38
  - type: bleu
 
39
  value: 9.54
40
+ name: en2ko-cot (ko-mecab)
41
  - task:
42
  type: automatic-speech-recognition
43
  dataset:
 
44
  name: zeroth_korean test
45
+ type: kresnik/zeroth_korean
46
  metrics:
47
  - type: cer
 
48
  value: 7.02
49
+ name: test CER
 
 
 
 
 
 
 
 
50
  ---
51
 
52
  # Phi-4-multimodal-finetune-ko-speech
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./output",
3
  "architectures": [
4
  "Phi4MMForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "./Phi-4-multimodal-finetune-ko-speech",
3
  "architectures": [
4
  "Phi4MMForCausalLM"
5
  ],
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5dd159a69ab84ceb84ae8495bc2c5b186bbdd666786e6fa95d8f3860119f3484
3
- size 4998420448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a1e57956624d5a96c5dcaab437261353efbcfaf5eff49d58d92329311020802
3
+ size 4997504848
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:402ce19eefd8c071927dd399bf6a84ce7a1ed1c400ed96cfe0f5b4e7872050f4
3
- size 4983891952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7ae17c45b5ab41c7659784e8cb2d6a79e6f886783b3c2af0297448a1f240c1c
3
+ size 4952333128
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dc9ce4edc2596f5d9439b5762f9b31b7e19342ea1305769388907d7bc015fc8
3
- size 1905111704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7277ccd89f1b88a99dd7ce37095510f2036e18e6289b732164704ee0d2934b92
3
+ size 1199389232
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 11887118272
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
@@ -1741,50 +1741,50 @@
1741
  "model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1742
  "model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1743
  "model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1744
- "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
1745
- "model.layers.25.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1746
- "model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1747
- "model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1748
- "model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1749
- "model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1750
- "model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1751
- "model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1752
- "model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1753
- "model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1754
- "model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1755
- "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
1756
  "model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1757
  "model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1758
  "model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1759
  "model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1760
  "model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1761
- "model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1762
- "model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1763
- "model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1764
- "model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1765
- "model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1766
- "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
1767
- "model.layers.26.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1768
- "model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1769
- "model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1770
- "model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1771
- "model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1772
- "model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1773
- "model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1774
- "model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1775
- "model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1776
- "model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1777
- "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
1778
- "model.layers.26.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1779
- "model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1780
- "model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1781
- "model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1782
- "model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1783
- "model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1784
- "model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1785
- "model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1786
- "model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1787
- "model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1788
  "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
1789
  "model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1790
  "model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
@@ -1797,16 +1797,16 @@
1797
  "model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1798
  "model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1799
  "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
1800
- "model.layers.27.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1801
- "model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1802
- "model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1803
- "model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1804
- "model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1805
- "model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1806
- "model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
1807
- "model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
1808
- "model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1809
- "model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1810
  "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
1811
  "model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1812
  "model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
@@ -1989,22 +1989,22 @@
1989
  "model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1990
  "model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1991
  "model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1992
- "model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1993
- "model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1994
- "model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1995
  "model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1996
- "model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1997
  "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
1998
  "model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
1999
- "model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
2000
- "model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
2001
- "model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
2002
- "model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
2003
- "model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
2004
- "model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
2005
- "model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
2006
- "model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
2007
- "model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
2008
  "model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
2009
  "model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
2010
  "model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 11148920768
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
 
1741
  "model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1742
  "model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1743
  "model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1744
+ "model.layers.25.input_layernorm.weight": "model-00002-of-00003.safetensors",
1745
+ "model.layers.25.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1746
+ "model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1747
+ "model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1748
+ "model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1749
+ "model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1750
+ "model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1751
+ "model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1752
+ "model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1753
+ "model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1754
+ "model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1755
+ "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
1756
  "model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1757
  "model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1758
  "model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1759
  "model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1760
  "model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1761
+ "model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1762
+ "model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1763
+ "model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1764
+ "model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1765
+ "model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1766
+ "model.layers.26.input_layernorm.weight": "model-00002-of-00003.safetensors",
1767
+ "model.layers.26.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1768
+ "model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1769
+ "model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1770
+ "model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1771
+ "model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1772
+ "model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1773
+ "model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1774
+ "model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1775
+ "model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1776
+ "model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1777
+ "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
1778
+ "model.layers.26.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1779
+ "model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1780
+ "model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1781
+ "model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1782
+ "model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1783
+ "model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1784
+ "model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1785
+ "model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1786
+ "model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1787
+ "model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1788
  "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
1789
  "model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1790
  "model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
 
1797
  "model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
1798
  "model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
1799
  "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
1800
+ "model.layers.27.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1801
+ "model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1802
+ "model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1803
+ "model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1804
+ "model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1805
+ "model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
1806
+ "model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
1807
+ "model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1808
+ "model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1809
+ "model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1810
  "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
1811
  "model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
1812
  "model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
 
1989
  "model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
1990
  "model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1991
  "model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
1992
+ "model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-00001-of-00003.safetensors",
1993
+ "model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
1994
+ "model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
1995
  "model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
1996
+ "model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
1997
  "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
1998
  "model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
1999
+ "model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
2000
+ "model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
2001
+ "model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
2002
+ "model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
2003
+ "model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-00001-of-00003.safetensors",
2004
+ "model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
2005
+ "model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
2006
+ "model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
2007
+ "model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
2008
  "model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
2009
  "model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
2010
  "model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",