Sarim-Hash commited on
Commit
b985af1
·
verified ·
1 Parent(s): dfd0908

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitignore +1 -0
  2. only_ce_really_fixed/adapter_1000/README.md +9 -0
  3. only_ce_really_fixed/adapter_1000/adapter_config.json +24 -0
  4. only_ce_really_fixed/adapter_1000/adapter_model.bin +3 -0
  5. only_ce_really_fixed/adapter_1500/README.md +9 -0
  6. only_ce_really_fixed/adapter_1500/adapter_config.json +24 -0
  7. only_ce_really_fixed/adapter_1500/adapter_model.bin +3 -0
  8. only_ce_really_fixed/adapter_2000/README.md +9 -0
  9. only_ce_really_fixed/adapter_2000/adapter_config.json +24 -0
  10. only_ce_really_fixed/adapter_2000/adapter_model.bin +3 -0
  11. only_ce_really_fixed/adapter_2500/README.md +9 -0
  12. only_ce_really_fixed/adapter_2500/adapter_config.json +24 -0
  13. only_ce_really_fixed/adapter_2500/adapter_model.bin +3 -0
  14. only_ce_really_fixed/adapter_3000/README.md +9 -0
  15. only_ce_really_fixed/adapter_3000/adapter_config.json +24 -0
  16. only_ce_really_fixed/adapter_3000/adapter_model.bin +3 -0
  17. only_ce_really_fixed/adapter_3500/README.md +9 -0
  18. only_ce_really_fixed/adapter_3500/adapter_config.json +24 -0
  19. only_ce_really_fixed/adapter_3500/adapter_model.bin +3 -0
  20. only_ce_really_fixed/adapter_4000/README.md +9 -0
  21. only_ce_really_fixed/adapter_4000/adapter_config.json +24 -0
  22. only_ce_really_fixed/adapter_4000/adapter_model.bin +3 -0
  23. only_ce_really_fixed/adapter_4500/README.md +9 -0
  24. only_ce_really_fixed/adapter_4500/adapter_config.json +24 -0
  25. only_ce_really_fixed/adapter_4500/adapter_model.bin +3 -0
  26. only_ce_really_fixed/adapter_500/README.md +9 -0
  27. only_ce_really_fixed/adapter_500/adapter_config.json +24 -0
  28. only_ce_really_fixed/adapter_500/adapter_model.bin +3 -0
  29. only_ce_really_fixed/adapter_5000/README.md +9 -0
  30. only_ce_really_fixed/adapter_5000/adapter_config.json +24 -0
  31. only_ce_really_fixed/adapter_5000/adapter_model.bin +3 -0
  32. only_ce_really_fixed/adapter_5500/README.md +9 -0
  33. only_ce_really_fixed/adapter_5500/adapter_config.json +24 -0
  34. only_ce_really_fixed/adapter_5500/adapter_model.bin +3 -0
  35. only_posix_100_really_fixed/adapter_1000/README.md +9 -0
  36. only_posix_100_really_fixed/adapter_1000/adapter_config.json +24 -0
  37. only_posix_100_really_fixed/adapter_1000/adapter_model.bin +3 -0
  38. only_posix_100_really_fixed/adapter_1500/README.md +9 -0
  39. only_posix_100_really_fixed/adapter_1500/adapter_config.json +24 -0
  40. only_posix_100_really_fixed/adapter_1500/adapter_model.bin +3 -0
  41. only_posix_100_really_fixed/adapter_2000/README.md +9 -0
  42. only_posix_100_really_fixed/adapter_2000/adapter_config.json +24 -0
  43. only_posix_100_really_fixed/adapter_2000/adapter_model.bin +3 -0
  44. only_posix_100_really_fixed/adapter_2500/README.md +9 -0
  45. only_posix_100_really_fixed/adapter_2500/adapter_config.json +24 -0
  46. only_posix_100_really_fixed/adapter_2500/adapter_model.bin +3 -0
  47. only_posix_100_really_fixed/adapter_3000/README.md +9 -0
  48. only_posix_100_really_fixed/adapter_3000/adapter_config.json +24 -0
  49. only_posix_100_really_fixed/adapter_3000/adapter_model.bin +3 -0
  50. only_posix_100_really_fixed/adapter_3500/README.md +9 -0
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ push.sh
only_ce_really_fixed/adapter_1000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_1000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_1000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3747633a6cc0c7403ef0712b9331822d258bfa9afe681f0981de565d5d6c564d
3
+ size 33640010
only_ce_really_fixed/adapter_1500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_1500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_1500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69a68e7fa2821bce5701d878508e842ad3b3339a81a9edb6ced52699f359e8c4
3
+ size 33640010
only_ce_really_fixed/adapter_2000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_2000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_2000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:600296c4f382bdb5fa06e9bc76e861e2e056427efc87522c73f3c73e462357c9
3
+ size 33640010
only_ce_really_fixed/adapter_2500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_2500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_2500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:571a77b3cd158060156062660f9eded8f69ae41679aca027cc996d4bd5a3f17a
3
+ size 33640010
only_ce_really_fixed/adapter_3000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_3000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_3000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18ffa70aa8bfbec20bc5e89d36d983444c600e56c4187aae6cbfd63a9a6c24a5
3
+ size 33640010
only_ce_really_fixed/adapter_3500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_3500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_3500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:050c37a454c05f3b766d63c9e6ce8d2e939649bb133237b685837feef047abe6
3
+ size 33640010
only_ce_really_fixed/adapter_4000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_4000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_4000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba50aaa20e07019c165adb3d7be91a3f8c42008f671370a14877c9f00ac993d5
3
+ size 33640010
only_ce_really_fixed/adapter_4500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_4500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_4500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:527a46e800550057014fa749ed854525a02987e0f6f55c7a6dc536b03dacad08
3
+ size 33640010
only_ce_really_fixed/adapter_500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:018a7a0ee929616b3f0751145f3353f7199e7329799c68bae10d946a588c1f11
3
+ size 33640010
only_ce_really_fixed/adapter_5000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_5000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_5000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa6b6f8ca9516cf3db8925ab2b637095977345d00ef99d7663e704081959a633
3
+ size 33640010
only_ce_really_fixed/adapter_5500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_ce_really_fixed/adapter_5500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_ce_really_fixed/adapter_5500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:433e92798f22b7df83fe7d1a52b6c2973e8d75476bcebd66f07894bdfad48acf
3
+ size 33640010
only_posix_100_really_fixed/adapter_1000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_posix_100_really_fixed/adapter_1000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_posix_100_really_fixed/adapter_1000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22ebf9d1f2eb99d9e34875e7f5fa6e0ad8dc6a015c282642d28a96b489ea5657
3
+ size 33640010
only_posix_100_really_fixed/adapter_1500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_posix_100_really_fixed/adapter_1500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_posix_100_really_fixed/adapter_1500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:995d8237ac7f6d30a79b2792d8e88ce82e9cd10ae7da2b09f4b77ba27719b8c5
3
+ size 33640010
only_posix_100_really_fixed/adapter_2000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_posix_100_really_fixed/adapter_2000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_posix_100_really_fixed/adapter_2000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b80ac7cbe1906c82508af649c67cc48cba2e4f3971ae40f30cfd4e9932db4f6
3
+ size 33640010
only_posix_100_really_fixed/adapter_2500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_posix_100_really_fixed/adapter_2500/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_posix_100_really_fixed/adapter_2500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e3ee65c584af5351d4518a08845fefcbcf568fbf840a3599dc0a143941d2f1d
3
+ size 33640010
only_posix_100_really_fixed/adapter_3000/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
only_posix_100_really_fixed/adapter_3000/adapter_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "LlavaMistralForCausalLM",
4
+ "parent_library": "llava.model.language_model.llava_mistral"
5
+ },
6
+ "base_model_name_or_path": "/share/ssddata/sarimhashmi/posix/llava_med/llava-med",
7
+ "bias": "none",
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layers_pattern": null,
12
+ "layers_to_transform": null,
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "modules_to_save": null,
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "revision": null,
19
+ "target_modules": [
20
+ "q_proj",
21
+ "v_proj"
22
+ ],
23
+ "task_type": null
24
+ }
only_posix_100_really_fixed/adapter_3000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73bf2dd8c04e51f4555c901a08e6e6592f80ea74e16c9e7847b51176bf4f5719
3
+ size 33640010
only_posix_100_really_fixed/adapter_3500/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0