supkon commited on
Commit
ec6e43b
·
verified ·
1 Parent(s): 102dd8e

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -11,7 +11,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 16,
14
- "lora_dropout": 0.1,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 16,
14
+ "lora_dropout": 0.5,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c2b0314111e3a7851a8eb89b55faf5a80d67aba991daf30a2d387899ca27aa6
3
- size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbcd7027318d72007204cd04910a68324a0bdb5c580f6d0daa183d3bba065e2d
3
+ size 109071352
runs/Dec07_04-57-48_e169ffdb36e0/events.out.tfevents.1733547477.e169ffdb36e0.4383.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf9dcad651b62d3e50d739d7361a7f8ddd3de4c609eae048b381f4a36969657c
3
+ size 7080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce382add86d3930aef7604cc169c888e6ba80577be8091f8c8d237bc812948fe
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f35f0228a40e342c6d45e9e015f55bc5edb30b9ab8728f15e87fc693c2bf00ae
3
  size 5560