rkumar1999 commited on
Commit
d9e8e8f
·
verified ·
1 Parent(s): e2fe931

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: meta-llama/Llama-3.1-8B-Instruct
3
- datasets: Tonic/MiniF2F
4
  library_name: transformers
5
  model_name: Llama-3.1-8B-Instruct-Open-R1-Distill-Lean
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - sft
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for Llama-3.1-8B-Instruct-Open-R1-Distill-Lean
15
 
16
- This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on the [Tonic/MiniF2F](https://huggingface.co/datasets/Tonic/MiniF2F) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/rohanbayya1205-san-jose-state-university/huggingface/runs/ws8rmwff)
33
 
34
 
35
  This model was trained with SFT.
 
1
  ---
2
  base_model: meta-llama/Llama-3.1-8B-Instruct
 
3
  library_name: transformers
4
  model_name: Llama-3.1-8B-Instruct-Open-R1-Distill-Lean
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for Llama-3.1-8B-Instruct-Open-R1-Distill-Lean
13
 
14
+ This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/rohanbayya1205-san-jose-state-university/huggingface/runs/rl8xdbu7)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -24,9 +24,9 @@
24
  "revision": null,
25
  "target_modules": [
26
  "o_proj",
 
27
  "k_proj",
28
- "v_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": true,
 
24
  "revision": null,
25
  "target_modules": [
26
  "o_proj",
27
+ "q_proj",
28
  "k_proj",
29
+ "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f9b528d33a706136dec74218a64d50402715fa9285eb6f9d674a1b99a1184e5
3
  size 7522752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d80adfc6a5efc9cfe88bcdaf9fee15d1da4c8e085b7b344176dd37d21f39eb80
3
  size 7522752
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 196696080384.0,
3
- "train_loss": 3363.7795448303223,
4
- "train_runtime": 10.1807,
5
  "train_samples": 488,
6
- "train_samples_per_second": 1.768,
7
- "train_steps_per_second": 0.196
8
  }
 
1
  {
2
+ "total_flos": 3737225527296.0,
3
+ "train_loss": 33168.128967285156,
4
+ "train_runtime": 78.7129,
5
  "train_samples": 488,
6
+ "train_samples_per_second": 2.287,
7
+ "train_steps_per_second": 0.254
8
  }
runs/Mar03_21-17-14_ip-10-192-12-214/events.out.tfevents.1741036644.ip-10-192-12-214.31425.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df0ca1acb1f86d1f3e1d89675ba8a6f99f2ab6dfcd414fad91a483b8e19a063b
3
+ size 6506
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 196696080384.0,
3
- "train_loss": 3363.7795448303223,
4
- "train_runtime": 10.1807,
5
  "train_samples": 488,
6
- "train_samples_per_second": 1.768,
7
- "train_steps_per_second": 0.196
8
  }
 
1
  {
2
+ "total_flos": 3737225527296.0,
3
+ "train_loss": 33168.128967285156,
4
+ "train_runtime": 78.7129,
5
  "train_samples": 488,
6
+ "train_samples_per_second": 2.287,
7
+ "train_steps_per_second": 0.254
8
  }
trainer_state.json CHANGED
@@ -10,19 +10,19 @@
10
  "log_history": [
11
  {
12
  "epoch": 0,
13
- "mean_token_accuracy": 0.6600732803344727,
14
  "step": 0,
15
- "total_flos": 196696080384.0,
16
- "train_loss": 3363.7795448303223,
17
- "train_runtime": 10.1807,
18
- "train_samples_per_second": 1.768,
19
- "train_steps_per_second": 0.196
20
  }
21
  ],
22
  "logging_steps": 500,
23
- "max_steps": 2,
24
  "num_input_tokens_seen": 0,
25
- "num_train_epochs": 2,
26
  "save_steps": 500,
27
  "stateful_callbacks": {
28
  "TrainerControl": {
@@ -36,7 +36,7 @@
36
  "attributes": {}
37
  }
38
  },
39
- "total_flos": 196696080384.0,
40
  "train_batch_size": 2,
41
  "trial_name": null,
42
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0,
13
+ "mean_token_accuracy": 0.6642943471670151,
14
  "step": 0,
15
+ "total_flos": 3737225527296.0,
16
+ "train_loss": 33168.128967285156,
17
+ "train_runtime": 78.7129,
18
+ "train_samples_per_second": 2.287,
19
+ "train_steps_per_second": 0.254
20
  }
21
  ],
22
  "logging_steps": 500,
23
+ "max_steps": 20,
24
  "num_input_tokens_seen": 0,
25
+ "num_train_epochs": 20,
26
  "save_steps": 500,
27
  "stateful_callbacks": {
28
  "TrainerControl": {
 
36
  "attributes": {}
37
  }
38
  },
39
+ "total_flos": 3737225527296.0,
40
  "train_batch_size": 2,
41
  "trial_name": null,
42
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02d9addc654a14dc4841e1d019ae713ceb25955f72671e8521e85905de27bf43
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe6a1413db8bfaa2c249f4c9a06e39b05fc03a29e90894f1e1a020a801111c97
3
  size 7352