erikbeltran commited on
Commit
8094502
·
verified ·
1 Parent(s): f56787c

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "HuggingFaceTB/SmolLM-135M",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "erikbeltran/ruby-autotab-v6",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
events.out.tfevents.1725658287.beb32e8e7a14.36.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f63dd710b98eb11941b5d54f895a3714856ed1909060fcea86dccc204b378ca
3
+ size 5741
events.out.tfevents.1725658398.beb32e8e7a14.36.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a499fb4db458cded5d76e8226f81575f4c9fac064cedc03469ac9b01003f28aa
3
+ size 4184
events.out.tfevents.1725658661.beb32e8e7a14.192.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20bee010965423405b8a920250a00ba563aac3bd3f4d487e809798c22d5090e7
3
+ size 6375
events.out.tfevents.1725659511.beb32e8e7a14.383.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c41f7d45a318efd207f9e6321b107009ce96a836ad0d9d21df48b493091e112
3
+ size 4184
events.out.tfevents.1725659622.beb32e8e7a14.474.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64bc9ffc1d37bbd378cdc2017ab4a294e46ee5a6f486aacd53c261cc44cf5926
3
+ size 8143
events.out.tfevents.1725662095.beb32e8e7a14.474.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90c5ba5a4a8b8e8bb13243070022e79a8f39534c6cb85a5a9fb89c99f8209a49
3
+ size 5469
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e60fc029fca287f98b13a92add90db482bc0b3d33800f43ee06351fad56818dd
3
  size 538090408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b86a22201b137f0553ed9faf8f3ab20c6961a6c55caad649c4c9ef1f51bd4054
3
  size 538090408
special_tokens_map.json CHANGED
@@ -32,7 +32,13 @@
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
36
  "unk_token": {
37
  "content": "<|endoftext|>",
38
  "lstrip": false,
 
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
+ "pad_token": {
36
+ "content": "<|endoftext|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false
41
+ },
42
  "unk_token": {
43
  "content": "<|endoftext|>",
44
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -160,9 +160,13 @@
160
  "bos_token": "<|endoftext|>",
161
  "clean_up_tokenization_spaces": false,
162
  "eos_token": "<|endoftext|>",
 
163
  "model_max_length": 1000000000000000019884624838656,
164
  "pad_token": "<|endoftext|>",
 
165
  "tokenizer_class": "GPT2Tokenizer",
 
 
166
  "unk_token": "<|endoftext|>",
167
  "vocab_size": 49152
168
  }
 
160
  "bos_token": "<|endoftext|>",
161
  "clean_up_tokenization_spaces": false,
162
  "eos_token": "<|endoftext|>",
163
+ "max_length": 1024,
164
  "model_max_length": 1000000000000000019884624838656,
165
  "pad_token": "<|endoftext|>",
166
+ "stride": 0,
167
  "tokenizer_class": "GPT2Tokenizer",
168
+ "truncation_side": "right",
169
+ "truncation_strategy": "longest_first",
170
  "unk_token": "<|endoftext|>",
171
  "vocab_size": 49152
172
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:129ba933e8ae3645669037fba183ba605aeb7726ed8a5b9c934841638b01fc4d
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c8acf4fd852feab5d5e82355d74f632d4a739b817fa0dba91bc79f88f2f478c
3
  size 5496