ianmckenzie commited on
Commit
f16954d
·
verified ·
1 Parent(s): 6173957

Pushing model and tokenizer to hub

Browse files
commit.txt CHANGED
@@ -1 +1 @@
1
- 2f355402bac52f09d975b92cf9199da42509a9ea
 
1
+ ca86dd7d550be87f21ead04ebb792e9fd7ccdb0c
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "EleutherAI/pythia-14m",
3
  "architectures": [
4
  "GPTNeoXForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "/robust_llm_data/checkpoints/6df328e8c92cfe5613549772d3a3e4276e232892aec3c486d5c03abbec7033f4/epoch_0004/model",
3
  "architectures": [
4
  "GPTNeoXForSequenceClassification"
5
  ],
disk-write-timestamp.txt CHANGED
@@ -1 +1 @@
1
- 20241022-234722
 
1
+ 20241023-182353
tokenizer_config.json CHANGED
@@ -207,9 +207,13 @@
207
  "bos_token": "<|endoftext|>",
208
  "clean_up_tokenization_spaces": false,
209
  "eos_token": "<|endoftext|>",
 
210
  "model_max_length": 2048,
211
  "pad_token": "<|endoftext|>",
212
  "padding_side": "right",
 
213
  "tokenizer_class": "GPTNeoXTokenizer",
 
 
214
  "unk_token": "<|endoftext|>"
215
  }
 
207
  "bos_token": "<|endoftext|>",
208
  "clean_up_tokenization_spaces": false,
209
  "eos_token": "<|endoftext|>",
210
+ "max_length": 2048,
211
  "model_max_length": 2048,
212
  "pad_token": "<|endoftext|>",
213
  "padding_side": "right",
214
+ "stride": 0,
215
  "tokenizer_class": "GPTNeoXTokenizer",
216
+ "truncation_side": "right",
217
+ "truncation_strategy": "longest_first",
218
  "unk_token": "<|endoftext|>"
219
  }