Kevin Fink commited on
Commit
b8f813e
·
1 Parent(s): 56a80ae
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -207,7 +207,8 @@ def fine_tune_model(model, dataset_name, hub_id, api_key, num_epochs, batch_size
207
  #dataset['train'] = dataset['train'].select(range(8000))
208
  dataset['train'] = dataset['train'].select(range(1000))
209
  dataset['validation'] = dataset['validation'].select(range(100))
210
- tokenized_first_third = dataset.map(tokenize_function, batched=True, batch_size=50, remove_columns=column_names,)
 
211
 
212
 
213
  print('DONE')
@@ -217,8 +218,8 @@ def fine_tune_model(model, dataset_name, hub_id, api_key, num_epochs, batch_size
217
  trainer = Trainer(
218
  model=model,
219
  args=training_args,
220
- train_dataset=dataset['train'],
221
- eval_dataset=dataset['validation'],
222
  compute_metrics=compute_metrics,
223
  data_collator=data_collator,
224
  #processing_class=tokenizer,
 
207
  #dataset['train'] = dataset['train'].select(range(8000))
208
  dataset['train'] = dataset['train'].select(range(1000))
209
  dataset['validation'] = dataset['validation'].select(range(100))
210
+ train_set = dataset['train'].map(tokenize_function, batched=True, batch_size=50, remove_columns=column_names,)
211
+ valid_set = dataset['validation'].map(tokenize_function, batched=True, batch_size=50, remove_columns=column_names,)
212
 
213
 
214
  print('DONE')
 
218
  trainer = Trainer(
219
  model=model,
220
  args=training_args,
221
+ train_dataset=train_set,
222
+ eval_dataset=valid_set,
223
  compute_metrics=compute_metrics,
224
  data_collator=data_collator,
225
  #processing_class=tokenizer,