dar-tau commited on
Commit
ee8bd6d
·
verified ·
1 Parent(s): 997caf4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -94,8 +94,7 @@ def run_interpretation(global_state, raw_interpretation_prompt, max_new_tokens,
94
  interpretation_prompt = InterpretationPrompt(tokenizer, interpretation_prompt)
95
 
96
  # generate the interpretations
97
- generate = generate_interpretation_gpu if use_gpu else lambda interpretation_prompt, *args,
98
- **kwargs: interpretation_prompt.generate(*args, **kwargs)
99
  generated = generate(interpretation_prompt, model, {0: interpreted_vectors}, k=3, **generation_kwargs)
100
  generation_texts = tokenizer.batch_decode(generated)
101
  progress_dummy_output = ''
 
94
  interpretation_prompt = InterpretationPrompt(tokenizer, interpretation_prompt)
95
 
96
  # generate the interpretations
97
+ generate = generate_interpretation_gpu if use_gpu else lambda interpretation_prompt, *args, **kwargs: interpretation_prompt.generate(*args, **kwargs)
 
98
  generated = generate(interpretation_prompt, model, {0: interpreted_vectors}, k=3, **generation_kwargs)
99
  generation_texts = tokenizer.batch_decode(generated)
100
  progress_dummy_output = ''