toilaluan commited on
Commit
18501f6
·
1 Parent(s): 3ef1793
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -42,8 +42,7 @@ def inference(question: str, doc_path: str, use_turbo=True) -> str:
42
  else:
43
  print("turbo-mode-off")
44
  kv_cache = get_model_kv_cache(context_ids)
45
-
46
- print("kv-length", kv_cache.get_seq_length())
47
 
48
  answer = generate_answer(MODEL, TOKENIZER, prompt_ids, kv_cache, context_length, 128)
49
  print(answer)
 
42
  else:
43
  print("turbo-mode-off")
44
  kv_cache = get_model_kv_cache(context_ids)
45
+
 
46
 
47
  answer = generate_answer(MODEL, TOKENIZER, prompt_ids, kv_cache, context_length, 128)
48
  print(answer)