abetlen commited on
Commit
1bb60da
1 Parent(s): 3dd5a0f
Files changed (2) hide show
  1. app.py +2 -2
  2. models.py +2 -1
app.py CHANGED
@@ -245,7 +245,7 @@ if __name__ == '__main__':
245
 
246
  gradio_helpers.set_warmup_function(warmup)
247
  for name, (repo, filename) in models.MODELS.items():
248
- gradio_helpers.register_download(name, repo, filename + "-text-model-q4_k_m.gguf")
249
- gradio_helpers.register_download(name, repo, filename + "-mmproj-f16.gguf")
250
 
251
  create_app().queue().launch()
 
245
 
246
  gradio_helpers.set_warmup_function(warmup)
247
  for name, (repo, filename) in models.MODELS.items():
248
+ gradio_helpers.register_download(name + "-text-model-q4_k_m.gguf", repo, filename + "-text-model-q4_k_m.gguf")
249
+ gradio_helpers.register_download(name + "-mmproj-f16.gguf", repo, filename + "-mmproj-f16.gguf")
250
 
251
  create_app().queue().launch()
models.py CHANGED
@@ -78,7 +78,8 @@ def generate(
78
  # with gradio_helpers.timed('computation', start_message=True):
79
  # tokens = model.predict(params, batch, sampler=sampler)
80
 
81
- model_path = gradio_helpers.get_paths()[model_name]
 
82
  print(model_path)
83
  print(gradio_helpers.get_paths())
84
  model = llama_cpp.Llama(
 
78
  # with gradio_helpers.timed('computation', start_message=True):
79
  # tokens = model.predict(params, batch, sampler=sampler)
80
 
81
+ model_path = gradio_helpers.get_paths()[model_name + "-text-model-q4_k_m.gguf"]
82
+ clip_path = gradio_helpers.get_paths()[model_name + "-mmproj-f16.gguf"]
83
  print(model_path)
84
  print(gradio_helpers.get_paths())
85
  model = llama_cpp.Llama(