Spaces:
Sleeping
Sleeping
import streamlit as st | |
import subprocess | |
import sys | |
import time | |
def install(package): | |
subprocess.check_call([sys.executable, "-m", "pip", "install", package]) | |
install("llama-cpp-python") | |
from llama_cpp import Llama | |
prompt = st.chat_input("Say something") | |
if prompt: | |
llm = Llama(model_path="Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf") | |
print(f"Started {time.time()}") | |
start = time.time() | |
r=llm(prompt, max_tokens=1000) | |
end = time.time() | |
print(f"The Generation time for 1000 tokens is : {end - start}") | |
l="Nothing" | |
try: | |
l=r["choices"][0]["text"] | |
except Exception as e: | |
print(e) | |
st.write(f"User has sent the following prompt: {prompt} with response: {l} ") |