Renegadesoffun commited on
Commit
7da9883
·
1 Parent(s): a71b8eb

Updated for CPU evalgguf

Browse files
Files changed (1) hide show
  1. app.py +19 -13
app.py CHANGED
@@ -1,14 +1,15 @@
1
  import streamlit as st
2
- from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
4
 
5
- # Use a medium sized model
6
- model_name = "microsoft/DialoGPT-medium"
7
 
8
- model = AutoModelForCausalLM.from_pretrained(model_name)
9
- model.eval() # Make sure model is in evaluation mode
 
10
 
11
- tokenizer = AutoTokenizer.from_pretrained(model_name)
 
12
 
13
  st.title("Buddy Christ Chatbot")
14
 
@@ -16,14 +17,19 @@ user_input = st.text_input("You:", "")
16
 
17
  if user_input:
18
 
 
19
  inputs = tokenizer.encode(user_input, return_tensors="pt")
20
-
 
21
  response = model.generate(inputs,
22
- max_length=1000,
23
- temperature=0.8,
24
- top_k=50,
25
- pad_token_id=tokenizer.eos_token_id)
26
-
27
- response_text = tokenizer.decode(response[0], skip_special_tokens=True)
 
 
 
28
 
29
  st.write("Buddy Christ:", response_text)
 
1
  import streamlit as st
2
+ from ggulf import GGUFModel, GGUFTokenizer
3
  import torch
4
 
5
+ model_name = "TheBloke/TinyLlama-1.1B-Chat-v0.3-GGUF"
 
6
 
7
+ # Load model and tokenizer
8
+ model = GGUFModel.from_pretrained(model_name)
9
+ tokenizer = GGUFTokenizer.from_pretrained(model_name)
10
 
11
+ # Set model to eval mode
12
+ model.eval()
13
 
14
  st.title("Buddy Christ Chatbot")
15
 
 
17
 
18
  if user_input:
19
 
20
+ # Encode input
21
  inputs = tokenizer.encode(user_input, return_tensors="pt")
22
+
23
+ # Generate response using GGUF
24
  response = model.generate(inputs,
25
+ max_length=1000,
26
+ temperature=1.0,
27
+ top_k=10,
28
+ pad_token_id=tokenizer.eos_token_id,
29
+ gguf_mode=True)
30
+
31
+ # Print and display full response
32
+ print(response_text)
33
+ response_text = tokenizer.decode(response[0])
34
 
35
  st.write("Buddy Christ:", response_text)