Update app.py
Browse files
app.py
CHANGED
@@ -59,21 +59,21 @@ model_id_6 = "sbcBI/sentiment_analysis_model"
|
|
59 |
model_id_7 = "oliverguhr/german-sentiment-bert"
|
60 |
|
61 |
# https://colab.research.google.com/drive/1hrS6_g14EcOD4ezwSGlGX2zxJegX5uNX#scrollTo=NUwUR9U7qkld
|
62 |
-
llm_hf_sentiment = HuggingFaceHub(
|
63 |
-
repo_id= model_id_7,
|
64 |
-
model_kwargs={"temperature":0.9 }
|
65 |
-
)
|
66 |
|
67 |
-
sentiment_prompt = PromptTemplate(
|
68 |
-
input_variables=["text_input"],
|
69 |
-
template="Extract the key facts out of this text. Don't include opinions. Give each fact a number and keep them short sentences. :\n\n {text_input}"
|
70 |
-
)
|
71 |
|
72 |
-
def sentiment ( message):
|
73 |
-
sentiment_chain = LLMChain(llm=llm_hf_sentiment, prompt=sentiment_prompt)
|
74 |
-
facts = sentiment_chain.run(message)
|
75 |
-
print(facts)
|
76 |
-
return facts
|
77 |
|
78 |
|
79 |
|
@@ -104,8 +104,8 @@ text = "Why did the chicken cross the road?"
|
|
104 |
# https://colab.research.google.com/drive/1hrS6_g14EcOD4ezwSGlGX2zxJegX5uNX#scrollTo=NUwUR9U7qkld
|
105 |
llm_factextract = HuggingFaceHub(
|
106 |
|
107 |
-
|
108 |
-
repo_id="google/flan-t5-small",
|
109 |
model_kwargs={"temperature":0.1,
|
110 |
"max_new_tokens":256})
|
111 |
|
|
|
59 |
model_id_7 = "oliverguhr/german-sentiment-bert"
|
60 |
|
61 |
# https://colab.research.google.com/drive/1hrS6_g14EcOD4ezwSGlGX2zxJegX5uNX#scrollTo=NUwUR9U7qkld
|
62 |
+
#llm_hf_sentiment = HuggingFaceHub(
|
63 |
+
# repo_id= model_id_7,
|
64 |
+
# model_kwargs={"temperature":0.9 }
|
65 |
+
#)
|
66 |
|
67 |
+
#sentiment_prompt = PromptTemplate(
|
68 |
+
# input_variables=["text_input"],
|
69 |
+
# template="Extract the key facts out of this text. Don't include opinions. Give each fact a number and keep them short sentences. :\n\n {text_input}"
|
70 |
+
#)
|
71 |
|
72 |
+
#def sentiment ( message):
|
73 |
+
# sentiment_chain = LLMChain(llm=llm_hf_sentiment, prompt=sentiment_prompt)
|
74 |
+
# facts = sentiment_chain.run(message)
|
75 |
+
# print(facts)
|
76 |
+
# return facts
|
77 |
|
78 |
|
79 |
|
|
|
104 |
# https://colab.research.google.com/drive/1hrS6_g14EcOD4ezwSGlGX2zxJegX5uNX#scrollTo=NUwUR9U7qkld
|
105 |
llm_factextract = HuggingFaceHub(
|
106 |
|
107 |
+
repo_id="google/flan-ul2",
|
108 |
+
# repo_id="google/flan-t5-small",
|
109 |
model_kwargs={"temperature":0.1,
|
110 |
"max_new_tokens":256})
|
111 |
|