Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -15,10 +15,11 @@ import re
|
|
15 |
HF_API_TOKEN = os.getenv("HUG_TOKEN_READ2")
|
16 |
|
17 |
# Hugging Face Inference API Client
|
18 |
-
client = InferenceClient(model="mistralai/Mistral-7B-Instruct-v0.1", token=HF_API_TOKEN)
|
19 |
#client = InferenceClient(model="openGPT-X/Teuken-7B-instruct-commercial-v0.4", token=HF_API_TOKEN)
|
20 |
#client = InferenceClient(model=qwenModel, token=HF_API_TOKEN)
|
21 |
#client = InferenceClient(model="microsoft/Phi-4-mini-instruct", token=HF_API_TOKEN)
|
|
|
22 |
|
23 |
# Function to translate text into emojis
|
24 |
def text_to_emoji(text):
|
@@ -38,4 +39,4 @@ iface = gr.Interface(
|
|
38 |
description="Enter a sentence, and the AI will transform it into an emoji-version 🥳"
|
39 |
)
|
40 |
|
41 |
-
iface.launch(
|
|
|
15 |
HF_API_TOKEN = os.getenv("HUG_TOKEN_READ2")
|
16 |
|
17 |
# Hugging Face Inference API Client
|
18 |
+
#client = InferenceClient(model="mistralai/Mistral-7B-Instruct-v0.1", token=HF_API_TOKEN)
|
19 |
#client = InferenceClient(model="openGPT-X/Teuken-7B-instruct-commercial-v0.4", token=HF_API_TOKEN)
|
20 |
#client = InferenceClient(model=qwenModel, token=HF_API_TOKEN)
|
21 |
#client = InferenceClient(model="microsoft/Phi-4-mini-instruct", token=HF_API_TOKEN)
|
22 |
+
client = InferenceClient(model="openai-community/gpt2", token=HF_API_TOKEN)
|
23 |
|
24 |
# Function to translate text into emojis
|
25 |
def text_to_emoji(text):
|
|
|
39 |
description="Enter a sentence, and the AI will transform it into an emoji-version 🥳"
|
40 |
)
|
41 |
|
42 |
+
iface.launch()
|