Spaces:
Running
Running
import os | |
import streamlit as st | |
from huggingface_hub import InferenceClient | |
client = InferenceClient(api_key=os.environ['HF_KEY']) | |
messages = [ | |
{ | |
"role": "user", | |
"content": "What is the capital of France?" | |
} | |
] | |
stream = client.chat.completions.create( | |
model="HuggingFaceTB/SmolLM2-1.7B-Instruct", | |
messages=messages, | |
max_tokens=500, | |
stream=True | |
) | |
for chunk in stream: | |
print(chunk.choices[0].delta.content, end="") |