
Add requirements.txt for synthetic dataset generator dependency, update import paths for InferenceEndpointsLLM in multiple files, and modify example deployment script to comment out HF_TOKEN.
1fc08db
# /// script | |
# requires-python = ">=3.11,<3.12" | |
# dependencies = [ | |
# "synthetic-dataset-generator", | |
# ] | |
# /// | |
import os | |
from synthetic_dataset_generator import launch | |
# os.environ["HF_TOKEN"] = "hf_..." # push the data to huggingface | |
os.environ["MODEL"] = "meta-llama/Llama-3.1-8B-Instruct" # use instruct model | |
os.environ["MAGPIE_PRE_QUERY_TEMPLATE"] = "llama3" # use the template for the model | |
launch() | |