Inference-API / main /config.yaml
AurelioAguirre's picture
added openAI schema based endpoint and response
02fd6bb
raw
history blame
511 Bytes
server:
host: "localhost"
port: 8001
timeout: 60
max_batch_size: 1
llm_server:
base_url: "https://teamgenki-llmserver.hf.space:7860"
timeout: 60.0
api_prefix: "/api/v1" # This will be used for route prefixing
endpoints:
generate: "/generate"
generate_stream: "/generate/stream"
embedding: "/embedding"
system_status: "/system/status"
system_validate: "/system/validate"
model_initialize: "/model/initialize"
model_initialize_embedding: "/model/initialize/embedding"