/** @format */ function getPayloadDoc(prompt) { return { model: "", temperature: 0.3, max_tokens: 1024, stream: false, random_seed: 42, messages: [{ role: "user", content: prompt }], safe_prompt: false, }; } function getPayloadBuildContext(prompt) { return { model: "", messages: [{ role: "user", content: prompt }], temperature: 0.3, max_tokens: 2000, stream: false, safe_prompt: false, random_seed: 42, }; } function getPayloadWithContext(prompt) { return { model: "", messages: [{ role: "user", content: prompt }], temperature: 0.3, max_tokens: 2000, stream: false, safe_prompt: false, random_seed: 42, }; } function getPayloadThread(prompt) { return { model: "", messages: [{ role: "user", content: prompt }], temperature: 0.7, max_tokens: 2000, stream: false, safe_prompt: false, random_seed: 42, }; } /* function getPayloadDoc(prompt) { const payload = { inputs: prompt, parameters: { task: "text2text-generation", max_new_tokens: 2000, num_return_sequences: 1, temperature: 0.4, top_p: 0.85, top_k: 30, do_sample: false, no_repeat_ngram_size: 3, num_beams: 4, repetition_penalty: 1.2, return_full_text: false, details: false, max_time: 90.0, seed: 42, }, options: { use_cache: false, wait_for_model: true, }, }; return payload; } function getPayloadBuildContext(prompt) { const payload = { inputs: prompt, parameters: { task: "text2text-generation", max_new_tokens: 6000, num_return_sequences: 1, temperature: 0.7, top_p: 0.85, top_k: 30, do_sample: false, no_repeat_ngram_size: 4, num_beams: 6, repetition_penalty: 1.2, return_full_text: false, details: false, max_time: 180.0, seed: 42, }, options: { use_cache: false, wait_for_model: true, }, }; return payload; } function getPayloadWithContext(prompt) { const payload = { inputs: prompt, parameters: { task: "text2text-generation", max_new_tokens: 4000, num_return_sequences: 1, temperature: 0.7, top_p: 0.85, top_k: 30, do_sample: false, no_repeat_ngram_size: 4, num_beams: 5, repetition_penalty: 1.4, return_full_text: false, details: false, max_time: 120.0, seed: 42, }, options: { use_cache: false, wait_for_model: true, }, }; return payload; } function getPayloadThread(prompt) { const payload = { inputs: prompt, parameters: { task: "text2text-generation", max_new_tokens: 6048, num_return_sequences: 1, temperature: 0.7, top_p: 0.85, top_k: 30, do_sample: false, no_repeat_ngram_size: 4, num_beams: 5, repetition_penalty: 1.4, return_full_text: false, details: false, max_time: 120.0, seed: 42, }, options: { use_cache: false, wait_for_model: true, }, }; return payload; } */