Deepak Yadav commited on
Commit
6fd83f3
·
1 Parent(s): 16e7ede

updated new version deepseek-r1

Browse files
Files changed (1) hide show
  1. services/llm.py +2 -0
services/llm.py CHANGED
@@ -23,6 +23,7 @@ def initialize_qa_chain(filepath, model_name, temperature, top_p, max_tokens):
23
  # Configure the LLM with additional parameters
24
  llm = OllamaLLM(
25
  model=model_name,
 
26
  temperature=temperature, # Controls randomness (0 = deterministic, 1 = max randomness)
27
  max_tokens=max_tokens, # Limit the number of tokens in the output
28
  top_p=top_p # Nucleus sampling for controlling diversity
@@ -70,6 +71,7 @@ def initialize_chain(model_name, temperature, top_p, max_tokens):
70
  # Configure the LLM with additional parameters
71
  llm = OllamaLLM(
72
  model=model_name,
 
73
  temperature=temperature, # Controls randomness (0 = deterministic, 1 = max randomness)
74
  max_tokens=max_tokens, # Limit the number of tokens in the output
75
  top_p=top_p # Nucleus sampling for controlling diversity
 
23
  # Configure the LLM with additional parameters
24
  llm = OllamaLLM(
25
  model=model_name,
26
+ base_url="https://deepak7376-ollama-server.hf.space",
27
  temperature=temperature, # Controls randomness (0 = deterministic, 1 = max randomness)
28
  max_tokens=max_tokens, # Limit the number of tokens in the output
29
  top_p=top_p # Nucleus sampling for controlling diversity
 
71
  # Configure the LLM with additional parameters
72
  llm = OllamaLLM(
73
  model=model_name,
74
+ base_url="https://deepak7376-ollama-server.hf.space",
75
  temperature=temperature, # Controls randomness (0 = deterministic, 1 = max randomness)
76
  max_tokens=max_tokens, # Limit the number of tokens in the output
77
  top_p=top_p # Nucleus sampling for controlling diversity