sachin commited on
Commit
f4d05d5
·
1 Parent(s): 6173695

temp-disable quantisation

Browse files
Files changed (1) hide show
  1. src/server/gemma_llm.py +1 -1
src/server/gemma_llm.py CHANGED
@@ -39,7 +39,7 @@ class LLMManager:
39
  self.model = Gemma3ForConditionalGeneration.from_pretrained(
40
  self.model_name,
41
  device_map="auto",
42
- quantization_config=quantization_config,
43
  torch_dtype=self.torch_dtype
44
  ).eval()
45
  self.processor = AutoProcessor.from_pretrained(self.model_name)
 
39
  self.model = Gemma3ForConditionalGeneration.from_pretrained(
40
  self.model_name,
41
  device_map="auto",
42
+ #quantization_config=quantization_config,
43
  torch_dtype=self.torch_dtype
44
  ).eval()
45
  self.processor = AutoProcessor.from_pretrained(self.model_name)