Update services/llama_generator.py
Browse files
services/llama_generator.py
CHANGED
@@ -6,7 +6,9 @@ from datetime import datetime
|
|
6 |
import logging
|
7 |
from config.config import settings
|
8 |
|
9 |
-
|
|
|
|
|
10 |
from services.base_generator import BaseGenerator
|
11 |
|
12 |
import asyncio
|
@@ -46,6 +48,7 @@ class LlamaGenerator(BaseGenerator):
|
|
46 |
print(llama_model_name)
|
47 |
print(prm_model_path)
|
48 |
|
|
|
49 |
|
50 |
self.tokenizer = model_manager.load_tokenizer(llama_model_name) # Add this line to initialize the tokenizer
|
51 |
|
|
|
6 |
import logging
|
7 |
from config.config import settings
|
8 |
|
9 |
+
from services.prompt_builder import LlamaPromptTemplate
|
10 |
+
from services.model_manager import ModelManager
|
11 |
+
|
12 |
from services.base_generator import BaseGenerator
|
13 |
|
14 |
import asyncio
|
|
|
48 |
print(llama_model_name)
|
49 |
print(prm_model_path)
|
50 |
|
51 |
+
self.model_manager = ModelManager()
|
52 |
|
53 |
self.tokenizer = model_manager.load_tokenizer(llama_model_name) # Add this line to initialize the tokenizer
|
54 |
|