Ahmed007 commited on
Commit
08921b4
·
verified ·
1 Parent(s): 4f8064c

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +10 -16
main.py CHANGED
@@ -5,8 +5,6 @@ from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
5
  from langchain.prompts import PromptTemplate
6
  from langchain.schema.output_parser import StrOutputParser
7
 
8
- # from langchain.llms import LlamaCpp
9
-
10
  app = Flask(__name__)
11
 
12
  n_gpu_layers = 0
@@ -15,7 +13,7 @@ n_batch = 1024
15
  callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
16
 
17
  llm = LlamaCpp(
18
- model_path="phi-2.Q4_K_M.gguf",
19
  temperature=0.1,
20
  n_gpu_layers=n_gpu_layers,
21
  n_batch=n_batch,
@@ -28,19 +26,15 @@ llm = LlamaCpp(
28
  def get_skills():
29
  cv_body = request.json.get('cv_body')
30
 
31
- template = """Instruct:
32
- Take a deep breath to deep understand, and don't this cv vc = {cv_body} . to answer this question and instructions </s>
33
- \nOutput:"""
34
-
35
- prompt = PromptTemplate(template=template, input_variables=["text"])
36
- chain = prompt | llm | StrOutputParser()
37
-
38
- ans = chain.invoke({"question": "What are his best skills? write in points","cv_body":cv_body},
39
- config={
40
- # "callbacks": [ConsoleCallbackHandler()]
41
- })
42
 
43
- return jsonify({'skills': ans})
44
 
45
  if __name__ == '__main__':
46
- app.run()
 
5
  from langchain.prompts import PromptTemplate
6
  from langchain.schema.output_parser import StrOutputParser
7
 
 
 
8
  app = Flask(__name__)
9
 
10
  n_gpu_layers = 0
 
13
  callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
14
 
15
  llm = LlamaCpp(
16
+ model_path="Phi-3-mini-4k-instruct-q4.gguf", # path to GGUF file
17
  temperature=0.1,
18
  n_gpu_layers=n_gpu_layers,
19
  n_batch=n_batch,
 
26
  def get_skills():
27
  cv_body = request.json.get('cv_body')
28
 
29
+ # Simple inference example
30
+ output = llm(
31
+ f"<|user|>\n{cv_body}<|end|>\n<|assistant|>Can you list the skills mentioned in the CV?<|end|>",
32
+ max_tokens=256, # Generate up to 256 tokens
33
+ stop=["<|end|>"],
34
+ echo=True, # Whether to echo the prompt
35
+ )
 
 
 
 
36
 
37
+ return jsonify({'skills': output})
38
 
39
  if __name__ == '__main__':
40
+ app.run()