Spaces:
Runtime error
Runtime error
Yash Sachdeva
commited on
Commit
·
21be15d
1
Parent(s):
a0347c2
python syntax
Browse files- question_paper.py +0 -6
question_paper.py
CHANGED
@@ -4,12 +4,8 @@ import os
|
|
4 |
|
5 |
from fastapi import FastAPI
|
6 |
|
7 |
-
from transformers import AutoTokenizer
|
8 |
-
|
9 |
from llama_cpp import Llama
|
10 |
|
11 |
-
from huggingface_hub import hf_hub_download
|
12 |
-
|
13 |
app = FastAPI()
|
14 |
@app.get("/")
|
15 |
def llama():
|
@@ -25,7 +21,5 @@ def llama():
|
|
25 |
max_tokens=32, # Generate up to 32 tokens, set to None to generate up to the end of the context window
|
26 |
echo=True # Echo the prompt back in the output
|
27 |
) # Generate a completion, can also call create_completion
|
28 |
-
|
29 |
-
print(output)
|
30 |
|
31 |
return {"output": output}
|
|
|
4 |
|
5 |
from fastapi import FastAPI
|
6 |
|
|
|
|
|
7 |
from llama_cpp import Llama
|
8 |
|
|
|
|
|
9 |
app = FastAPI()
|
10 |
@app.get("/")
|
11 |
def llama():
|
|
|
21 |
max_tokens=32, # Generate up to 32 tokens, set to None to generate up to the end of the context window
|
22 |
echo=True # Echo the prompt back in the output
|
23 |
) # Generate a completion, can also call create_completion
|
|
|
|
|
24 |
|
25 |
return {"output": output}
|