Spaces:
Running
on
Zero
Running
on
Zero
Commit
·
e49ede1
1
Parent(s):
4954b56
Use separate var name
Browse files
app.py
CHANGED
@@ -14,7 +14,6 @@ huggingface_hub.login(token=hf_key)
|
|
14 |
|
15 |
tokenizer = AutoTokenizer.from_pretrained(
|
16 |
"bigcode/starcoderbase-3b"
|
17 |
-
# , use_auth_token=hf_key
|
18 |
)
|
19 |
vardecoder_model = AutoModelForCausalLM.from_pretrained(
|
20 |
"ejschwartz/resym-vardecoder", torch_dtype=torch.bfloat16, device_map="auto"
|
@@ -38,10 +37,10 @@ example = """{
|
|
38 |
|
39 |
|
40 |
@spaces.GPU
|
41 |
-
def infer(input):
|
42 |
line = json.loads(input)
|
43 |
first_token = line["output"].split(":")[0]
|
44 |
-
prompt = line["input"] +
|
45 |
|
46 |
input_ids = tokenizer.encode(prompt, return_tensors="pt").cuda()[:, : 8192 - 1024]
|
47 |
output = vardecoder_model.generate(
|
|
|
14 |
|
15 |
tokenizer = AutoTokenizer.from_pretrained(
|
16 |
"bigcode/starcoderbase-3b"
|
|
|
17 |
)
|
18 |
vardecoder_model = AutoModelForCausalLM.from_pretrained(
|
19 |
"ejschwartz/resym-vardecoder", torch_dtype=torch.bfloat16, device_map="auto"
|
|
|
37 |
|
38 |
|
39 |
@spaces.GPU
|
40 |
+
def infer(var_name, input):
|
41 |
line = json.loads(input)
|
42 |
first_token = line["output"].split(":")[0]
|
43 |
+
prompt = line["input"] + var_name + ":"
|
44 |
|
45 |
input_ids = tokenizer.encode(prompt, return_tensors="pt").cuda()[:, : 8192 - 1024]
|
46 |
output = vardecoder_model.generate(
|