Update run.py
Browse files
run.py
CHANGED
@@ -54,13 +54,13 @@ def biomarker_prediction(smile_acc, smile_don):
|
|
54 |
|
55 |
|
56 |
das_input =smile_don
|
57 |
-
d_inputs = tokenizer(aas_input, padding='max_length', max_length=
|
58 |
# d_inputs = tokenizer(smiles, truncation=True, return_tensors="pt")
|
59 |
drug_input_ids = d_inputs['input_ids'].to(device)
|
60 |
drug_attention_mask = d_inputs['attention_mask'].to(device)
|
61 |
drug_inputs = {'input_ids': drug_input_ids, 'attention_mask': drug_attention_mask}
|
62 |
|
63 |
-
p_inputs = prot_tokenizer(das_input, padding='max_length', max_length=
|
64 |
# p_inputs = prot_tokenizer(aas_input, truncation=True, return_tensors="pt")
|
65 |
prot_input_ids = p_inputs['input_ids'].to(device)
|
66 |
prot_attention_mask = p_inputs['attention_mask'].to(device)
|
|
|
54 |
|
55 |
|
56 |
das_input =smile_don
|
57 |
+
d_inputs = tokenizer(aas_input, padding='max_length', max_length=510, truncation=True, return_tensors="pt")
|
58 |
# d_inputs = tokenizer(smiles, truncation=True, return_tensors="pt")
|
59 |
drug_input_ids = d_inputs['input_ids'].to(device)
|
60 |
drug_attention_mask = d_inputs['attention_mask'].to(device)
|
61 |
drug_inputs = {'input_ids': drug_input_ids, 'attention_mask': drug_attention_mask}
|
62 |
|
63 |
+
p_inputs = prot_tokenizer(das_input, padding='max_length', max_length=510, truncation=True, return_tensors="pt")
|
64 |
# p_inputs = prot_tokenizer(aas_input, truncation=True, return_tensors="pt")
|
65 |
prot_input_ids = p_inputs['input_ids'].to(device)
|
66 |
prot_attention_mask = p_inputs['attention_mask'].to(device)
|