Spaces:
Running
on
Zero
Running
on
Zero
File size: 1,811 Bytes
57a9ebf 97a6c80 57a9ebf 97a6c80 57a9ebf eba3124 57a9ebf 87edf46 57a9ebf 9da82c0 87edf46 9da82c0 87edf46 eba3124 57a9ebf |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
import gradio as gr
import torch
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch.nn.functional as F
model_path = "ssocean/NAIP" # 更换为你的模型路径
model = AutoModelForSequenceClassification.from_pretrained(model_path, num_labels=1, load_in_8bit=True)
tokenizer = AutoTokenizer.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
model.eval()
def predict(title, abstract):
# 将标题和摘要处理为一个单一的字符串
text = f'''Given a certain paper, Title: {title}\n Abstract: {abstract}. \n Predict its normalized academic impact (between 0 and 1):'''
inputs = tokenizer(text, return_tensors="pt")
with torch.no_grad():
outputs = model(**inputs.to(device))
# 应用 Sigmoid 函数来获取概率输出
probability = torch.sigmoid(outputs.logits).item()
return {"Predicted Impact": round(probability, 4)}
# 创建 Gradio 界面
iface = gr.Interface(
fn=predict,
inputs=[
gr.Textbox(lines=2, placeholder="Enter Paper Title Here...", label="Paper Title"),
gr.Textbox(lines=5, placeholder="Enter Paper Abstract Here... (Do not input line breaks. No more than 1024 tokens.)", label="Paper Abstract")
],
outputs=gr.Label(label="Predicted Impact"),
title="Predict academic future impact with LLMs",
description="Predict the normalized academic impact (0-1) of a paper based on its title and abstract. !!!Important!!! Please note that the predicted impact is a probabilistic value generated by the model and does not accurately reflect the article's future citation performance. It should not be associated with writing quality, novelty, or other attributes. The author assumes no responsibility for the predictive metrics."
)
iface.launch()
|