Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -19,7 +19,8 @@ from transformers import pipeline
|
|
19 |
|
20 |
# from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
21 |
|
22 |
-
from transformers import
|
|
|
23 |
|
24 |
def search_related_articles_crossref(query, max_results=3):
|
25 |
"""Search for related articles using CrossRef API."""
|
@@ -64,8 +65,9 @@ def extract_text_from_html(url):
|
|
64 |
except Exception as e:
|
65 |
return f"Error extracting text: {str(e)}"
|
66 |
|
67 |
-
|
68 |
-
tokenizer =
|
|
|
69 |
|
70 |
def summarize_article(article_text):
|
71 |
"""Summarize a given article's text."""
|
|
|
19 |
|
20 |
# from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
21 |
|
22 |
+
from transformers import T5Tokenizer, T5ForConditionalGeneration
|
23 |
+
|
24 |
|
25 |
def search_related_articles_crossref(query, max_results=3):
|
26 |
"""Search for related articles using CrossRef API."""
|
|
|
65 |
except Exception as e:
|
66 |
return f"Error extracting text: {str(e)}"
|
67 |
|
68 |
+
|
69 |
+
tokenizer = T5Tokenizer.from_pretrained("google-t5/t5-small")
|
70 |
+
model = T5ForConditionalGeneration.from_pretrained("google-t5/t5-small")
|
71 |
|
72 |
def summarize_article(article_text):
|
73 |
"""Summarize a given article's text."""
|