import streamlit as st import pandas as pd import spacy from transformers import pipeline, AutoModelForTokenClassification, AutoTokenizer import PyPDF2 import docx import io st.set_page_config(layout="wide") # Function to read text from uploaded file def read_file(file): if file.type == "text/plain": return file.getvalue().decode("utf-8") elif file.type == "application/pdf": pdf_reader = PyPDF2.PdfReader(io.BytesIO(file.getvalue())) return " ".join(page.extract_text() for page in pdf_reader.pages) elif file.type == "application/vnd.openxmlformats-officedocument.wordprocessingml.document": doc = docx.Document(io.BytesIO(file.getvalue())) return " ".join(paragraph.text for paragraph in doc.paragraphs) else: st.error("Unsupported file type") return None # Rest of your code remains the same example_list = [ "Mustafa Kemal Atatürk 1919 yılında Samsun'a çıktı.", """Mustafa Kemal Atatürk, Türk asker, devlet adamı ve Türkiye Cumhuriyeti'nin kurucusudur. # ... (rest of the example text) """ ] st.title("Demo for Turkish NER Models") model_list = [ 'akdeniz27/bert-base-turkish-cased-ner', 'akdeniz27/convbert-base-turkish-cased-ner', 'girayyagmur/bert-base-turkish-ner-cased', 'FacebookAI/xlm-roberta-large', 'savasy/bert-base-turkish-ner-cased', 'xlm-roberta-large-finetuned-conll03-english', 'asahi417/tner-xlm-roberta-base-ontonotes5' ] st.sidebar.header("Select NER Model") model_checkpoint = st.sidebar.radio("", model_list) st.sidebar.write("For details of models: 'https://huggingface.co/akdeniz27/") st.sidebar.write("") if model_checkpoint in ["akdeniz27/xlm-roberta-base-turkish-ner", "xlm-roberta-large-finetuned-conll03-english", "asahi417/tner-xlm-roberta-base-ontonotes5"]: aggregation = "simple" if model_checkpoint != "akdeniz27/xlm-roberta-base-turkish-ner": st.sidebar.write("The selected NER model is included just to show the zero-shot transfer learning capability of XLM-Roberta pretrained language model.") else: aggregation = "first" st.subheader("Select Text Input Method") input_method = st.radio("", ('Select from Examples', 'Write or Paste New Text', 'Upload File')) if input_method == 'Select from Examples': selected_text = st.selectbox('Select Text from List', example_list, index=0, key=1) input_text = st.text_area("Selected Text", selected_text, height=128, max_chars=None, key=2) elif input_method == "Write or Paste New Text": input_text = st.text_area('Write or Paste Text Below', value="", height=128, max_chars=None, key=2) else: uploaded_file = st.file_uploader("Choose a file", type=["txt", "pdf", "docx"]) if uploaded_file is not None: input_text = read_file(uploaded_file) if input_text: st.text_area("Extracted Text", input_text, height=128, max_chars=None, key=2) else: input_text = "" # Rest of your functions (setModel, get_html, entity_comb) remain the same @st.cache_resource def setModel(model_checkpoint, aggregation): model = AutoModelForTokenClassification.from_pretrained(model_checkpoint) tokenizer = AutoTokenizer.from_pretrained(model_checkpoint) return pipeline('ner', model=model, tokenizer=tokenizer, aggregation_strategy=aggregation) @st.cache_resource def get_html(html: str): WRAPPER = """