Vineedhar commited on
Commit
89796f5
·
verified ·
1 Parent(s): 568454a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -4
app.py CHANGED
@@ -10,10 +10,14 @@ from transformers import AutoTokenizer, AutoModelForSequenceClassification, pipe
10
  import tensorflow as tf
11
  tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR)
12
 
13
- # LICESNE.tensorflow - https://github.com/tensorflow/tensorflow/blob/master/LICENSE
14
- # LICENSE.Roberta -https://github.com/cardiffnlp/xlm-t/blob/main/LICENSE
15
- # LICENSE.streamlit - https://github.com/streamlit/streamlit/blob/develop/LICENSE
16
- # LICENSE.tranformers - https://github.com/huggingface/huggingface_hub/blob/main/LICENSE
 
 
 
 
17
 
18
  # Load pre-trained model and tokenizer
19
  saved_directory = "orYx-models/finetuned-roberta-leadership-sentiment-analysis"
 
10
  import tensorflow as tf
11
  tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR)
12
 
13
+ # LICENSE.streamlit.Apachev2 - Copyright (c) Streamlit Inc. (2018-2022) Snowflake Inc. (2022-2024) (https://github.com/streamlit/streamlit/blob/develop/LICENSE)
14
+
15
+ # LICENSE.tranformers.Apachev2 - Copyright 2020 The HuggingFace Team. All rights reserved. (https://github.com/huggingface/huggingface_hub/blob/main/LICENSE)
16
+
17
+ # LICESNE.tensorflow.Apachev2 - Copyright 2015 The TensorFlow Authors. All Rights Reserved. (https://github.com/tensorflow/tensorflow/blob/master/LICENSE)
18
+
19
+ # LICENSE.Roberta.Apachev2 - @InProceedings{barbieri-espinosaanke-camachocollados:2022:LREC,author = {Barbieri, Francesco and Espinosa Anke, Luis and Camacho-Collados, Jose}, title = {XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond}, booktitle = {Proceedings of the Language Resources and Evaluation Conference}, month = {June}, year = {2022}, address = {Marseille, France}, publisher = {European Language Resources Association}, pages = {258--266}, abstract = {Language models are ubiquitous in current NLP, and their multilingual capacity has recently attracted considerable attention. However, current analyses have almost exclusively focused on (multilingual variants of) standard benchmarks, and have relied on clean pre-training and task-specific corpora as multilingual signals. In this paper, we introduce XLM-T, a model to train and evaluate multilingual language models in Twitter. In this paper we provide: (1) a new strong multilingual baseline consisting of an XLM-R (Conneau et al. 2020) model pre-trained on millions of tweets in over thirty languages, alongside starter code to subsequently fine-tune on a target task; and (2) a set of unified sentiment analysis Twitter datasets in eight different languages and a XLM-T model trained on this dataset.}, url = {https://aclanthology.org/2022.lrec-1.27}} (https://github.com/cardiffnlp/xlm-t/blob/main/LICENSE)
20
+
21
 
22
  # Load pre-trained model and tokenizer
23
  saved_directory = "orYx-models/finetuned-roberta-leadership-sentiment-analysis"