Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -18,6 +18,7 @@ nltk.download('dolch') #Sight words
|
|
18 |
nltk.download('names') #People Names NER
|
19 |
nltk.download('gazetteers') #Location NER
|
20 |
nltk.download('opinion_lexicon') #Sentiment words
|
|
|
21 |
|
22 |
|
23 |
spacy.cli.download("en_core_web_sm")
|
@@ -120,11 +121,6 @@ def filter_words(words):
|
|
120 |
|
121 |
return filtered_words
|
122 |
|
123 |
-
# Call the function to get the filtered list of words
|
124 |
-
filtered_words = filter_words(words)
|
125 |
-
|
126 |
-
print(filtered_words)
|
127 |
-
|
128 |
def SepHypandSynExpansion(text):
|
129 |
# Tokenize the text
|
130 |
tokens = nltk.word_tokenize(text)
|
|
|
18 |
nltk.download('names') #People Names NER
|
19 |
nltk.download('gazetteers') #Location NER
|
20 |
nltk.download('opinion_lexicon') #Sentiment words
|
21 |
+
nltk.download('averaged_perceptron_tagger') #Parts of Speech Tagging
|
22 |
|
23 |
|
24 |
spacy.cli.download("en_core_web_sm")
|
|
|
121 |
|
122 |
return filtered_words
|
123 |
|
|
|
|
|
|
|
|
|
|
|
124 |
def SepHypandSynExpansion(text):
|
125 |
# Tokenize the text
|
126 |
tokens = nltk.word_tokenize(text)
|