Update app.py
Browse files
app.py
CHANGED
@@ -254,6 +254,16 @@ def text_to_kenlm(
|
|
254 |
line = line.lower()
|
255 |
results.append(line)
|
256 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
257 |
generate_files(results)
|
258 |
|
259 |
# Write to intermediate file
|
@@ -314,6 +324,7 @@ def text_to_kenlm(
|
|
314 |
if r.returncode != 0:
|
315 |
raise gr.Error("Failed to filter the model.")
|
316 |
|
|
|
317 |
generate_files(vocab_str.split("\n"))
|
318 |
|
319 |
if _do_quantize:
|
|
|
254 |
line = line.lower()
|
255 |
results.append(line)
|
256 |
|
257 |
+
# Remove previous files
|
258 |
+
for file in [
|
259 |
+
"/tmp/intermediate.txt", "/tmp/my_model.arpa", "/tmp/my_model-trie.bin", "/tmp/my_model_correct.arpa",
|
260 |
+
"/tmp/my_model-trie-10000-words.arpa", "/tmp/my_model-trie-10000-words.bin",
|
261 |
+
"/tmp/model_vocab.txt", "/tmp/model_lexicon.txt", "/tmp/model_tokens.txt",
|
262 |
+
]:
|
263 |
+
if os.path.exists(file):
|
264 |
+
os.remove(file)
|
265 |
+
|
266 |
+
# Generate files: vocab, lexicon, tokens
|
267 |
generate_files(results)
|
268 |
|
269 |
# Write to intermediate file
|
|
|
324 |
if r.returncode != 0:
|
325 |
raise gr.Error("Failed to filter the model.")
|
326 |
|
327 |
+
# Regenerate files: vocab, lexicon, tokens
|
328 |
generate_files(vocab_str.split("\n"))
|
329 |
|
330 |
if _do_quantize:
|