Papers
arxiv:2405.01873

Enhancing Bangla Language Next Word Prediction and Sentence Completion through Extended RNN with Bi-LSTM Model On N-gram Language

Published on May 3, 2024
Authors:
,
,

Abstract

Texting stands out as the most prominent form of communication worldwide. Individual spend significant amount of time writing whole texts to send emails or write something on social media, which is time consuming in this modern era. Word prediction and sentence completion will be suitable and appropriate in the Bangla language to make textual information easier and more convenient. This paper expands the scope of Bangla language processing by introducing a Bi-LSTM model that effectively handles Bangla next-<PRE_TAG><PRE_TAG>word prediction</POST_TAG></POST_TAG> and Bangla sentence generation, demonstrating its versatility and potential impact. We proposed a new <PRE_TAG>Bi-LSTM model</POST_TAG> to predict a following word and complete a sentence. We constructed a <PRE_TAG>corpus dataset</POST_TAG> from various news portals, including bdnews24, BBC News Bangla, and Prothom Alo. The proposed approach achieved superior results in <PRE_TAG>word prediction</POST_TAG>, reaching 99\% accuracy for both 4-gram and 5-gram word predictions. Moreover, it demonstrated significant improvement over existing methods, achieving 35\%, 75\%, and 95\% accuracy for uni-gram, bi-gram, and tri-gram <PRE_TAG>word prediction</POST_TAG>, respectively

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2405.01873 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2405.01873 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2405.01873 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.