titulm-mpt-1b-v1.0 / README.md
sagorsarker's picture
Update README.md
0c58c9f verified
|
raw
history blame
762 Bytes
metadata
language:
  - bn
license: apache-2.0
datasets:
  - uonlp/CulturaX
pipeline_tag: text-generation

TituLM-1B-BN-V1

TituLM-1B-BN-V1 is a large language model specifically trained for generating and understanding Bangla text. Utilizing a decoder-style transformer architecture, this model has been extensively trained on a dataset comprising 4.51 billion Bangla tokens. This model is the part of iterative train and release Bangla LLM from Hishab.

Training

The training process was managed using the robust framework provided by MosaicML's llm-foundry repository. Throughout the training phase, titulm-1b-bn-v1 underwent a total of 59 iterations, allowing for iterative refinements and optimization

Datasets