l3cube-pune's picture
Create README.md
8b0de1d verified
|
raw
history blame
555 Bytes
metadata
license: cc-by-4.0
language: mr

MahaGemma

MahaGemma is a Marathi Gemma model. It is a Gemma 7B (google/gemma-7b) model LoRA fine-tuned on a translated Marathi instruction tuning datasets. [dataset link] (https://github.com/l3cube-pune/MarathiNLP)

This is part of the MahaNLP initiative. More details coming soon.

Citing

@article{joshi2022l3cube,
  title={L3cube-mahanlp: Marathi natural language processing datasets, models, and library},
  author={Joshi, Raviraj},
  journal={arXiv preprint arXiv:2205.14728},
  year={2022}
}