--- license: cc-by-4.0 language: mr --- ## MahaGemma MahaGemma is a Marathi Gemma model. It is a Gemma 7B (google/gemma-7b) model LoRA fine-tuned on a translated Marathi instruction tuning datasets. [dataset link] (https://github.com/l3cube-pune/MarathiNLP) This is part of the MahaNLP initiative. More details coming soon.
Citing ``` @article{joshi2022l3cube, title={L3cube-mahanlp: Marathi natural language processing datasets, models, and library}, author={Joshi, Raviraj}, journal={arXiv preprint arXiv:2205.14728}, year={2022} } ```