NeMo
PyTorch
English
Hindi
nemotron
ravirajoshi commited on
Commit
8fcd1d9
1 Parent(s): b2efad2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -14,6 +14,7 @@ library_name: nemo
14
  ## Model Overview
15
 
16
  Nemotron-4-Mini-Hindi-4B-Instruct is a model for generating responses to questions grounded in the Indian context. It supports queries in Hindi, English, and Hinglish. It is a small language model (SLM) which is fine-tuned/aligned version of [nvidia/Nemotron-4-Mini-Hindi-4B-Base](https://huggingface.co/nvidia/Nemotron-4-Mini-Hindi-4B-Base), which was continuously pre-trained on top of Nemotron-Mini-4B-Base (Minitron-4B-Base). The alignment stage uses a mix of real and synthetically generated alignment corpus. It supports a context length of 4,096 tokens. This model is ready for commercial use.
 
17
 
18
  Try this model on [build.nvidia.com](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct).
19
 
@@ -146,5 +147,15 @@ The model was trained on data that contains toxic language and societal biases o
146
  NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the [Model Card++](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct/modelcard). Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
147
 
148
 
 
149
 
150
-
 
 
 
 
 
 
 
 
 
 
14
  ## Model Overview
15
 
16
  Nemotron-4-Mini-Hindi-4B-Instruct is a model for generating responses to questions grounded in the Indian context. It supports queries in Hindi, English, and Hinglish. It is a small language model (SLM) which is fine-tuned/aligned version of [nvidia/Nemotron-4-Mini-Hindi-4B-Base](https://huggingface.co/nvidia/Nemotron-4-Mini-Hindi-4B-Base), which was continuously pre-trained on top of Nemotron-Mini-4B-Base (Minitron-4B-Base). The alignment stage uses a mix of real and synthetically generated alignment corpus. It supports a context length of 4,096 tokens. This model is ready for commercial use.
17
+ Please refer to our [arXiv paper](https://arxiv.org/abs/2410.14815) for more details.
18
 
19
  Try this model on [build.nvidia.com](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct).
20
 
 
147
  NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the [Model Card++](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct/modelcard). Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
148
 
149
 
150
+ ## Citation
151
 
152
+ If you find our work helpful, please consider citing our paper:
153
+ ```
154
+ @article{hindiminitron2024,
155
+ title={Adapting Multilingual LLMs to Low-Resource Languages using Continued Pre-training and Synthetic Corpus},
156
+ author={Raviraj Joshi and Kanishk Singla and Anusha Kamath and Raunak Kalani and Rakesh Paul and Utkarsh Vaidya and Sanjay Singh Chauhan and Niranjan Wartikar and Eileen Long},
157
+ journal={arXiv preprint arXiv:2410.14815},
158
+ year={2024},
159
+ url={https://arxiv.org/abs/2410.14815},
160
+ }
161
+ ```