ravirajoshi
commited on
Commit
•
8fcd1d9
1
Parent(s):
b2efad2
Update README.md
Browse files
README.md
CHANGED
@@ -14,6 +14,7 @@ library_name: nemo
|
|
14 |
## Model Overview
|
15 |
|
16 |
Nemotron-4-Mini-Hindi-4B-Instruct is a model for generating responses to questions grounded in the Indian context. It supports queries in Hindi, English, and Hinglish. It is a small language model (SLM) which is fine-tuned/aligned version of [nvidia/Nemotron-4-Mini-Hindi-4B-Base](https://huggingface.co/nvidia/Nemotron-4-Mini-Hindi-4B-Base), which was continuously pre-trained on top of Nemotron-Mini-4B-Base (Minitron-4B-Base). The alignment stage uses a mix of real and synthetically generated alignment corpus. It supports a context length of 4,096 tokens. This model is ready for commercial use.
|
|
|
17 |
|
18 |
Try this model on [build.nvidia.com](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct).
|
19 |
|
@@ -146,5 +147,15 @@ The model was trained on data that contains toxic language and societal biases o
|
|
146 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the [Model Card++](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct/modelcard). Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|
147 |
|
148 |
|
|
|
149 |
|
150 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
## Model Overview
|
15 |
|
16 |
Nemotron-4-Mini-Hindi-4B-Instruct is a model for generating responses to questions grounded in the Indian context. It supports queries in Hindi, English, and Hinglish. It is a small language model (SLM) which is fine-tuned/aligned version of [nvidia/Nemotron-4-Mini-Hindi-4B-Base](https://huggingface.co/nvidia/Nemotron-4-Mini-Hindi-4B-Base), which was continuously pre-trained on top of Nemotron-Mini-4B-Base (Minitron-4B-Base). The alignment stage uses a mix of real and synthetically generated alignment corpus. It supports a context length of 4,096 tokens. This model is ready for commercial use.
|
17 |
+
Please refer to our [arXiv paper](https://arxiv.org/abs/2410.14815) for more details.
|
18 |
|
19 |
Try this model on [build.nvidia.com](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct).
|
20 |
|
|
|
147 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the [Model Card++](https://build.nvidia.com/nvidia/nemotron-4-mini-hindi-4b-instruct/modelcard). Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|
148 |
|
149 |
|
150 |
+
## Citation
|
151 |
|
152 |
+
If you find our work helpful, please consider citing our paper:
|
153 |
+
```
|
154 |
+
@article{hindiminitron2024,
|
155 |
+
title={Adapting Multilingual LLMs to Low-Resource Languages using Continued Pre-training and Synthetic Corpus},
|
156 |
+
author={Raviraj Joshi and Kanishk Singla and Anusha Kamath and Raunak Kalani and Rakesh Paul and Utkarsh Vaidya and Sanjay Singh Chauhan and Niranjan Wartikar and Eileen Long},
|
157 |
+
journal={arXiv preprint arXiv:2410.14815},
|
158 |
+
year={2024},
|
159 |
+
url={https://arxiv.org/abs/2410.14815},
|
160 |
+
}
|
161 |
+
```
|