Model Card for Rashik24/Mistral-Instruct-Bangla

Blog post: https://blog.rashik.sh/mistral-instruct-bangla-bridging-the-gap-in-bengali-ai

The Rashik24/Mistral-Instruct-Bangla model is a language model specifically tailored for the Bengali language. Based on the Mistralai/Mistral-7B-Instruct-v0.2 base model, it has been fine-tuned using the iamshnoo/alpaca-cleaned-bengali dataset. This model is designed to understand and generate Bengali text, making it a valuable tool for a variety of natural language processing tasks in the Bengali language context.

Uses

The Mistral-Instruct-Bangla model is intended for a range of applications where understanding and generating Bengali text is crucial. This includes but is not limited to machine translation, content creation, sentiment analysis, and language understanding tasks in Bengali. The model is suited for both academic researchers and industry practitioners who are working on Bengali language processing.

Direct Use

This model can be directly used for generating Bengali text, understanding Bengali context in conversations, and translating between Bengali and other languages. It is designed to be straightforward to implement in various software environments, requiring minimal additional setup for direct use cases.

How to Get Started with the Model

To start using the Rashik24/Mistral-Instruct-Bangla model, you can use the following code as a basic guide. This will help you integrate the model into your application or research project.

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
from peft import PeftModel, PeftConfig
from transformers import AutoModelForCausalLM


def load_model(model_name):
    tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", trust_remote_code=True)
    tokenizer.pad_token = tokenizer.eos_token
    tokenizer.padding_side = "right"
    config = PeftConfig.from_pretrained("Rashik24/Mistral-Instruct-Bangla")
    model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2").to("cuda")
    model = PeftModel.from_pretrained(model, "Rashik24/Mistral-Instruct-Bangla").to("cuda")
    model.eval()
    return model, tokenizer

def generate_text(prompt, model, tokenizer):
    inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
    with torch.no_grad():
        generated_code = tokenizer.decode(model.generate(**inputs, max_new_tokens=256, pad_token_id=2)[0], skip_special_tokens=True)
    print(generated_code)
    return generated_code

#Load the model
model_name = 'Rashik24/Mistral-Instruct-Bangla'
model, tokenizer = load_model(model_name)

prompt = "একটি গ্রামের বর্ণনা করুন।"
generated_text = generate_text(prompt, model, tokenizer)

Training Details

Training Data

The model has been trained on the 'iamshnoo/alpaca-cleaned-bengali' dataset.

For more details on the training data and methodology, refer to the dataset card linked here:https://huggingface.co/datasets/iamshnoo/alpaca-cleaned-bengali

Downloads last month
2
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for Rashik24/Mistral-Instruct-Bangla

Adapter
(886)
this model

Dataset used to train Rashik24/Mistral-Instruct-Bangla

Space using Rashik24/Mistral-Instruct-Bangla 1