Base model: CorticalStack/gemma-7b-ultrachat-sft

This is finetuned from above base model and to be used for multi-turn chat based use-cases. Unlike our AryaBhatta-GemmaOrca model which is skilled in science, literature and finetuned on Orca datasets, this model is fine-tuned on Ultra-Chat datasets. And show improved performance over AryaBhatta-GemmaOrca on Hellaswag datasets and in multi-turn conversations. It is finetuned on 9 Indian languages (Hindi, Tamil, Punjabi, Bengali, Gujarati, Oriya, Telugu, Kannada, Malayalam) plus English.

Benchmarked on Indic LLM leaderboard: https://huggingface.co/spaces/Cognitive-Lab/indic_llm_leaderboard

Release post: https://www.linkedin.com/feed/update/urn:li:activity:7184856055565180928

Downloads last month
29
Safetensors
Model size
8.54B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.