Update README.md
Browse files
README.md
CHANGED
@@ -3,7 +3,7 @@ license: apache-2.0
|
|
3 |
language:
|
4 |
- en
|
5 |
base_model:
|
6 |
-
- unsloth/Qwen2.5-72B-Instruct
|
7 |
pipeline_tag: text-generation
|
8 |
library_name: transformers
|
9 |
tags:
|
@@ -21,7 +21,7 @@ tags:
|
|
21 |
This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
|
22 |
|
23 |
### **Key Features**
|
24 |
-
- **Base Model:** `unsloth/qwen2.5-72b-instruct
|
25 |
- **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
|
26 |
- **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
|
27 |
- **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
|
@@ -53,7 +53,7 @@ This model is released under the **Apache-2.0 License**, which allows for free u
|
|
53 |
## **Model Training**
|
54 |
|
55 |
### **Base Model**
|
56 |
-
The model is derived from the `unsloth/qwen2.5-72b-instruct
|
57 |
|
58 |
### **Training Process**
|
59 |
- **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
|
|
|
3 |
language:
|
4 |
- en
|
5 |
base_model:
|
6 |
+
- unsloth/Qwen2.5-72B-Instruct
|
7 |
pipeline_tag: text-generation
|
8 |
library_name: transformers
|
9 |
tags:
|
|
|
21 |
This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
|
22 |
|
23 |
### **Key Features**
|
24 |
+
- **Base Model:** `unsloth/qwen2.5-72b-instruct`
|
25 |
- **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
|
26 |
- **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
|
27 |
- **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
|
|
|
53 |
## **Model Training**
|
54 |
|
55 |
### **Base Model**
|
56 |
+
The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
|
57 |
|
58 |
### **Training Process**
|
59 |
- **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
|