Daemontatox commited on
Commit
5b57128
·
verified ·
1 Parent(s): e061542

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -3,7 +3,7 @@ license: apache-2.0
3
  language:
4
  - en
5
  base_model:
6
- - unsloth/Qwen2.5-72B-Instruct-bnb-4bit
7
  pipeline_tag: text-generation
8
  library_name: transformers
9
  tags:
@@ -21,7 +21,7 @@ tags:
21
  This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
22
 
23
  ### **Key Features**
24
- - **Base Model:** `unsloth/qwen2.5-72b-instruct-bnb-4bit`
25
  - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
26
  - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
27
  - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
@@ -53,7 +53,7 @@ This model is released under the **Apache-2.0 License**, which allows for free u
53
  ## **Model Training**
54
 
55
  ### **Base Model**
56
- The model is derived from the `unsloth/qwen2.5-72b-instruct-bnb-4bit`, a quantized version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
57
 
58
  ### **Training Process**
59
  - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
 
3
  language:
4
  - en
5
  base_model:
6
+ - unsloth/Qwen2.5-72B-Instruct
7
  pipeline_tag: text-generation
8
  library_name: transformers
9
  tags:
 
21
  This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
22
 
23
  ### **Key Features**
24
+ - **Base Model:** `unsloth/qwen2.5-72b-instruct`
25
  - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
26
  - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
27
  - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
 
53
  ## **Model Training**
54
 
55
  ### **Base Model**
56
+ The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
57
 
58
  ### **Training Process**
59
  - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.