File size: 3,731 Bytes
7c6ce59 866b4c6 7c6ce59 7c4ee5d 7c6ce59 7c4ee5d 5026185 7c6ce59 5026185 5b57128 5026185 af3c521 5026185 af3c521 5026185 5b57128 5026185 5dff8db 5026185 af3c521 5026185 af3c521 5026185 1042431 5026185 7c6ce59 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 |
---
license: apache-2.0
language:
- en
base_model:
- Qwen/Qwen2.5-72B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- reasoning
- logic
- cot
- text-generation-inference
new_version: Daemontatox/Cogito-Maximus
---

## **Model Overview**
This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
### **Key Features**
- **Base Model:** `unsloth/qwen2.5-72b-instruct`
- **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
- **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
- **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
- **License:** Apache-2.0
---
## **Model Details**
### **Developed by**
- **Author:** Daemontatox
- **Organization:** Independent Contributor
### **Tags**
- Text Generation Inference
- Transformers
- Unsloth
- Qwen2
- TRL
### **Language**
- English (`en`)
### **License**
This model is released under the **Apache-2.0 License**, which allows for free use, modification, and distribution, provided the original license and copyright notice are included.
---
## **Model Training**
### **Base Model**
The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
### **Training Process**
- **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
- **Acceleration:** Training was completed **2x faster** compared to traditional methods, thanks to Unsloth's optimizations.
- **Reinforcement Learning:** Fine-tuning incorporated techniques from Huggingface's **TRL** library, enabling advanced instruction-tuning and alignment with human preferences.
---
## **Intended Use**
### **Primary Use Case**
This model is designed for **text generation tasks**, including but not limited to:
- Instruction-following
- Question answering
- Content creation
- Dialogue systems
### **Limitations**
- The model is trained primarily on English data and may not perform as well on other languages.
- While fine-tuned for instruction-following, outputs should be reviewed for accuracy and relevance in critical applications.
---
## **How to Use**
### **Installation**
To use this model, ensure you have the following libraries installed:
```bash
pip install transformers torch bitsandbytes unsloth trl
```
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the tokenizer and model
model_name = "Daemontatox/Cogito-Maximus"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_4bit=True)
# Generate text
input_text = "Explain the concept of machine learning in simple terms."
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_length=100)
# Decode and print the output
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
```
@misc{daemontatox_cogito_maximus,
author = {Daemontatox},
title = {Cogito-Maximus: Fine-tuned Qwen2.5-72B Instruct Model},
year = {2025},
publisher = {Hugging Face},
journal = {Hugging Face Model Repository},
howpublished = {\url{https://huggingface.co/Daemontatox/Cogito-Maximus}}
}
``` |