kevin009's picture
Trained with Unsloth
8914bc5 verified
---
tags:
- unsloth
---
# Llama 3.1 Merged Adapters
## Model Description
This is a merged model combining multiple fine-tuned LoRA adapters using TIES (Task Inference with Expert Selection) merging technique. The model combines the strengths of multiple specialized adapters into a single model.
## Base Model
- [unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit](https://huggingface.co/unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit)
## Merged Adapters
The following adapters were merged:
- llama337 - Specialized for creative writing
- llama338 - Specialized for logical reasoning
- llama340 - Specialized for code generation
- llama344 - Specialized for instruction following
- llama345 - Specialized for factual knowledge
- llama346 - Specialized for conversational abilities
- llama349 - Specialized for problem solving
- llama350 - Specialized for structured output
## Merging Parameters
- **Merging Method**: TIES (Task Inference with Expert Selection)
- **Density**: 0.2 (controls parameter sparsity)
- **Weights**: Equal weighting (1.0 for each adapter)
- **Merge Date**: 2025-03-09
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the model and tokenizer
model = AutoModelForCausalLM.from_pretrained("kevin009/llama3-merged-adapters")
tokenizer = AutoTokenizer.from_pretrained("kevin009/llama3-merged-adapters")
# Example usage
prompt = "Write a short story about a robot learning to paint."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Model Performance
This merged model combines the capabilities of multiple specialized adapters, resulting in improved performance across a variety of tasks compared to individual adapters.
## Limitations
- The model inherits limitations from the base Llama 3.1 model
- May produce inconsistent outputs for certain edge cases
- As with all language models, can produce incorrect or misleading information
## License
This model is subject to the license of the original Llama 3.1 model.