File size: 473 Bytes
6311882 f3b764c 392acc9 6311882 392acc9 6311882 392acc9 6311882 392acc9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 |
---
library_name: transformers
pipeline_tag: text-generation
license: apache-2.0
datasets:
- euclaise/reddit-instruct
language:
- en
---
# Model Card for Model ID
This is the fine-tuned model of Mistral-7B-v0.1 (https://huggingface.co/mistralai/Mistral-7B-v0.1) with reddit instruct dataset
Model Architecture
Mistral-7B-v0.1 is a transformer model, with the following architecture choices:
Grouped-Query Attention
Sliding-Window Attention
Byte-fallback BPE tokenizer
|