Model Details
- Model Description: This model is test for data ordering.
- Developed by: Juhwan Lee
- Model Type: Large Language Model
Model Architecture
This model is based on Mistral-7B-v0.1. We fine-tuning this model for data ordering task.
Mistral-7B-v0.1 is a transformer model, with the following architecture choices:
- Grouped-Query Attention
- Sliding-Window Attention
- Byte-fallback BPE tokenizer
Dataset
We random sample SlimOrca dataset.
Guthub
License
Apache License 2.0
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 60.44 |
AI2 Reasoning Challenge (25-Shot) | 60.32 |
HellaSwag (10-Shot) | 82.92 |
MMLU (5-Shot) | 62.30 |
TruthfulQA (0-shot) | 45.47 |
Winogrande (5-shot) | 78.06 |
GSM8k (5-shot) | 33.59 |
Dataset used to train NLUHOPOE/experiment2-cause-non-qLoRa
Evaluation results
- normalized accuracy on AI2 Reasoning Challenge (25-Shot)test set Open LLM Leaderboard60.320
- normalized accuracy on HellaSwag (10-Shot)validation set Open LLM Leaderboard82.920
- accuracy on MMLU (5-Shot)test set Open LLM Leaderboard62.300
- mc2 on TruthfulQA (0-shot)validation set Open LLM Leaderboard45.470
- accuracy on Winogrande (5-shot)validation set Open LLM Leaderboard78.060
- accuracy on GSM8k (5-shot)test set Open LLM Leaderboard33.590