千尋 7B v0.1

Zebrafish 7B 加上 Breeze 7B 的 slerp merge 試驗性通用繁中基座模型 📚

GGUF Quants 👉 Chihiro-7B-v0.1-GGUF

請用 Mistral 7B Instruct 或是 Breeze 7B Instruct 所推薦的 Prompt 格式進行操作;以下為模型配置。

Chihiro 7B v0.1

This is an experimental Mistral-architecture SLERP merge with two brilliant base models. Zebrafish and Breeze were used together in this work.

Model configuration is as follows:

To use the model, please use either prompt templates suggested by the base models, or just slap the Mistral one on.



Benchmarks

Evaluation suite: OpenLLM

Model ARC HellaSwag MMLU TruthfulQA Winogrande GSM8K
Chihiro-7B-v0.1 68.52 85.95 (not yet evaluated) 63.81 81.77 64.22

Evaluation suite: Nous

Model AGIEval GPT4All TruthfulQA Bigbench Average
Chihiro-7B-v0.1 45.16 75.26 63.82 47.38 57.91

Average: 47.38%

Average score: 57.91%

Evaluated Apr. 27, 2024, NVIDIA RTX 4090



Downloads last month
17
Safetensors
Model size
7.49B params
Tensor type
F32
·
BF16
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for yuuko-eth/Chihiro-7B-v0.1

Quantizations
1 model