|
--- |
|
library_name: transformers |
|
license: apache-2.0 |
|
pipeline_tag: text-generation |
|
datasets: |
|
- maywell/ko_Ultrafeedback_binarized |
|
base model: |
|
- yanolja/EEVE-Korean-Instruct-10.8B-v1.0 |
|
--- |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65f22e4076fedc4fd11e978f/MoTedec_ZL8GM2MmGyAPs.png) |
|
|
|
|
|
|
|
|
|
# T3Q-LLM-MG-v1.0 |
|
|
|
## This model is a version of T3Q-LLM/T3Q-LLM-solar10.8-sft-v1.0 that has been fine-tuned with DPO. |
|
|
|
## Model Developers Chihoon Lee(chihoonlee10), T3Q |
|
|
|
### Python code |
|
```python |
|
import torch |
|
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer |
|
|
|
MODEL_DIR = "chihoonlee10/T3Q-LLM-MG-v1.0" |
|
model = AutoModelForCausalLM.from_pretrained(MODEL_DIR, torch_dtype=torch.float16).to("cuda") |
|
tokenizer = AutoTokenizer.from_pretrained(MODEL_DIR) |
|
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) |
|
|
|
s = "한국의 수도는 어디?" |
|
conversation = [{'role': 'user', 'content': s}] |
|
inputs = tokenizer.apply_chat_template( |
|
conversation, |
|
tokenize=True, |
|
add_generation_prompt=True, |
|
return_tensors='pt').to("cuda") |
|
_ = model.generate(inputs, streamer=streamer, max_new_tokens=1024) |
|
|
|
``` |
|
hf (pretrained=chihoonlee10/T3Q-LLM-MG-v1.0), limit: None, provide_description: False, num_fewshot: 0, batch_size: None |
|
|
|
| Task |Version| Metric |Value | |Stderr| |
|
|----------------|------:|--------|-----:|---|-----:| |
|
|kobest_boolq | 0|acc |0.9523|± |0.0057| |
|
| | |macro_f1|0.9523|± |0.0057| |
|
|kobest_copa | 0|acc |0.7740|± |0.0132| |
|
| | |macro_f1|0.7737|± |0.0133| |
|
|kobest_hellaswag| 0|acc |0.4980|± |0.0224| |
|
| | |acc_norm|0.5920|± |0.0220| |
|
| | |macro_f1|0.4950|± |0.0223| |
|
|kobest_sentineg | 0|acc |0.7254|± |0.0224| |
|
| | |macro_f1|0.7106|± |0.0234| |
|
|
|
|
|
|
|
### T3Q-LLM/T3Q-LLM-sft1.0-dpo1.0 |
|
|
|
| Task |Version| Metric |Value | |Stderr| |
|
|----------------|------:|--------|-----:|---|-----:| |
|
|kobest_boolq | 0|acc |0.9387|± |0.0064| |
|
| | |macro_f1|0.9387|± |0.0064| |
|
|kobest_copa | 0|acc |0.7590|± |0.0135| |
|
| | |macro_f1|0.7585|± |0.0135| |
|
|kobest_hellaswag| 0|acc |0.5080|± |0.0224| |
|
| | |acc_norm|0.5580|± |0.0222| |
|
| | |macro_f1|0.5049|± |0.0224| |
|
|kobest_sentineg | 0|acc |0.8489|± |0.0180| |
|
| | |macro_f1|0.8483|± |0.0180| |
|
|