<Midm ๋ชจ๋ธ์„ nsmc ๋ฐ์ดํ„ฐ์…‹์„ ํ•ด๊ฒฐํ•˜๋Š” ๋ชจ๋ธ์ด ๋˜๋„๋ก ๋ฏธ์„ธํŠœ๋‹ ํ•˜๊ธฐ>

๋ชจ๋ธ: Midm
๋ฐ์ดํ„ฐ์…‹: nsmc
https://huggingface.co/datasets/nsmc
Train ๋ฐ์ดํ„ฐ: 3000
Test ๋ฐ์ดํ„ฐ: 1000

[ํ…Œ์ŠคํŠธ ๊ฒฐ๊ณผ]

์ •ํ™•๋„: 89.00%

ํ˜ผ๋™ํ–‰๋ ฌ(Confusion Matrix)

์ •๋‹ต Positive ์ •๋‹ต Negative
์˜ˆ์ธก Positive 474 76
์˜ˆ์ธก Negative 34 416

ํ‰๊ฐ€์ง€ํ‘œ

์ •๋ฐ€๋„(Precision) 0.862
์žฌํ˜„์œจ(Recall) 0.933
F1 Score 0.927

[์„ฑ๋Šฅ ํ–ฅ์ƒ]

train ๋ฐ์ดํ„ฐ ์ˆ˜๋ฅผ 2000์—์„œ 2500, 3000์œผ๋กœ ๋Š˜๋ ค๊ฐ€๋ฉฐ ์„ฑ๋Šฅ์„ ์•ฝ 8% ์ •๋„ ๋†’์˜€์œผ๋ฉฐ, TrainingArguments์˜ max_steps ๋“ฑ์˜ ํŒŒ๋ผ๋ฏธํ„ฐ๋ฅผ ์กฐ์ ˆํ•ด๊ฐ€๋ฉฐ ์„ฑ๋Šฅ์„ ๋†’์ด๊ณ ์ž ๋…ธ๋ ฅํ•˜์˜€๋‹ค.


lora-midm-7b-food-order-understanding

This model is a fine-tuned version of KT-AI/midm-bitext-S-7B-inst-v1 on an unknown dataset.

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 2
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.03
  • training_steps: 300
  • mixed_precision_training: Native AMP

Training results

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu118
  • Datasets 2.15.0
  • Tokenizers 0.15.0
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Model tree for ChloeKa/lora-midm-7b-food-order-understanding

Finetuned
(11)
this model