--- base_model: meta-llama/Llama-2-7b-chat-hf tags: - generated_from_trainer model-index: - name: lora-llama-2-7b-nsmc results: [] --- # lora-llama-2-7b-nsmc This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on 'nsmc' dataset. ## Model description [meta-llama/Llama-2-7b-chat-hf]를 nsmc 데이터셋을 이용하여 미세튜닝함. ## Intended uses & limitations 목적: 영화 리뷰 판단 (긍정/부정) ## Training and evaluation data - training data: nsmc의 train 데이터 중 상위 2000개 - evaluation data: nsmc의 test 데이터 중 상위 1000개 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 2 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.03 - training_steps: 1000 - mixed_precision_training: Native AMP ### Training results - global_step=1000 - training_loss=0.7012106285095215 - metrics={'train_runtime': 3567.3706, 'train_samples_per_second': 0.561, 'train_steps_per_second': 0.28, 'total_flos': 4.0718328397824e+16, 'train_loss': 0.7012106285095215, 'epoch': 1.0} ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu118 - Datasets 2.15.0 - Tokenizers 0.15.0 ### Accuracy ||TP|TN| |------|---|---| |PP|436|36| |PN|72|456| - accuracy: 0.892