stash commited on
Commit
ea3765e
1 Parent(s): 84ad8c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -7,7 +7,41 @@ library_name: peft
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  ## Model Details
13
 
 
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ echo "Starting LoRA fine-tuning..."
11
+ deepspeed LLaVA/llava/train/train_mem.py \
12
+ --lora_enable True --lora_r 128 --lora_alpha 256 --mm_projector_lr 2e-5 \
13
+ --deepspeed ./LLaVA/scripts/zero3.json \
14
+ --model_name_or_path liuhaotian/llava-v1.5-7b \
15
+ --version v1 \
16
+ --data_path "${output_dir}/processed_dataset_correct_path.json" \
17
+ --image_folder "${image_dir}" \
18
+ --vision_tower openai/clip-vit-large-patch14-336 \
19
+ --mm_projector_type mlp2x_gelu \
20
+ --mm_vision_select_layer -2 \
21
+ --mm_use_im_start_end False \
22
+ --mm_use_im_patch_token False \
23
+ --image_aspect_ratio pad \
24
+ --group_by_modality_length True \
25
+ --bf16 True \
26
+ --output_dir "${output_dir}/checkpoints/${model_name}" \
27
+ --num_train_epochs 1 \
28
+ --per_device_train_batch_size 2 \
29
+ --per_device_eval_batch_size 2 \
30
+ --gradient_accumulation_steps 5 \
31
+ --evaluation_strategy "no" \
32
+ --save_strategy "steps" \
33
+ --save_steps 50000 \
34
+ --save_total_limit 1 \
35
+ --learning_rate 2e-6 \
36
+ --weight_decay 0. \
37
+ --warmup_ratio 0.03 \
38
+ --lr_scheduler_type "cosine" \
39
+ --logging_steps 1 \
40
+ --tf32 True \
41
+ --model_max_length 2048 \
42
+ --gradient_checkpointing True \
43
+ --dataloader_num_workers 1 \
44
+ --lazy_preprocess True
45
 
46
  ## Model Details
47