stash commited on
Commit
71b2a35
·
verified ·
1 Parent(s): e03ef2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -6,7 +6,41 @@ library_name: peft
6
  # Model Card for Model ID
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
 
12
  ## Model Details
 
6
  # Model Card for Model ID
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
+ echo "Starting LoRA fine-tuning..."
10
+ deepspeed LLaVA/llava/train/train_mem.py \
11
+ --lora_enable True --lora_r 128 --lora_alpha 256 --mm_projector_lr 2e-5 \
12
+ --deepspeed ./LLaVA/scripts/zero3.json \
13
+ --model_name_or_path liuhaotian/llava-v1.5-7b \
14
+ --version v1 \
15
+ --data_path "${output_dir}/processed_dataset_correct_path.json" \
16
+ --image_folder "${image_dir}" \
17
+ --vision_tower openai/clip-vit-large-patch14-336 \
18
+ --mm_projector_type mlp2x_gelu \
19
+ --mm_vision_select_layer -2 \
20
+ --mm_use_im_start_end False \
21
+ --mm_use_im_patch_token False \
22
+ --image_aspect_ratio pad \
23
+ --group_by_modality_length True \
24
+ --bf16 True \
25
+ --output_dir "${output_dir}/checkpoints/${model_name}" \
26
+ --num_train_epochs 1 \
27
+ --per_device_train_batch_size 3 \
28
+ --per_device_eval_batch_size 3 \
29
+ --gradient_accumulation_steps 10 \
30
+ --evaluation_strategy "no" \
31
+ --save_strategy "steps" \
32
+ --save_steps 50000 \
33
+ --save_total_limit 1 \
34
+ --learning_rate 2e-4 \
35
+ --weight_decay 0. \
36
+ --warmup_ratio 0.03 \
37
+ --lr_scheduler_type "cosine" \
38
+ --logging_steps 1 \
39
+ --tf32 True \
40
+ --model_max_length 2048 \
41
+ --gradient_checkpointing True \
42
+ --dataloader_num_workers 1 \
43
+ --lazy_preprocess True
44
 
45
 
46
  ## Model Details