-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Expand file tree
/
Copy pathWan-Distill.sh
More file actions
13 lines (13 loc) · 866 Bytes
/
Wan-Distill.sh
File metadata and controls
13 lines (13 loc) · 866 Bytes
1
2
3
4
5
6
7
8
9
10
11
12
13
accelerate launch examples/wanvideo/model_training/train_distill.py \
--dataset_base_path "data/paired_beverage_video_advertising" \
--dataset_metadata_path "data/paired_beverage_video_advertising/metadata.csv" \
--model_id_with_origin_paths "Wan-AI/Wan2.1-T2V-1.3B:diffusion_pytorch_model*.safetensors,Wan-AI/Wan2.1-T2V-1.3B:Wan2.1_VAE.pth,Wan-AI/Wan2.1-T2V-1.3B:models_t5_umt5-xxl-enc-bf16.pth" \
--teacher_model_id_with_origin_paths "Wan-AI/Wan2.2-TI2V-5B:DiT-g-1001.safetensors,Wan-AI/Wan2.2-TI2V-5B:vae.safetensors,Wan-AI/Wan2.2-TI2V-5B:text_encoder.safetensors,Wan-AI/Wan2.2-TI2V-5B:image_encoder.safetensors" \
--lora_base_model "dit" \
--output_path "./models/train/wan_distill_lora" \
--lora_rank 32 \
--distillation_weight 0.5 \
--learning_rate 1e-4 \
--num_epochs 5 \
--trainable_models "dit" \
--use_gradient_checkpointing_offload