| accelerate launch --config_file config.yaml peft_adalora_whisper_large_training.py \ | |
| --model_name_or_path "openai/whisper-large-v2" \ | |
| --language "Marathi" \ | |
| --language_abbr "mr" \ | |
| --task "transcribe" \ | |
| --dataset_name "mozilla-foundation/common_voice_11_0" \ | |
| --push_to_hub \ | |
| --preprocessing_num_workers 2 \ | |
| --per_device_train_batch_size 8 \ | |
| --per_device_eval_batch_size 8 \ | |
| --dataloader_pin_memory \ | |
| --dataloader_num_workers 2 \ | |
| --learning_rate 1e-3 \ | |
| --weight_decay 1e-4 \ | |
| --num_train_epochs 3 \ | |
| --gradient_accumulation_steps 1 \ | |
| --lr_scheduler_type "linear" \ | |
| --num_warmup_steps 50 \ | |
| --output_dir "adalora_whisper_large_marathi_multi_adapter" \ | |
| --seed 42 \ | |
| --load_best_model \ | |
| --with_tracking \ | |
| --report_to "wandb" \ | |
| --hub_token $HUB_TOKEN \ | |
| --checkpointing_steps 2000 \ | |
| --evaluation_steps 2000 \ | |
| --logging_steps 25 \ | |
| --use_peft \ | |
| --use_adalora \ | |
| --init_r 12 \ | |
| --target_r 8 \ | |
| --tinit 100 \ | |
| --tfinal 800 \ | |
| --delta_t 10 \ | |
| --lora_alpha 32 \ | |
| --lora_dropout 0.1 \ | |
| --orth_reg_weight 0.5 |