| export WANDB_MODE=disabled # 禁用wandb | |
| # 使用chinese-alpaca-plus-7b-merged模型在law_data.json数据集上finetune | |
| experiment_name="chinese-alpaca-plus-7b-law-e1" | |
| # 单卡或者模型并行 | |
| python finetune.py \ | |
| --base_model "minlik/chinese-alpaca-plus-7b-merged" \ | |
| --data_path "./data/finetune_law_data.json" \ | |
| --output_dir "./outputs/"${experiment_name} \ | |
| --batch_size 64 \ | |
| --micro_batch_size 8 \ | |
| --num_epochs 20 \ | |
| --learning_rate 3e-4 \ | |
| --cutoff_len 256 \ | |
| --val_set_size 0 \ | |
| --lora_r 8 \ | |
| --lora_alpha 16 \ | |
| --lora_dropout 0.05 \ | |
| --lora_target_modules "[q_proj,v_proj]" \ | |
| --train_on_inputs False \ | |
| --add_eos_token True \ | |
| --group_by_length False \ | |
| --wandb_project "" \ | |
| --wandb_run_name "" \ | |
| --wandb_watch "" \ | |
| --wandb_log_model "" \ | |
| --resume_from_checkpoint "./outputs/"${experiment_name} \ | |
| --prompt_template_name "alpaca" \ | |
| # 多卡数据并行 | |
| # WORLD_SIZE=8 CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 torchrun --nproc_per_node=8 --master_port=1234 finetune.py \ | |
| # --base_model "minlik/chinese-alpaca-plus-7b-merged" \ | |
| # --data_path "./data/finetune_law_data.json" \ | |
| # --output_dir "./outputs/"${experiment_name} \ | |
| # --batch_size 64 \ | |
| # --micro_batch_size 8 \ | |
| # --num_epochs 20 \ | |
| # --learning_rate 3e-4 \ | |
| # --cutoff_len 256 \ | |
| # --val_set_size 0 \ | |
| # --lora_r 8 \ | |
| # --lora_alpha 16 \ | |
| # --lora_dropout 0.05 \ | |
| # --lora_target_modules "[q_proj,v_proj]" \ | |
| # --train_on_inputs True \ | |
| # --add_eos_token True \ | |
| # --group_by_length False \ | |
| # --wandb_project \ | |
| # --wandb_run_name \ | |
| # --wandb_watch \ | |
| # --wandb_log_model \ | |
| # --resume_from_checkpoint "./outputs/"${experiment_name} \ | |
| # --prompt_template_name "alpaca" \ |