CUDA_VISIBLE_DEVICES=1,2,3 python src/train_sft.py \ --model_name_or_path /home/majiahui/project/models-llm/openbuddy-llama-7b-v1.4-fp16/ \ --do_train \ --dataset paper_data_v6_prompt \ --dataset_dir /home/majiahui/project/LLaMA-Efficient-Tuning/data \ --finetuning_type freeze \ --max_source_length 700 \ --max_target_length 1400 \ --output_dir path_to_sft_openbuddy_llama_paper_checkpoint_prompt_freeze_checkpoint_data_v6 \ --overwrite_cache \ --per_device_train_batch_size 2 \ --gradient_accumulation_steps 1 \ --lr_scheduler_type cosine \ --logging_steps 10 \ --save_steps 6000 \ --learning_rate 5e-5 \ --num_train_epochs 3.0 \ --plot_loss \ --fp16 \