|
|
export OMINI_CONFIG=./config/math395.yaml |
|
|
|
|
|
|
|
|
export TOKENIZERS_PARALLELISM=true |
|
|
|
|
|
|
|
|
CUDA_INCLUDE_PATH="/home/work/miniconda3/envs/allm/include" |
|
|
|
|
|
|
|
|
export CPATH=$CPATH:$CUDA_INCLUDE_PATH |
|
|
export CPLUS_INCLUDE_PATH=$CPLUS_INCLUDE_PATH:$CUDA_INCLUDE_PATH |
|
|
|
|
|
|
|
|
|
|
|
export WANDB_PROJECT="Llama2_7B_FT_Math_395k" |
|
|
|
|
|
export OMP_NUM_THREADS=1 |
|
|
export MKL_NUM_THREADS=1 |
|
|
export OPENBLAS_NUM_THREADS=1 |
|
|
export NUMEXPR_NUM_THREADS=1 |
|
|
|
|
|
date +"%F %T" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex06" --trainer_args.learning_rate=1e-2 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 2.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
sleep 5 |
|
|
echo "6th exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex07" --trainer_args.learning_rate=1e-2 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 3.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
sleep 5 |
|
|
echo "6th exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex08" --trainer_args.learning_rate=2e-2 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 2.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
sleep 5 |
|
|
echo "8th exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex09" --trainer_args.learning_rate=2e-2 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 3.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
sleep 5 |
|
|
echo "9th exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex10" --trainer_args.learning_rate=3e-2 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 2.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
sleep 5 |
|
|
echo "10th exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
|
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex11" --trainer_args.learning_rate=8e-3 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 2.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
|
|
|
sleep 5 |
|
|
echo "11 exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
|
|
|
accelerate launch --main_process_port 41353 -m src.ft_mathQ \ |
|
|
--config_path $OMINI_CONFIG --trainer_args.output_dir "./exp100/run_ex12" --trainer_args.learning_rate=8e-3 \ |
|
|
--rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \ |
|
|
--trainer_args.num_train_epochs 3.0 --data.dataset_split train[:100000] \ |
|
|
--run_text 'def|o100k|b48' --trainer_args.per_device_train_batch_size 48 |
|
|
|
|
|
sleep 5 |
|
|
echo "12 exp finishes" |
|
|
date +"%F %T" |
|
|
wandb sync wandb/latest-run |
|
|
|
|
|
|
|
|
bash ./scripts/merge_100k.sh |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|