File size: 3,374 Bytes
dad14e4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
#!/bin/bash

export OMINI_CONFIG=./config/glue.yaml
export TOKENIZERS_PARALLELISM=true

export CPATH=$CPATH:$CUDA_INCLUDE_PATH
export CPLUS_INCLUDE_PATH=$CPLUS_INCLUDE_PATH:$CUDA_INCLUDE_PATH
# echo "CPATH is set to: $CPATH"
# echo "CPLUS_INCLUDE_PATH is set to: $CPLUS_INCLUDE_PATH"
# export PYTHONPATH=$PYTHONPATH:$(pwd)/../nl_tasks # for rpeft

export WANDB_PROJECT="DeBERTaV3-GLUE-Test"

export OMP_NUM_THREADS=1
export MKL_NUM_THREADS=1
export OPENBLAS_NUM_THREADS=1
export NUMEXPR_NUM_THREADS=1

date +"%F %T"

MODEL_LRS=("1e-4")
CLS_LRS=("2e-3")
DROPOUT_RATES=("0.1")
TEXT=("oft" "boft" "hra" "loco")

STEPS=2000
EPOCHS=11
for m_lr in "${MODEL_LRS[@]}"; do
  for c_lr in "${CLS_LRS[@]}"; do
    for drop_out in "${DROPOUT_RATES[@]}"; do

        echo ">>> Params: model_lr=$m_lr, cls_lr=$c_lr, dropout=$drop_out, step=$STEPS, epoch=$EPOCHS"
        python -m src.test \
            --config_path $OMINI_CONFIG --trainer_args.output_dir "./glue_testYY" --run_text 'oft' \
            --rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 6 \
            --trainer_args.gradient_accumulation_steps 1 \
            --glue.is_debug False --rotation_adapter_config.drop_out "$drop_out" \
            --glue.task_name qnli --trainer_args.metric_for_best_model accuracy \
            --trainer_args.num_train_epochs $EPOCHS --trainer_args.max_steps=405 --trainer_args.warmup_steps 200 \
            --glue.model_lr "$m_lr" --glue.cls_lr "$c_lr" \
            --trainer_args.logging_step $STEPS --trainer_args.eval_step $STEPS --trainer_args.save_steps $STEPS \
            --trainer_args.report_to none \
            --glue.max_seq_length 512 \
            --trainer_args.per_device_train_batch_size 64 --trainer_args.per_device_eval_batch_size 128 \
            --trainer_args.eval_strategy '"no"' \
            --trainer_args.load_best_model_at_end False \
            --trainer_args.save_strategy '"no"'


    done
  done
done

date +"%F %T"

# for m_lr in "${MODEL_LRS[@]}"; do
#   for c_lr in "${CLS_LRS[@]}"; do
#     for drop_out in "${DROPOUT_RATES[@]}"; do
#       for text in "${TEXT[@]}"; do
#         echo ">>> Params: model_lr=$m_lr, cls_lr=$c_lr, dropout=$drop_out, step=$STEPS, epoch=$EPOCHS"
#         python -m src.test \
#             --config_path $OMINI_CONFIG --trainer_args.output_dir "./glue_testYY" --run_text "$text" \
#             --rotation_adapter_config.num_rotations 1 --rotation_adapter_config.r 16 \
#             --trainer_args.gradient_accumulation_steps 1 \
#             --glue.is_debug False --rotation_adapter_config.drop_out "$drop_out" \
#             --glue.task_name qnli --trainer_args.metric_for_best_model accuracy \
#             --trainer_args.num_train_epochs $EPOCHS --trainer_args.max_steps=405 --trainer_args.warmup_steps 200 \
#             --glue.model_lr "$m_lr" --glue.cls_lr "$c_lr" \
#             --trainer_args.logging_step $STEPS --trainer_args.eval_step $STEPS --trainer_args.save_steps $STEPS \
#             --trainer_args.report_to none \
#             --glue.max_seq_length 512 \
#             --trainer_args.per_device_train_batch_size 64 --trainer_args.per_device_eval_batch_size 128 \
#             --trainer_args.eval_strategy '"no"' \
#             --trainer_args.load_best_model_at_end False \
#             --trainer_args.save_strategy '"no"'
#       done

#     done
#   done
# done