| LANG=java |
| DATADIR=../dataset/javaCorpus/token_completion |
| LITFILE=../dataset/javaCorpus/literals.json |
| OUTPUTDIR=../model/javaCorpus |
| PRETRAINDIR=microsoft/CodeGPT-small-java |
| LOGFILE=train_javaCorpus.log |
| PER_NODE_GPU=4 |
|
|
| CUDA_VISIBLE_DEVICES=0,1,2,3 python run_lm.py \ |
| --data_dir=$DATADIR \ |
| --lit_file=$LITFILE \ |
| --langs=$LANG \ |
| --output_dir=$OUTPUTDIR \ |
| --pretrain_dir=$PRETRAINDIR \ |
| --log_file=$LOGFILE \ |
| --model_type=gpt2 \ |
| --block_size=512 \ |
| --do_train \ |
| --gpu_per_node $PER_NODE_GPU \ |
| --learning_rate=8e-5 \ |
| --weight_decay=0.01 \ |
| --evaluate_during_training \ |
| --per_gpu_train_batch_size=1 \ |
| --per_gpu_eval_batch_size=4 \ |
| --gradient_accumulation_steps=4 \ |
| --num_train_epochs=5 \ |
| --logging_steps=100 \ |
| --save_steps=1000 \ |
| --seed=42 \ |
| --overwrite_output_dir \ |
| --not_pretrain |