Thesis/scripts/umls/pretrain_umls.sh

14 lines
430 B
Bash

nohup python -u main.py --gpus "0," --max_epochs=20 --num_workers=32 \
--model_name_or_path bert-base-uncased \
--accumulate_grad_batches 1 \
--model_class BertKGC \
--batch_size 128 \
--pretrain 1 \
--bce 0 \
--check_val_every_n_epoch 1 \
--overwrite_cache \
--data_dir xxx/Relphormer/dataset/umls \
--eval_batch_size 256 \
--max_seq_length 64 \
--lr 1e-4 \
>logs/pretrain_umls.log 2>&1 &