1,微调transformer中mlm任务中(多卡跑),nohup运行出现如下错误,经搜索说是nohup的bug
{'loss': 1.5461, 'learning_rate': 3.933343085625122e-05, 'epoch': 0.64} 21%|██▏ | 35000/164064 [2:03:12<7:30:17, 4.78it/s][INFO|trainer.py:2700] 2022-12-28 20:39:56,894 >> Saving model checkpoint to /data//models/myhugBert30w2/checkpoint-35000 [INFO|configuration_utils.py:447] 2022-12-28 20:39:56,895 >> Configuration saved in /data//models/myhugBert30w2/checkpoint-35000/config.json [INFO|modeling_utils.py:1702] 2022-12-28 20:39:57,345 >> Model weights saved in /data//models/myhugBert30w2/checkpoint-35000/pytorch_model.bin [INFO|tokenization_utils_base.py:2157] 2022-12-28 20:39:57,346 >> tokenizer config file saved in /data//models/myhugBert30w2/checkpoint-35000/tokenizer_config.json [INFO|tokeniz