seed:0output_dir:'./output'# path to save checkpoint/strategyload_checkpoint:'/mnt/dataset/mindformers/research/pretrain/ckpt'src_strategy_path_or_dir:''auto_trans_ckpt:True# If true, auto transform load_checkpoint to load in distributed modelonly_save_strategy:Falseresume_training:Falseuse_parallel:Truerun_mode:'finetune'# trainer configtrainer:type: CausalLanguageModelingTrainer
model_name:'qwen2_7b'# datasettrain_dataset:&train_datasetdata_loader:type: MindDataset
dataset_dir:"/mnt/dataset/alpaca_ds/mindrecord/alpaca-messages.mindrecord"shuffle:Trueinput_columns:["input_ids","target_ids","attention_mask"]num_parallel_workers:8python_multiprocessing:Falsedrop_remainder:Truebatch_size:2repeat:1numa_enable:Falseprefetch_size:1train_dataset_task:type: CausalLanguageModelDataset
dataset_config:*train_dataset# runner configrunner_config:epochs:5batch_size:2sink_mode:True