1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85
| # cuda = 0表示GPU 正常识别,可以启动cuda训练,world size反应GPU的卡数,compute dtype: torch.bfloat16表示使用的是bfloat16精度进行训练,比较节省内存,适合加速训练的情况 [INFO|2025-10-17 14:56:31] llamafactory.hparams.parser:423 >> Process rank: 0, world size: 1, device: cuda:0, distributed training: False, compute dtype: torch.bfloat16 # 加载微调训练数据集ruozhiba_qa_input_little.json [INFO|2025-10-17 14:56:32] llamafactory.data.loader:143 >> Loading dataset ruozhiba_qa_input_little.json... # 训练微调数据集中的数据格式是标准SFT格式,模型正确解析input_ids 和 label_ids, 说明数据预处理没问题,Tokenizer与数据格式兼容,标签正确对齐,其中label_ids是output的token IDs。 training example: input_ids: [151644, 8948, 198, 2610, 525, 1207, 16948, 11, 3465, 553, 54364, 14817, 13, 1446, 525, 264, 10950, 17847, 13, 151645, 198, 151644, 872, 198, 103434, 46944, 103023, 34187, 104246, 75606, 101037, 11319, 151645, 198, 151644, 77091, 198, 26232, 3837, 17340, 112706, 110382, 103023, 1773, 151645, 198] inputs: <|im_start|>system You are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|> <|im_start|>user 只剩一个心脏了还能活吗?<|im_end|> <|im_start|>assistant 能,人本来就只有一个心脏。<|im_end|>
label_ids: [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 26232, 3837, 17340, 112706, 110382, 103023, 1773, 151645, 198] labels: 能,人本来就只有一个心脏。<|im_end|>
[INFO|2025-10-17 14:57:19] llamafactory.model.model_utils.kv_cache:143 >> KV cache is disabled during training. [INFO|2025-10-17 14:57:48] llamafactory.model.model_utils.checkpointing:143 >> Gradient checkpointing enabled. [INFO|2025-10-17 14:57:48] llamafactory.model.model_utils.attention:143 >> Using torch SDPA for faster training and inference. [INFO|2025-10-17 14:57:48] llamafactory.model.adapter:143 >> Upcasting trainable params to float32. # LoRA训练正常启动 [INFO|2025-10-17 14:57:48] llamafactory.model.adapter:143 >> Fine-tuning method: LoRA [INFO|2025-10-17 14:57:48] llamafactory.model.model_utils.misc:143 >> Found linear modules: down_proj,gate_proj,up_proj,v_proj,k_proj,q_proj,o_proj
INFO ENV: Auto setting PYTORCH_CUDA_ALLOC_CONF='expandable_segments:True' for memory saving. INFO ENV: Auto setting CUDA_DEVICE_ORDER=PCI_BUS_ID for correctness. # 参与这次训练的参数是20,185,088,而基础模型总参数是7,635,801,600,被训练参数的参数只占总参数的0.2643%,这也反应出使用LoRA微调模型的好处,微调参数少、节省内存、训练快效果好 [INFO|2025-10-17 14:57:49] llamafactory.model.loader:143 >> trainable params: 20,185,088 || all params: 7,635,801,600 || trainable%: 0.2643 [INFO|2025-10-17 15:01:14] llamafactory.train.callbacks:143 >> {'loss': 1.5699, 'learning_rate': 4.9803e-05, 'epoch': 5.00, 'throughput': 7.96} {'loss': 1.5699, 'grad_norm': 4.0874104499816895, 'learning_rate': 4.980286753286195e-05, 'epoch': 5.0, 'num_input_tokens_seen': 1632, 'train_runtime': 205.0118, 'train_tokens_per_second': 7.961} [INFO|2025-10-17 15:04:41] llamafactory.train.callbacks:143 >> {'loss': 0.5215, 'learning_rate': 4.9007e-05, 'epoch': 10.00, 'throughput': 7.98} {'loss': 0.5215, 'grad_norm': 1.5765197277069092, 'learning_rate': 4.900734214192358e-05, 'epoch': 10.0, 'num_input_tokens_seen': 3280, 'train_runtime': 411.1345, 'train_tokens_per_second': 7.978} [INFO|2025-10-17 15:08:06] llamafactory.train.callbacks:143 >> {'loss': 0.0859, 'learning_rate': 4.7621e-05, 'epoch': 15.00, 'throughput': 7.96} {'loss': 0.0859, 'grad_norm': 0.27630168199539185, 'learning_rate': 4.762067631165049e-05, 'epoch': 15.0, 'num_input_tokens_seen': 4912, 'train_runtime': 617.0055, 'train_tokens_per_second': 7.961} [INFO|2025-10-17 15:11:31] llamafactory.train.callbacks:143 >> {'loss': 0.0102, 'learning_rate': 4.5677e-05, 'epoch': 20.00, 'throughput': 7.97} {'loss': 0.0102, 'grad_norm': 0.08512571454048157, 'learning_rate': 4.567701435686404e-05, 'epoch': 20.0, 'num_input_tokens_seen': 6544, 'train_runtime': 821.1588, 'train_tokens_per_second': 7.969} [INFO|2025-10-17 15:14:57] llamafactory.train.callbacks:143 >> {'loss': 0.0033, 'learning_rate': 4.3224e-05, 'epoch': 25.00, 'throughput': 7.99} {'loss': 0.0033, 'grad_norm': 0.04416470229625702, 'learning_rate': 4.3224215685535294e-05, 'epoch': 25.0, 'num_input_tokens_seen': 8208, 'train_runtime': 1027.3333, 'train_tokens_per_second': 7.99} [INFO|2025-10-17 15:18:25] llamafactory.train.callbacks:143 >> {'loss': 0.0018, 'learning_rate': 4.0323e-05, 'epoch': 30.00, 'throughput': 8.01} {'loss': 0.0018, 'grad_norm': 0.029697248712182045, 'learning_rate': 4.0322676341324415e-05, 'epoch': 30.0, 'num_input_tokens_seen': 9888, 'train_runtime': 1235.0521, 'train_tokens_per_second': 8.006} [INFO|2025-10-17 15:21:53] llamafactory.train.callbacks:143 >> {'loss': 0.0010, 'learning_rate': 3.7044e-05, 'epoch': 35.00, 'throughput': 8.01} {'loss': 0.001, 'grad_norm': 0.012993291951715946, 'learning_rate': 3.704384185254288e-05, 'epoch': 35.0, 'num_input_tokens_seen': 11568, 'train_runtime': 1443.7118, 'train_tokens_per_second': 8.013} [INFO|2025-10-17 15:25:18] llamafactory.train.callbacks:143 >> {'loss': 0.0007, 'learning_rate': 3.3468e-05, 'epoch': 40.00, 'throughput': 8.01} {'loss': 0.0007, 'grad_norm': 0.0086961779743433, 'learning_rate': 3.346844800613229e-05, 'epoch': 40.0, 'num_input_tokens_seen': 13200, 'train_runtime': 1648.2519, 'train_tokens_per_second': 8.008} [INFO|2025-10-17 15:28:42] llamafactory.train.callbacks:143 >> {'loss': 0.0005, 'learning_rate': 2.9685e-05, 'epoch': 45.00, 'throughput': 8.01} {'loss': 0.0005, 'grad_norm': 0.006480826064944267, 'learning_rate': 2.9684532864643122e-05, 'epoch': 45.0, 'num_input_tokens_seen': 14848, 'train_runtime': 1852.9815, 'train_tokens_per_second': 8.013} [INFO|2025-10-17 15:32:08] llamafactory.train.callbacks:143 >> {'loss': 0.0004, 'learning_rate': 2.5785e-05, 'epoch': 50.00, 'throughput': 8.02} {'loss': 0.0004, 'grad_norm': 0.005220182705670595, 'learning_rate': 2.578526897695321e-05, 'epoch': 50.0, 'num_input_tokens_seen': 16512, 'train_runtime': 2058.6258, 'train_tokens_per_second': 8.021} [INFO|2025-10-17 15:35:35] llamafactory.train.callbacks:143 >> {'loss': 0.0004, 'learning_rate': 2.1867e-05, 'epoch': 55.00, 'throughput': 8.02} {'loss': 0.0004, 'grad_norm': 0.0046224468387663364, 'learning_rate': 2.186666916089239e-05, 'epoch': 55.0, 'num_input_tokens_seen': 18176, 'train_runtime': 2265.8238, 'train_tokens_per_second': 8.022} [INFO|2025-10-17 15:38:59] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 1.8025e-05, 'epoch': 60.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.004301400855183601, 'learning_rate': 1.802522234901927e-05, 'epoch': 60.0, 'num_input_tokens_seen': 19824, 'train_runtime': 2469.7645, 'train_tokens_per_second': 8.027} [INFO|2025-10-17 15:42:24] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 1.4356e-05, 'epoch': 65.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.003765035653486848, 'learning_rate': 1.4355517710873184e-05, 'epoch': 65.0, 'num_input_tokens_seen': 21488, 'train_runtime': 2674.4326, 'train_tokens_per_second': 8.035} [INFO|2025-10-17 15:45:55] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 1.0948e-05, 'epoch': 70.00, 'throughput': 8.02} {'loss': 0.0003, 'grad_norm': 0.0038004806265234947, 'learning_rate': 1.0947915553696742e-05, 'epoch': 70.0, 'num_input_tokens_seen': 23152, 'train_runtime': 2885.9865, 'train_tokens_per_second': 8.022} [INFO|2025-10-17 15:49:23] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 7.8863e-06, 'epoch': 75.00, 'throughput': 8.02} {'loss': 0.0003, 'grad_norm': 0.003600555472075939, 'learning_rate': 7.886322351782783e-06, 'epoch': 75.0, 'num_input_tokens_seen': 24816, 'train_runtime': 3093.3623, 'train_tokens_per_second': 8.022} [INFO|2025-10-17 15:52:51] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 5.2461e-06, 'epoch': 80.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.0036511190701276064, 'learning_rate': 5.24612469060774e-06, 'epoch': 80.0, 'num_input_tokens_seen': 26496, 'train_runtime': 3301.4524, 'train_tokens_per_second': 8.026} [INFO|2025-10-17 15:56:17] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 3.0923e-06, 'epoch': 85.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.0035285328049212694, 'learning_rate': 3.0923329989034132e-06, 'epoch': 85.0, 'num_input_tokens_seen': 28176, 'train_runtime': 3507.221, 'train_tokens_per_second': 8.034} [INFO|2025-10-17 15:59:45] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 1.4780e-06, 'epoch': 90.00, 'throughput': 8.04} {'loss': 0.0003, 'grad_norm': 0.003523440333083272, 'learning_rate': 1.4779807761443636e-06, 'epoch': 90.0, 'num_input_tokens_seen': 29856, 'train_runtime': 3715.0666, 'train_tokens_per_second': 8.036} [INFO|2025-10-17 16:03:12] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 4.4282e-07, 'epoch': 95.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.003556356066837907, 'learning_rate': 4.4281873178278475e-07, 'epoch': 95.0, 'num_input_tokens_seen': 31504, 'train_runtime': 3922.8514, 'train_tokens_per_second': 8.031} [INFO|2025-10-17 16:06:36] llamafactory.train.callbacks:143 >> {'loss': 0.0003, 'learning_rate': 1.2336e-08, 'epoch': 100.00, 'throughput': 8.03} {'loss': 0.0003, 'grad_norm': 0.0034833240788429976, 'learning_rate': 1.233599085671e-08, 'epoch': 100.0, 'num_input_tokens_seen': 33152, 'train_runtime': 4126.4772, 'train_tokens_per_second': 8.034} {'train_runtime': 4127.2231, 'train_samples_per_second': 0.097, 'train_steps_per_second': 0.024, 'train_loss': 0.10991070384858176, 'epoch': 100.0, 'num_input_tokens_seen': 33152} ***** train metrics ***** epoch = 100.0 num_input_tokens_seen = 33152 total_flos = 1313580GF train_loss = 0.1099 train_runtime = 1:08:47.22 train_samples_per_second = 0.097 train_steps_per_second = 0.024 Figure saved at: saves\Qwen2.5-7B-Instruct\lora\train_2025-10-17-14-25-15\training_loss.png [WARNING|2025-10-17 16:06:37] llamafactory.extras.ploting:148 >> No metric eval_loss to plot. [WARNING|2025-10-17 16:06:37] llamafactory.extras.ploting:148 >> No metric eval_accuracy to plot.
|