0%| | 0/1000 [00:00> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-10\config.json [INFO|configuration_utils.py:362] 2023-04-22 15:58:29,416 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-10\generation_config.json [INFO|modeling_utils.py:1762] 2023-04-22 15:58:29,657 >> Model weights saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-10\pytorch_model.bin [INFO|tokenization_utils_base.py:2163] 2023-04-22 15:58:29,662 >> tokenizer config file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-10\tokenizer_config.json [INFO|tokenization_utils_base.py:2170] 2023-04-22 15:58:29,664 >> Special tokens file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-10\special_tokens_map.json 2%|█▉ | 19/1000 [08:55<8:52:26, 32.56s/it] {'loss': 1.6609, 'learning_rate': 0.0196, 'epoch': 16.84} 2%|██ | 20/1000 [09:09<7:20:43, 26.98s/it][INFO|configuration_utils.py:457] 2023-04-22 16:03:18,188 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-20\config.json [INFO|configuration_utils.py:362] 2023-04-22 16:03:18,191 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-20\generation_config.json [INFO|modeling_utils.py:1762] 2023-04-22 16:03:18,399 >> Model weights saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-20\pytorch_model.bin [INFO|tokenization_utils_base.py:2163] 2023-04-22 16:03:18,403 >> tokenizer config file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-20\tokenizer_config.json [INFO|tokenization_utils_base.py:2170] 2023-04-22 16:03:18,405 >> Special tokens file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-20\special_tokens_map.json 3%|██▉ | 29/1000 [12:01<4:03:58, 15.08s/it] {'loss': 0.38, 'learning_rate': 0.0194, 'epoch': 25.26} 3%|███ | 30/1000 [12:14<3:54:04, 14.48s/it][INFO|configuration_utils.py:457] 2023-04-22 16:06:23,084 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-30\config.json [INFO|configuration_utils.py:362] 2023-04-22 16:06:23,086 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-30\generation_config.json [INFO|modeling_utils.py:1762] 2023-04-22 16:06:23,292 >> Model weights saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-30\pytorch_model.bin [INFO|tokenization_utils_base.py:2163] 2023-04-22 16:06:23,296 >> tokenizer config file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-30\tokenizer_config.json [INFO|tokenization_utils_base.py:2170] 2023-04-22 16:06:23,296 >> Special tokens file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-30\special_tokens_map.json 4%|████ | 39/1000 [14:51<4:40:29, 17.51s/it] {'loss': 0.0535, 'learning_rate': 0.0192, 'epoch': 33.68} 4%|████ | 40/1000 [15:11<4:49:19, 18.08s/it][INFO|configuration_utils.py:457] 2023-04-22 16:09:20,023 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-40\config.json [INFO|configuration_utils.py:362] 2023-04-22 16:09:20,027 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-40\generation_config.json [INFO|modeling_utils.py:1762] 2023-04-22 16:09:20,233 >> Model weights saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-40\pytorch_model.bin [INFO|tokenization_utils_base.py:2163] 2023-04-22 16:09:20,237 >> tokenizer config file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-40\tokenizer_config.json [INFO|tokenization_utils_base.py:2170] 2023-04-22 16:09:20,238 >> Special tokens file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-40\special_tokens_map.json 5%|█████▏ | 50/1000 [18:38<5:22:50, 20.39s/it][INFO|configuration_utils.py:457] 2023-04-22 16:12:47,553 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-50\config.json [INFO|configuration_utils.py:362] 2023-04-22 16:12:47,556 >> Configuration saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-50\generation_config.json [INFO|modeling_utils.py:1762] 2023-04-22 16:12:47,773 >> Model weights saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-50\pytorch_model.bin [INFO|tokenization_utils_base.py:2163] 2023-04-22 16:12:47,780 >> tokenizer config file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-50\tokenizer_config.json [INFO|tokenization_utils_base.py:2170] 2023-04-22 16:12:47,781 >> Special tokens file saved in output\adgen-chatglm-6b-pt-128-2e-2\checkpoint-50\special_tokens_map.json {'loss': 0.0295, 'learning_rate': 0.019, 'epoch': 42.11} Saving PrefixEncoder 5%|█████▎ | 51/1000 [18:59<5:26:08, 20.62s/it]Traceback (most recent call last): File "main.py", line 444, in main() File "main.py", line 383, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 1635, in train return inner_training_loop( File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 1904, in _inner_training_loop tr_loss_step = self.training_step(model, inputs) File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 2665, in training_step loss.backward() File "D:\Program\Python38\lib\site-packages\torch\_tensor.py", line 487, in backward torch.autograd.backward( File "D:\Program\Python38\lib\site-packages\torch\autograd\__init__.py", line 200, in backward Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass KeyboardInterrupt Error in sys.excepthook: Traceback (most recent call last): File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1694, in print extend(render(renderable, render_options)) File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1330, in render yield from self.render(render_output, _options) File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1326, in render for render_output in iter_render: File "D:\Program\Python38\lib\site-packages\rich\constrain.py", line 29, in __rich_console__ yield from console.render(self.renderable, child_options) File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1326, in render for render_output in iter_render: File "D:\Program\Python38\lib\site-packages\rich\panel.py", line 220, in __rich_console__ lines = console.render_lines(renderable, child_options, style=style) File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1366, in render_lines lines = list( File "D:\Program\Python38\lib\site-packages\rich\segment.py", line 292, in split_and_crop_lines for segment in segments: File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1326, in render for render_output in iter_render: File "D:\Program\Python38\lib\site-packages\rich\padding.py", line 97, in __rich_console__ lines = console.render_lines( File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1366, in render_lines lines = list( File "D:\Program\Python38\lib\site-packages\rich\segment.py", line 292, in split_and_crop_lines for segment in segments: File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1330, in render yield from self.render(render_output, _options) File "D:\Program\Python38\lib\site-packages\rich\console.py", line 1326, in render for render_output in iter_render: File "D:\Program\Python38\lib\site-packages\rich\syntax.py", line 609, in __rich_console__ segments = Segments(self._get_syntax(console, options)) File "D:\Program\Python38\lib\site-packages\rich\segment.py", line 668, in __init__ self.segments = list(segments) File "D:\Program\Python38\lib\site-packages\rich\syntax.py", line 637, in _get_syntax text = self.highlight(processed_code, self.line_range) File "D:\Program\Python38\lib\site-packages\rich\syntax.py", line 509, in highlight text.append_tokens(tokens_to_spans()) File "D:\Program\Python38\lib\site-packages\rich\text.py", line 995, in append_tokens for content, style in tokens: File "D:\Program\Python38\lib\site-packages\rich\syntax.py", line 497, in tokens_to_spans _token_type, token = next(tokens) File "D:\Program\Python38\lib\site-packages\rich\syntax.py", line 484, in line_tokenize for token_type, token in lexer.get_tokens(code): File "D:\Program\Python38\lib\site-packages\pygments\lexer.py", line 190, in streamer for _, t, v in self.get_tokens_unprocessed(text): File "D:\Program\Python38\lib\site-packages\pygments\lexer.py", line 632, in get_tokens_unprocessed m = rexmatch(text, pos) KeyboardInterrupt Original exception was: Traceback (most recent call last): File "main.py", line 444, in main() File "main.py", line 383, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 1635, in train return inner_training_loop( File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 1904, in _inner_training_loop tr_loss_step = self.training_step(model, inputs) File "E:\Documents\Desktop\ChatGLM-6B\ptuning\trainer.py", line 2665, in training_step loss.backward() File "D:\Program\Python38\lib\site-packages\torch\_tensor.py", line 487, in backward torch.autograd.backward( File "D:\Program\Python38\lib\site-packages\torch\autograd\__init__.py", line 200, in backward Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass KeyboardInterrupt