We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When I want to run the pt_example, the Error happened.
Error displaying widget: model not found
TypeError Traceback (most recent call last) Cell In[15], line 2 1 # Train the model for 400 epochs (= training steps) ----> 2 trainer.fit(model=learn, train_dataloaders=dataloader) File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py:520, in Trainer.fit(self, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path) 518 model = _maybe_unwrap_optimized(model) 519 self.strategy._lightning_module = model --> 520 call._call_and_handle_interrupt( 521 self, self._fit_impl, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path 522 ) File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/call.py:44, in _call_and_handle_interrupt(trainer, trainer_fn, *args, **kwargs) 42 return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) 43 else: ---> 44 return trainer_fn(*args, **kwargs) 46 except _TunerExitException: 47 _call_teardown_hook(trainer) File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py:559, in Trainer._fit_impl(self, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path) 549 self._data_connector.attach_data( 550 model, train_dataloaders=train_dataloaders, val_dataloaders=val_dataloaders, datamodule=datamodule 551 ) 553 ckpt_path = self._checkpoint_connector._select_ckpt_path( 554 self.state.fn, 555 ckpt_path, 556 model_provided=True, 557 model_connected=self.lightning_module is not None, 558 ) --> 559 self._run(model, ckpt_path=ckpt_path) 561 assert self.state.stopped 562 self.training = False File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py:935, in Trainer._run(self, model, ckpt_path) 930 self._signal_connector.register_signal_handlers() 932 # ---------------------------- 933 # RUN THE TRAINER 934 # ---------------------------- --> 935 results = self._run_stage() 937 # ---------------------------- 938 # POST-Training CLEAN UP 939 # ---------------------------- 940 log.debug(f"{self.__class__.__name__}: trainer tearing down") File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py:978, in Trainer._run_stage(self) 976 self._run_sanity_check() 977 with torch.autograd.set_detect_anomaly(self._detect_anomaly): --> 978 self.fit_loop.run() 979 return None 980 raise RuntimeError(f"Unexpected state {self.state}") File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/fit_loop.py:201, in _FitLoop.run(self) 199 try: 200 self.on_advance_start() --> 201 self.advance() 202 self.on_advance_end() 203 self._restarting = False File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/fit_loop.py:354, in _FitLoop.advance(self) 352 self._data_fetcher.setup(combined_loader) 353 with self.trainer.profiler.profile("run_training_epoch"): --> 354 self.epoch_loop.run(self._data_fetcher) File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/training_epoch_loop.py:133, in _TrainingEpochLoop.run(self, data_fetcher) 131 while not self.done: 132 try: --> 133 self.advance(data_fetcher) 134 self.on_advance_end() 135 self._restarting = False File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/training_epoch_loop.py:218, in _TrainingEpochLoop.advance(self, data_fetcher) 215 with trainer.profiler.profile("run_training_batch"): 216 if trainer.lightning_module.automatic_optimization: 217 # in automatic optimization, there can only be one optimizer --> 218 batch_output = self.automatic_optimization.run(trainer.optimizers[0], kwargs) 219 else: 220 batch_output = self.manual_optimization.run(kwargs) File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/optimization/automatic.py:185, in _AutomaticOptimization.run(self, optimizer, kwargs) 178 closure() 180 # ------------------------------ 181 # BACKWARD PASS 182 # ------------------------------ 183 # gradient update with accumulated gradients 184 else: --> 185 self._optimizer_step(kwargs.get("batch_idx", 0), closure) 187 result = closure.consume_result() 188 if result.loss is None: File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/loops/optimization/automatic.py:261, in _AutomaticOptimization._optimizer_step(self, batch_idx, train_step_and_backward_closure) 258 self.optim_progress.optimizer.step.increment_ready() 260 # model hook --> 261 call._call_lightning_module_hook( 262 trainer, 263 "optimizer_step", 264 trainer.current_epoch, 265 batch_idx, 266 optimizer, 267 train_step_and_backward_closure, 268 ) 270 if not should_accumulate: 271 self.optim_progress.optimizer.step.increment_completed() File /opt/conda/lib/python3.10/site-packages/pytorch_lightning/trainer/call.py:142, in _call_lightning_module_hook(trainer, hook_name, pl_module, *args, **kwargs) 139 pl_module._current_fx_name = hook_name 141 with trainer.profiler.profile(f"[LightningModule]{pl_module.__class__.__name__}.{hook_name}"): --> 142 output = fn(*args, **kwargs) 144 # restore current_fx when nested context 145 pl_module._current_fx_name = prev_fx_name TypeError: SequenceLearner.optimizer_step() missing 1 required positional argument: 'closure'```
The text was updated successfully, but these errors were encountered:
The pt_example.py was outdated. Please refer to https://ncps.readthedocs.io/en/latest/examples/torch_first_steps.html on how to use NCPs with Pytorch
Sorry, something went wrong.
No branches or pull requests
When I want to run the pt_example, the Error happened.
Error displaying widget: model not found
The text was updated successfully, but these errors were encountered: