Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Softmax fix #83

Open
wants to merge 2 commits into
base: main
Choose a base branch
from
Open

Softmax fix #83

wants to merge 2 commits into from

Conversation

rw404
Copy link

@rw404 rw404 commented Feb 21, 2023

  • Softmax layer in eval mode added;
  • model.train(), model.eval() in train.py placed for better performance

Fixes from issue #51

@twofeetcat
Copy link

twofeetcat commented Jun 30, 2023

hello,I can probably understand what you mean and follow your code to repeat it, but when calculating the loss, I wonder whether the following error occurred because my real label had 20 classes(the value of y_batch is 0, 1, ..., 19, shape of y_batch is [batch_size, 256, 256])
Error reported as follows:
Traceback (most recent call last): File "train.py", line 141, in <module> loss.backward() File "D:\software\anaconda\lib\site-packages\torch\_tensor.py", line 489, in backward self, gradient, retain_graph, create_graph, inputs=inputs File "D:\software\anaconda\lib\site-packages\torch\autograd\__init__.py", line 199, in backward allow_unreachable=True, accumulate_grad=True) # Calls into the C++ engine to run the backward pass RuntimeError: Unable to find a valid cuDNN algorithm to run convolution
But when I changed all values greater than 0 in y_batch to 1, that is, 20 classes to 0 and 1, the program did not have this error

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants