-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Loss rises during training。 #1
Comments
Thanks for your answer, but clustering loss is also increased. |
Hi, I still have a problem. In the code below: attn_for_self = torch.mm(h, self.a_self) # (N,1)
attn_for_neighs = torch.mm(h, self.a_neighs) # (N,1)
attn_dense = attn_for_self + torch.transpose(attn_for_neighs, 0, 1) # 妙啊
attn_dense = torch.mul(attn_dense, M)
attn_dense = self.leakyrelu(attn_dense) # (N,N)
zero_vec = -9e15 * torch.ones_like(adj)
adj = torch.where(adj > 0, attn_dense, zero_vec)
attention = F.softmax(adj, dim=1) according to the paper, we need to use generalized neighbors of node, so in code adj = torch.where(adj > 0, attn_dense, zero_vec), why use adj, instead of attn_dense. just like: adj = torch.where(attn_dense > 0, attn_dense, zero_vec) |
Thanks for your answer. The paper proposes to exploit high-order neighbors. but |
when i run your model, loss and acc are keep increasing, just like the picture below.

the model users the gradient descent method to optimize, but the loss is increased. I know that in the KL
loss, the Q distribution is approached to the P distribution, and both distributions are changed during training. So i want to know whether it is normal for loss and acc to increase during training, and why?
The text was updated successfully, but these errors were encountered: