You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Dear Prof. He:
I am a junior PhD student from Peking University, and I has read your paper and code, but still has a problem on understanding this whole idea essentially.
In previous months, i felt into thinking using GAN in Recommendation systems, while it's hard since rating is discrete and perturbing on rating is not proper!
We know GAN is used to enhance the robustness of model by creating more fine-grained negative samples using noise. Your paper has perturbed on the parameters (latent factors) of BPR-MF after its convergence.
While the parameters of BPR are stable and almost invariable after convergence. Then, the loss of adversarial part, plus the loss of BPR, reflecting on optimization, is a small-size gradient ascent plus a large-size gradient descent (value of delta = gradient of BPR, can be seen as gradient ascent to get larger adversarial loss), which is very like a large-step backward followed by a small-step forward to adjust the value of parameters (trade-off).
So the whole idea in essence can be interpreted as adjusting gradient optimization, right? GAN is just the outer wear.
But the improvement from BPR-MF to AMF is amazing, how did it get without more noisy input or less paramters or adjusting regulizer? It is very puzzling for me to find a good explanation.
Hope for your reply!
The text was updated successfully, but these errors were encountered:
Dear Prof. He:
I am a junior PhD student from Peking University, and I has read your paper and code, but still has a problem on understanding this whole idea essentially.
In previous months, i felt into thinking using GAN in Recommendation systems, while it's hard since rating is discrete and perturbing on rating is not proper!
We know GAN is used to enhance the robustness of model by creating more fine-grained negative samples using noise. Your paper has perturbed on the parameters (latent factors) of BPR-MF after its convergence.
While the parameters of BPR are stable and almost invariable after convergence. Then, the loss of adversarial part, plus the loss of BPR, reflecting on optimization, is a small-size gradient ascent plus a large-size gradient descent (value of delta = gradient of BPR, can be seen as gradient ascent to get larger adversarial loss), which is very like a large-step backward followed by a small-step forward to adjust the value of parameters (trade-off).
So the whole idea in essence can be interpreted as adjusting gradient optimization, right? GAN is just the outer wear.
But the improvement from BPR-MF to AMF is amazing, how did it get without more noisy input or less paramters or adjusting regulizer? It is very puzzling for me to find a good explanation.
Hope for your reply!
The text was updated successfully, but these errors were encountered: