-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Test results, the acceptance rate is always 1 #5
Comments
Hi, could you provide me with more details so I can better understand the problems?
|
1.I am verifying HRL-ACRA, and I have trained the upper and lower agents. 2.The p_net_setting.yaml is the default value, which I did not change. The v_sim_setting.yaml is based on the comparative experiments in the paper. I am very curious why the acceptance rate is always 1 or 0.999 during the verification process. During training, there is a verification every ten rounds, and the acceptance rate finally converges to about 0.89. 3.I have trained the A3C-GCN model before, and this is more normal and closer to the data in the paper. This problem seems to have been mentioned in issue 1, and it looks similar to mine |
Hi, the results of my running testing show that the AC rate is not always 1. Could you provide me with more details on any modifications to the original code? |
I may know what happens... if you set the decoding strategy to "beam," the AC will have a very high probability of being 1... |
Thanks for your feedback. I will check this issue this week. |
After I trained the model and ran several rounds of tests, I found that the acceptance rate was always 1. The acceptance rate during the validation process during training was about 0.9, which was quite different from the test results. I did not modify the core code, but I could not get results close to those in the paper. Why is this?
The text was updated successfully, but these errors were encountered: