-
Notifications
You must be signed in to change notification settings - Fork 4.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multiagent Bug in Gym-Unity 14.1 #3531
Comments
Hi @sahandrez, the multi-agent gym currently doesn't support environments with multiple types of behaviors (WallJump has two), or on-demand decisions (e.g. Bouncer). This is due to how gym expects all data per step to be from the same observation/action space. |
Hi @ervteng,
Any idea what might have caused it? Thanks. |
Hi @sahandrez, are the agents done at different times/do they use on demand decisions? Could you post the Inspector for your agents? Would be really helpful for us to debug. Thanks! |
Hi @ervteng, |
Hello @sahandrez , |
Hi @batu, |
Hi @batu , |
Unfortunately no. But I have something that is good enough. In my case, I believe it happened when my agent got out of bounds and didn't reset properly. It was going out of bound because my implementation was acting weird when I push the simulation speed very high. Currently, it happens only every 10-100 episodes, so when it does I just catch it and... pass. |
For this error:
We found a bug in the gym wrapper which would appear if an Agent called EndEpisode multiple times in the same step. The fix for this is in #3700, and we'll be doing a patch release next week that contains this fix (and a few others). |
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
Hello,
There seems to be a bug with multiagent environments in gym-unity release 14.1.
Some of the example environments work fine (e.g. 3D ball) while some of them (e.g. WallJump) fail with this error:
It is possible that the issue is originated in the Unity environment side, but I can't seem to find the reason behind it. Single Agent environments work fine.
I'm using the provided notebooks here.
Thanks,
Sahand
The text was updated successfully, but these errors were encountered: