-
Notifications
You must be signed in to change notification settings - Fork 14
Did you forget about the reward from the second agent? #1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents. |
Hi! If i change the environment, eg. env_CatchPigs, which the two agents have different reward, What should be stored in r_list[t] used in line 94? The sum of rewards of agent1 and agent2? I'm new with COMA too... |
It seems that in this environment, the rewards of the two agents may be different at the end of an episode! |
Seems so. But they surely can have different rewards. |
I tried to provide the two agents with the same reward (sum up their independent rewards) as in the paper of COMA, and your implementation still works! |
Hi... I'm new with COMA and PyTorch.
As I read your code, it's impressive and useful.
However, I saw at line 165 you didn't include reward from the second agent.
I would like to know why...
The text was updated successfully, but these errors were encountered: