Skip to content

Did you forget about the reward from the second agent? #1

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
FSNStefan opened this issue Jun 26, 2020 · 5 comments
Open

Did you forget about the reward from the second agent? #1

FSNStefan opened this issue Jun 26, 2020 · 5 comments

Comments

@FSNStefan
Copy link

Hi... I'm new with COMA and PyTorch.
As I read your code, it's impressive and useful.
However, I saw at line 165 you didn't include reward from the second agent.
I would like to know why...

@Bigpig4396
Copy link
Owner

That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents.

@SunnyWangGitHub
Copy link

That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents.

Hi! If i change the environment, eg. env_CatchPigs, which the two agents have different reward, What should be stored in r_list[t] used in line 94? The sum of rewards of agent1 and agent2? I'm new with COMA too...

@hccz95
Copy link

hccz95 commented Jul 28, 2022

That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents.

It seems that in this environment, the rewards of the two agents may be different at the end of an episode!

@Bigpig4396
Copy link
Owner

That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents.

It seems that in this environment, the rewards of the two agents may be different at the end of an episode!

Seems so. But they surely can have different rewards.

@hccz95
Copy link

hccz95 commented Jul 28, 2022

That is because in this environment, the two agents share the same reward. You can definitely record reward from other agents.

I tried to provide the two agents with the same reward (sum up their independent rewards) as in the paper of COMA, and your implementation still works!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants