Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How you compare performance of algorithms in the battle game? #8

Open
woaipichuli opened this issue Sep 4, 2018 · 4 comments
Open

Comments

@woaipichuli
Copy link

Recently, I have run the train_battle.py and battle.py. However, I found the performance of one algorithms may vary quite obviously in the battle game. I consider this is quite normal. As the performance of models saved in the 2000th generation may not be the best. Besides, with different random seeds, final performance of models obtained after the self-play process should be different. In this case, I wonder how you evaluate their performance and gain results in Fig.8. How many independent runs you have made for each algorithm in the comparative battles.

@woaipichuli
Copy link
Author

I wonder why the simulation results output by this program is different from that in the paper. Eg: The performance of MFAC outperforms MFQ obviously in this program, which is totally opposite to the results in the paper.

@lzh-awesome
Copy link

I wonder why the simulation results output by this program is different from that in the paper. Eg: The performance of MFAC outperforms MFQ obviously in this program, which is totally opposite to the results in the paper.

@woaipichuli hi,i have the same situation as you.i find even AC performs better than MFAC,and MFAC get more reward than MFQ.that's completely oppsite to the ressults in the paper.

@QianZhao-xd
Copy link

I wonder why the simulation results output by this program is different from that in the paper. Eg: The performance of MFAC outperforms MFQ obviously in this program, which is totally opposite to the results in the paper.

@woaipichuli hi,i have the same situation as you.i find even AC performs better than MFAC,and MFAC get more reward than MFQ.that's completely oppsite to the ressults in the paper.

Have you found a solution to the problem? I had the same problem

@lzh-awesome
Copy link

I wonder why the simulation results output by this program is different from that in the paper. Eg: The performance of MFAC outperforms MFQ obviously in this program, which is totally opposite to the results in the paper.

@woaipichuli hi,i have the same situation as you.i find even AC performs better than MFAC,and MFAC get more reward than MFQ.that's completely oppsite to the ressults in the paper.

Have you found a solution to the problem? I had the same problem

I don't use this code at present. I think it can be explained that the performance results of MFAC and MFQ are different from those mentioned in the paper, such as from the perspective of algorithm structure, model training episode, random seed number of algorithm training, operating equipment, etc. It can also be explained that the AC effect is better than the MFAC effect. The possible reason is that when the interaction between agents becomes complex, agents must act cooperatively according to the global state, which means that the local approximation assumption in the mean field method is not clear, leading to a significant decline in performance.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants