Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
Happy 83a6dcaa3e | 1 year ago | |
---|---|---|
.. | ||
.benchmark | 2 years ago | |
README.md | 1 year ago | |
atari_agent.py | 1 year ago | |
atari_model.py | 2 years ago | |
replay_memory.py | 2 years ago | |
requirements.txt | 1 year ago | |
train.py | 1 year ago |
Based on PARL, the DQN algorithm of deep reinforcement learning has been reproduced, reaching the same level of indicators as the paper in Atari benchmarks.
DQN in Human-level Control Through Deep Reinforcement Learning
Dueling DQN in Dueling Network Architectures for Deep Reinforcement Learning
Please see here to know more about Atari games.
Benchmark results are obtained using different random seeds.
Performance of Dueling DQN on various environments:
Performance of Dueling DQN on 55 Atari environments:
Alien (5977) | Amidar (364) | Assault (9676) | Asterix (23800) | Asteroids (657) |
Atlantis (85633) | WizardOfWor (2767) | BankHeist (1143) | BattleZone (37667) | BeamRider (13570) |
Berzerk (827) | Bowling (47) | Boxing (100) | Breakout (409) | Centipede (5103) |
ChopperCommand (1300) | CrazyClimber (118733) | DemonAttack (167200) | DoubleDunk (-1) | Enduro (4153) |
FishingDerby (-64) | Freeway (22) | Frostbite (5273) | Gopher (11187) | Gravitar (0) |
Hero (14613) | IceHockey (2) | Jamesbond (767) | Kangaroo (4133) | Krull (8856) |
KungFuMaster (19933) | MontezumaRevenge (0) | MsPacman (4013) | NameThisGame (10327) | Phoenix (7333) |
Pitfall (0) | Pong (21) | PrivateEye (49) | Qbert (15275) | Riverraid (13410) |
RoadRunner (47167) | Robotank (27) | Seaquest (16573) | Skiing (-14409) | Solaris (53) |
SpaceInvaders (2797) | StarGunner (59367) | Tennis (0) | TimePilot (8200) | Tutankham (235) |
UpNDown (18153) | Venture (0) | VideoPinball (745800) | YarsRevenge (34346) | Zaxxon (13233) |
# To train a model for Pong game
python train.py
# For more customized arguments
python train.py --help
PARL 是一个高性能、灵活的强化学习框架
Python C++ JavaScript Shell Markdown other
Dear OpenI User
Thank you for your continuous support to the Openl Qizhi Community AI Collaboration Platform. In order to protect your usage rights and ensure network security, we updated the Openl Qizhi Community AI Collaboration Platform Usage Agreement in January 2024. The updated agreement specifies that users are prohibited from using intranet penetration tools. After you click "Agree and continue", you can continue to use our services. Thank you for your cooperation and understanding.
For more agreement content, please refer to the《Openl Qizhi Community AI Collaboration Platform Usage Agreement》