atari-impala-multi-gpu.yaml 602 B

1234567891011121314151617181920
  1. # Runs on a p2.8xlarge single head node machine.
  2. # Should reach ~400 reward in about 1h and after 15-20M ts.
  3. atari-impala:
  4. env:
  5. BreakoutNoFrameskip-v4
  6. run: IMPALA
  7. config:
  8. # Works for both torch and tf.
  9. framework: torch
  10. rollout_fragment_length: 50
  11. train_batch_size: 4000
  12. num_gpus: 4
  13. num_workers: 31
  14. num_gpus_per_worker: 0 # works also for partial GPUs (<1.0) per worker
  15. num_envs_per_worker: 5
  16. clip_rewards: True
  17. lr_schedule: [
  18. [0, 0.0005],
  19. [20000000, 0.000000000001],
  20. ]