Sven Mika d5bfb7b7da [RLlib] Preparatory PR for multi-agent multi-GPU learner (alpha-star style) #03 (#21652) 2 年之前
..
buffers c4636c7c05 [RLlib] Issue 21633: SimpleQ should not use a prio. replay buffer. (#21665) 2 年之前
tests d07e50e957 [RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552) 2 年之前
__init__.py 62dbf26394 [RLlib] POC: Run PGTrainer w/o the distr. exec API (Trainer's new training_iteration method). (#20984) 2 年之前
common.py 62dbf26394 [RLlib] POC: Run PGTrainer w/o the distr. exec API (Trainer's new training_iteration method). (#20984) 2 年之前
concurrency_ops.py 1520c3d147 [RLlib] Deepcopy env_ctx for vectorized sub-envs AND add eval-worker-option to `Trainer.add_policy()` (#18428) 3 年之前
learner_thread.py 49cd7ea6f9 [RLlib] Trainer sub-class PPO/DDPPO (instead of `build_trainer()`). (#20571) 2 年之前
metric_ops.py d5bfb7b7da [RLlib] Preparatory PR for multi-agent multi-GPU learner (alpha-star style) #03 (#21652) 2 年之前
multi_gpu_impl.py a931076f59 [RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981) 3 年之前
multi_gpu_learner.py a931076f59 [RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981) 3 年之前
multi_gpu_learner_thread.py d07e50e957 [RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552) 2 年之前
replay_ops.py d07e50e957 [RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552) 2 年之前
rollout_ops.py 90c6b10498 [RLlib] Decentralized multi-agent learning; PR #01 (#21421) 2 年之前
segment_tree.py 4888d7c9af [RLlib] Replay buffers: Add config option to store contents in checkpoints. (#17999) 3 年之前
train_ops.py 7dc3de4eed [RLlib] Fix config mismatch for train_one_step. num_sgd_iter instead of sgd_num_iter. (#21555) 2 年之前
tree_agg.py 90c6b10498 [RLlib] Decentralized multi-agent learning; PR #01 (#21421) 2 年之前