.. |
buffers
|
c4636c7c05
[RLlib] Issue 21633: SimpleQ should not use a prio. replay buffer. (#21665)
|
2 年之前 |
tests
|
d07e50e957
[RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552)
|
2 年之前 |
__init__.py
|
62dbf26394
[RLlib] POC: Run PGTrainer w/o the distr. exec API (Trainer's new training_iteration method). (#20984)
|
2 年之前 |
common.py
|
62dbf26394
[RLlib] POC: Run PGTrainer w/o the distr. exec API (Trainer's new training_iteration method). (#20984)
|
2 年之前 |
concurrency_ops.py
|
1520c3d147
[RLlib] Deepcopy env_ctx for vectorized sub-envs AND add eval-worker-option to `Trainer.add_policy()` (#18428)
|
3 年之前 |
learner_thread.py
|
49cd7ea6f9
[RLlib] Trainer sub-class PPO/DDPPO (instead of `build_trainer()`). (#20571)
|
2 年之前 |
metric_ops.py
|
d5bfb7b7da
[RLlib] Preparatory PR for multi-agent multi-GPU learner (alpha-star style) #03 (#21652)
|
2 年之前 |
multi_gpu_impl.py
|
a931076f59
[RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981)
|
3 年之前 |
multi_gpu_learner.py
|
a931076f59
[RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981)
|
3 年之前 |
multi_gpu_learner_thread.py
|
d07e50e957
[RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552)
|
2 年之前 |
replay_ops.py
|
d07e50e957
[RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552)
|
2 年之前 |
rollout_ops.py
|
90c6b10498
[RLlib] Decentralized multi-agent learning; PR #01 (#21421)
|
2 年之前 |
segment_tree.py
|
4888d7c9af
[RLlib] Replay buffers: Add config option to store contents in checkpoints. (#17999)
|
3 年之前 |
train_ops.py
|
7dc3de4eed
[RLlib] Fix config mismatch for train_one_step. num_sgd_iter instead of sgd_num_iter. (#21555)
|
2 年之前 |
tree_agg.py
|
90c6b10498
[RLlib] Decentralized multi-agent learning; PR #01 (#21421)
|
2 年之前 |