.. |
atari-ddppo.yaml
|
93c0a5549b
[RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397)
|
3 年之前 |
atari-ppo.yaml
|
93c0a5549b
[RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397)
|
3 年之前 |
cartpole-appo-vtrace-fake-gpus.yaml
|
e6ae08f416
[RLlib] Optionally don't drop last ts in v-trace calculations (APPO and IMPALA). (#19601)
|
3 年之前 |
cartpole-appo-vtrace-separate-losses.yaml
|
2d24ef0d32
[RLlib] Add all simple learning tests as `framework=tf2`. (#19273)
|
3 年之前 |
cartpole-appo-vtrace.yaml
|
f3397b6f48
[RLlib] Minor fixes/cleanups; chop_into_sequences now handles nested data. (#19408)
|
3 年之前 |
cartpole-appo.yaml
|
5a313ba3d6
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
3 年之前 |
cartpole-ddppo.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
cartpole-grid-search-example.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
cartpole-ppo-fake-gpus.yaml
|
599e589481
[RLlib] Move existing fake multi-GPU learning tests into separate buildkite job. (#18065)
|
3 年之前 |
cartpole-ppo-hyperband.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
cartpole-ppo.yaml
|
93c0a5549b
[RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397)
|
3 年之前 |
frozenlake-appo-vtrace.yaml
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
halfcheetah-appo.yaml
|
5a313ba3d6
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
3 年之前 |
halfcheetah-ppo.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
hopper-ppo.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
humanoid-ppo-gae.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
humanoid-ppo.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |
pendulum-appo.yaml
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
pendulum-ppo.yaml
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
pendulum-transformed-actions-ppo.yaml
|
53206dd440
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531)
|
3 年之前 |
pong-appo.yaml
|
5a313ba3d6
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
3 年之前 |
pong-ppo.yaml
|
93c0a5549b
[RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397)
|
3 年之前 |
repeatafterme-ppo-lstm.yaml
|
8a72824c63
[RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591)
|
3 年之前 |
unity3d-soccer-strikers-vs-goalie-ppo.yaml
|
e74947cc94
[RLlib] Env directory cleanup and tests. (#13082)
|
3 年之前 |
walker2d-ppo.yaml
|
2746fc0476
[RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520)
|
4 年之前 |