Sven Mika f3397b6f48 [RLlib] Minor fixes/cleanups; chop_into_sequences now handles nested data. (#19408) 3 年之前
..
atari-ddppo.yaml 93c0a5549b [RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397) 3 年之前
atari-ppo.yaml 93c0a5549b [RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397) 3 年之前
cartpole-appo-vtrace-fake-gpus.yaml e6ae08f416 [RLlib] Optionally don't drop last ts in v-trace calculations (APPO and IMPALA). (#19601) 3 年之前
cartpole-appo-vtrace-separate-losses.yaml 2d24ef0d32 [RLlib] Add all simple learning tests as `framework=tf2`. (#19273) 3 年之前
cartpole-appo-vtrace.yaml f3397b6f48 [RLlib] Minor fixes/cleanups; chop_into_sequences now handles nested data. (#19408) 3 年之前
cartpole-appo.yaml 5a313ba3d6 [RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169) 3 年之前
cartpole-ddppo.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
cartpole-grid-search-example.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
cartpole-ppo-fake-gpus.yaml 599e589481 [RLlib] Move existing fake multi-GPU learning tests into separate buildkite job. (#18065) 3 年之前
cartpole-ppo-hyperband.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
cartpole-ppo.yaml 93c0a5549b [RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397) 3 年之前
frozenlake-appo-vtrace.yaml 026bf01071 [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 3 年之前
halfcheetah-appo.yaml 5a313ba3d6 [RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169) 3 年之前
halfcheetah-ppo.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
hopper-ppo.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
humanoid-ppo-gae.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
humanoid-ppo.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前
pendulum-appo.yaml 026bf01071 [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 3 年之前
pendulum-ppo.yaml 026bf01071 [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 3 年之前
pendulum-transformed-actions-ppo.yaml 53206dd440 [RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531) 3 年之前
pong-appo.yaml 5a313ba3d6 [RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169) 3 年之前
pong-ppo.yaml 93c0a5549b [RLlib] Deprecate `vf_share_layers` in top-level PPO/MAML/MB-MPO configs. (#13397) 3 年之前
repeatafterme-ppo-lstm.yaml 8a72824c63 [RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591) 3 年之前
unity3d-soccer-strikers-vs-goalie-ppo.yaml e74947cc94 [RLlib] Env directory cleanup and tests. (#13082) 3 年之前
walker2d-ppo.yaml 2746fc0476 [RLlib] Auto-framework, retire `use_pytorch` in favor of `framework=...` (#8520) 4 年之前