.. |
documentation
|
7a585fb275
[RLlib; Documentation] RLlib README overhaul. (#20249)
|
2 年之前 |
env
|
c288b97e5f
[RLlib] Issue 21629: Video recorder env wrapper not working. Added test case. (#21670)
|
2 年之前 |
export
|
596c8e2772
[RLlib] Experimental no-flatten option for actions/prev-actions. (#20918)
|
2 年之前 |
inference_and_serving
|
143d23a278
[RLlib] Issue 20062: Action inference examples missing (#20144)
|
2 年之前 |
models
|
2317c693cf
[RLlib] Use SampleBrach instead of input dict whenever possible (#20746)
|
2 年之前 |
policy
|
daa4304a91
[RLlib] Switch off preprocessors by default for PGTrainer. (#21008)
|
2 年之前 |
serving
|
05a55a9335
[RLlib] Issue 18668: Unity3D env client/server example not working (fix + add to test cases). (#18942)
|
3 年之前 |
simulators
|
e735add268
[RLlib] Integration with SUMO Simulator (#11710)
|
4 年之前 |
tune
|
3408b60d2b
[Release] Refactor User Tests (#20028)
|
3 年之前 |
__init__.py
|
5d7afe8092
[rllib] Try moving RLlib to top level dir (#5324)
|
5 年之前 |
action_masking.py
|
2b3d0c691f
[RLlib] Document and extend action mask example. (#20390)
|
2 年之前 |
attention_net.py
|
f82880eda1
Revert "Revert [RLlib] POC: Deprecate `build_policy` (policy template) for torch only; PPOTorchPolicy (#20061) (#20399)" (#20417)
|
2 年之前 |
attention_net_supervised.py
|
9eba1871bb
[RLlib] Support easy `use_attention=True` flag for using the GTrXL model. (#11698)
|
3 年之前 |
autoregressive_action_dist.py
|
eab9c25856
[RLlib] Better example scripts: Description --no-tune and --local-mode CLI options (autoregressive_action_dist.py) (#17705)
|
3 年之前 |
bare_metal_policy_with_custom_view_reqs.py
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
batch_norm_model.py
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
cartpole_lstm.py
|
53206dd440
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531)
|
3 年之前 |
centralized_critic.py
|
49cd7ea6f9
[RLlib] Trainer sub-class PPO/DDPPO (instead of `build_trainer()`). (#20571)
|
2 年之前 |
centralized_critic_2.py
|
be6db06485
[RLlib] Re-do: Trainer: Support add and delete Policies. (#16569)
|
3 年之前 |
checkpoint_by_custom_criteria.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |
coin_game_env.py
|
be6db06485
[RLlib] Re-do: Trainer: Support add and delete Policies. (#16569)
|
3 年之前 |
complex_struct_space.py
|
daa4304a91
[RLlib] Switch off preprocessors by default for PGTrainer. (#21008)
|
2 年之前 |
compute_adapted_gae_on_postprocess_trajectory.py
|
9acf2f954d
[RLlib] Example containing a proposal for computing an adapted (time-dependent) GAE used by the PPO algorithm (via callback on_postprocess_trajectory) (#20850)
|
2 年之前 |
curriculum_learning.py
|
d89fb82bfb
[RLlib] Add simple curriculum learning API and example script. (#15740)
|
3 年之前 |
custom_env.py
|
8a72824c63
[RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591)
|
3 年之前 |
custom_eval.py
|
60b2219d72
[RLlib] Allow for evaluation to run by `timesteps` (alternative to `episodes`) and add auto-setting to make sure train doesn't ever have to wait for eval (e.g. long episodes) to finish. (#20757)
|
2 年之前 |
custom_experiment.py
|
53206dd440
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531)
|
3 年之前 |
custom_fast_model.py
|
5a313ba3d6
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
3 年之前 |
custom_input_api.py
|
db058d0fb3
[RLlib] Rename `metrics_smoothing_episodes` into `metrics_num_episodes_for_smoothing` for clarity. (#20983)
|
2 年之前 |
custom_keras_model.py
|
ed85f59194
[RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879)
|
3 年之前 |
custom_logger.py
|
026bf01071
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
3 年之前 |
custom_loss.py
|
7eb1a29426
[RLlib] Fix ModelV2 custom metrics for torch. (#16734)
|
3 年之前 |
custom_metrics_and_callbacks.py
|
9c73871da0
[RLlib; Docs overhaul] Docstring cleanup: Evaluation (#19783)
|
3 年之前 |
custom_metrics_and_callbacks_legacy.py
|
c17169dc11
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
4 年之前 |
custom_model_api.py
|
2317c693cf
[RLlib] Use SampleBrach instead of input dict whenever possible (#20746)
|
2 年之前 |
custom_model_loss_and_metrics.py
|
ed85f59194
[RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879)
|
3 年之前 |
custom_observation_filters.py
|
8a066474d4
[RLlib] No Preprocessors; preparatory PR #1 (#18367)
|
3 年之前 |
custom_rnn_model.py
|
53206dd440
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531)
|
3 年之前 |
custom_tf_policy.py
|
b213565783
[RLlib] Fix failing test cases: Soft-deprecate ModelV2.from_batch (in favor of ModelV2.__call__). (#19693)
|
3 年之前 |
custom_torch_policy.py
|
99ae7bae05
[RLlib] JAXPolicy prep. PR #1. (#13077)
|
3 年之前 |
custom_train_fn.py
|
8a72824c63
[RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591)
|
3 年之前 |
custom_vector_env.py
|
0d8fce8fd8
[RLlib] Discussion 2294: Custom vector env example and fix. (#16083)
|
3 年之前 |
deterministic_training.py
|
ad87ddf93e
[rllib] Add deterministic test to gpu (#19306)
|
3 年之前 |
dmlab_watermaze.py
|
60d4d5e1aa
Remove future imports (#6724)
|
4 年之前 |
eager_execution.py
|
b213565783
[RLlib] Fix failing test cases: Soft-deprecate ModelV2.from_batch (in favor of ModelV2.__call__). (#19693)
|
3 年之前 |
env_rendering_and_recording.py
|
60b2219d72
[RLlib] Allow for evaluation to run by `timesteps` (alternative to `episodes`) and add auto-setting to make sure train doesn't ever have to wait for eval (e.g. long episodes) to finish. (#20757)
|
2 年之前 |
fractional_gpus.py
|
7eb1a29426
[RLlib] Fix ModelV2 custom metrics for torch. (#16734)
|
3 年之前 |
hierarchical_training.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
iterated_prisoners_dilemma_env.py
|
be6db06485
[RLlib] Re-do: Trainer: Support add and delete Policies. (#16569)
|
3 年之前 |
lstm_auto_wrapping.py
|
6f342a2221
[RLlib] Preparatory PR for: Documentation on Model Building. (#13260)
|
3 年之前 |
mobilenet_v2_with_lstm.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |
multi_agent_cartpole.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
multi_agent_custom_policy.py
|
f94bd99ce4
[RLlib] Issue 21044: Improve error message for "multiagent" dict checks. (#21448)
|
2 年之前 |
multi_agent_independent_learning.py
|
c01245763e
[RLlib] Revert "Revert "updated pettingzoo wrappers, env versions, urls"" (#21339)
|
2 年之前 |
multi_agent_parameter_sharing.py
|
c01245763e
[RLlib] Revert "Revert "updated pettingzoo wrappers, env versions, urls"" (#21339)
|
2 年之前 |
multi_agent_two_trainers.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
nested_action_spaces.py
|
59f796edf3
[RLlib] Fix crash when using StochasticSampling exploration (most PG-style algos) w/ tf and numpy > 1.19.5 (#18366)
|
3 年之前 |
offline_rl.py
|
60b2219d72
[RLlib] Allow for evaluation to run by `timesteps` (alternative to `episodes`) and add auto-setting to make sure train doesn't ever have to wait for eval (e.g. long episodes) to finish. (#20757)
|
2 年之前 |
parallel_evaluation_and_training.py
|
7517aefe05
[RLlib] Bring back BC and Marwil learning tests. (#21574)
|
2 年之前 |
parametric_actions_cartpole.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |
parametric_actions_cartpole_embeddings_learnt_by_model.py
|
a7f8dc9d77
[RLlib] New and changed version of parametric actions cartpole example + small suggested update in policy_client.py (#15664)
|
3 年之前 |
partial_gpus.py
|
7eb1a29426
[RLlib] Fix ModelV2 custom metrics for torch. (#16734)
|
3 年之前 |
preprocessing_disabled.py
|
61a1274619
[RLlib] No Preprocessors (part 2). (#18468)
|
3 年之前 |
random_parametric_agent.py
|
99a0088233
[RLlib] Unify the way we create local replay buffer for all agents (#19627)
|
3 年之前 |
re3_exploration.py
|
2868d1a2cf
[RLlib] Support for RE3 exploration algorithm (for tf) (#19551)
|
2 年之前 |
recsim_with_slateq.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |
remote_base_env_with_custom_api.py
|
65bd8e29f8
[RLlib] Update a few things to get rid of the `remote_vector_env` deprecation warning. (#20753)
|
2 年之前 |
remote_envs_with_inference_done_on_main_node.py
|
49cd7ea6f9
[RLlib] Trainer sub-class PPO/DDPPO (instead of `build_trainer()`). (#20571)
|
2 年之前 |
restore_1_of_n_agents_from_checkpoint.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
rnnsac_stateless_cartpole.py
|
f7a5fc36eb
[rllib] Give rnnsac_stateless cartpole gpu, increase timeout (#21407)
|
2 年之前 |
rock_paper_scissors_multiagent.py
|
62dbf26394
[RLlib] POC: Run PGTrainer w/o the distr. exec API (Trainer's new training_iteration method). (#20984)
|
2 年之前 |
rollout_worker_custom_workflow.py
|
cdf70c2900
[Tune] Remove legacy resources implementations in Runner and Executor. (#19773)
|
2 年之前 |
saving_experiences.py
|
5a788474aa
[Core] First pass at privatizing non-public Python APIs. (#14607)
|
3 年之前 |
sb2rllib_rllib_example.py
|
489febc6b2
[RLlib] Better example scripts: Description --no-tune and --local-mode CLI options (#17038)
|
3 年之前 |
sb2rllib_sb_example.py
|
55709bac7a
[RLlib] Examples for training, saving, loading, testing an agent with SB & RLlib (#15897)
|
3 年之前 |
self_play_league_based_with_open_spiel.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
self_play_with_open_spiel.py
|
fd13bac9b3
[RLlib] Add `worker` arg (optional) to `policy_mapping_fn`. (#18184)
|
3 年之前 |
sumo_env_local.py
|
d5bfb7b7da
[RLlib] Preparatory PR for multi-agent multi-GPU learner (alpha-star style) #03 (#21652)
|
2 年之前 |
trajectory_view_api.py
|
828f5d26b7
[RLlib] Custom view requirements (e.g. for prev-n-obs) work with `compute_single_action` and `compute_actions_from_input_dict`. (#18921)
|
3 年之前 |
two_step_game.py
|
abd3bef63b
[RLlib] QMIX better defaults + added to CI learning tests (#21332)
|
2 年之前 |
two_trainer_workflow.py
|
d07e50e957
[RLlib] Replay buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (#20552)
|
2 年之前 |
unity3d_env_local.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |
vizdoom_with_attention_net.py
|
d2c755ccef
[RLlib] Examples scripts add argparse help and replace `--torch` with `--framework`. (#15832)
|
3 年之前 |