123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960 |
- .. _replay-buffer-api-reference-docs:
- Replay Buffer API
- =================
- The following classes don't take into account the separation of experiences from different policies, multi-agent replay buffers will be explained further below.
- Replay Buffer Base Classes
- --------------------------
- .. currentmodule:: ray.rllib.utils.replay_buffers
- .. autosummary::
- :toctree: doc/
- ~replay_buffer.StorageUnit
- ~replay_buffer.ReplayBuffer
- ~prioritized_replay_buffer.PrioritizedReplayBuffer
- ~reservoir_replay_buffer.ReservoirReplayBuffer
- Public Methods
- --------------
- .. currentmodule:: ray.rllib.utils.replay_buffers.replay_buffer
- .. autosummary::
- :toctree: doc/
- ~ReplayBuffer.sample
- ~ReplayBuffer.add
- ~ReplayBuffer.get_state
- ~ReplayBuffer.set_state
- Multi Agent Buffers
- -------------------
- The following classes use the above, "single-agent", buffers as underlying buffers to facilitate splitting up experiences between the different agents' policies.
- In multi-agent RL, more than one agent exists in the environment and not all of these agents may utilize the same policy (mapping M agents to N policies, where M <= N).
- This leads to the need for MultiAgentReplayBuffers that store the experiences of different policies separately.
- .. currentmodule:: ray.rllib.utils.replay_buffers
- .. autosummary::
- :toctree: doc/
- ~multi_agent_replay_buffer.MultiAgentReplayBuffer
- ~multi_agent_prioritized_replay_buffer.MultiAgentPrioritizedReplayBuffer
- Utility Methods
- ---------------
- .. autosummary::
- :toctree: doc/
- ~utils.update_priorities_in_replay_buffer
- ~utils.sample_min_n_steps_from_buffer
|