You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- `MuJoCo and Atari benchmarks <https://github.com/ray-project/rl-experiments>`__:
19
-
Collection of reasonably optimized Atari and MuJoCo results.
17
+
Collection of tuned hyperparameters sorted by algorithm.
20
18
21
-
Blog Posts
22
-
----------
23
-
24
-
- `Attention Nets and More with RLlib’s Trajectory View API <https://medium.com/distributed-computing-with-ray/attention-nets-and-more-with-rllibs-trajectory-view-api-d326339a6e65>`__:
25
-
This blog describes RLlib's new "trajectory view API" and how it enables implementations of GTrXL (attention net) architectures.
26
-
- `Reinforcement Learning with RLlib in the Unity Game Engine <https://medium.com/distributed-computing-with-ray/reinforcement-learning-with-rllib-in-the-unity-game-engine-1a98080a7c0d>`__:
27
-
A how-to on connecting RLlib with the Unity3D game engine for running visual- and physics-based RL experiments.
28
-
- `Lessons from Implementing 12 Deep RL Algorithms in TF and PyTorch <https://medium.com/distributed-computing-with-ray/lessons-from-implementing-12-deep-rl-algorithms-in-tf-and-pytorch-1b412009297d>`__:
29
-
Discussion on how we ported 12 of RLlib's algorithms from TensorFlow to PyTorch and what we learnt on the way.
This blog post is a brief tutorial on multi-agent RL and its design in RLlib.
32
-
- `Functional RL with Keras and TensorFlow Eager <https://medium.com/riselab/functional-rl-with-keras-and-tensorflow-eager-7973f81d6345>`__:
33
-
Exploration of a functional paradigm for implementing reinforcement learning (RL) algorithms.
34
19
35
20
Environments and Adapters
36
21
-------------------------
@@ -47,7 +32,7 @@ Environments and Adapters
47
32
Custom- and Complex Models
48
33
--------------------------
49
34
50
-
- `Custom Keras model <https://github.com/ray-project/ray/blob/master/rllib/examples/custom_keras_model.py>`__:
35
+
- `Custom Keras model <https://github.com/ray-project/ray/blob/master/rllib/examples/_old_api_stack/custom_keras_model.py>`__:
51
36
Example of using a custom Keras model.
52
37
- `Registering a custom model with supervised loss <https://github.com/ray-project/ray/blob/master/rllib/examples/custom_model_loss_and_metrics.py>`__:
53
38
Example of defining and registering a custom model with a supervised loss.
@@ -83,9 +68,9 @@ Training Workflows
83
68
84
69
Evaluation:
85
70
-----------
86
-
- `Custom evaluation function <https://github.com/ray-project/ray/blob/master/rllib/examples/custom_eval.py>`__:
71
+
- `Custom evaluation function <https://github.com/ray-project/ray/blob/master/rllib/examples/evaluation/custom_evaluation.py>`__:
87
72
Example of how to write a custom evaluation function that is called instead of the default behavior, which is running with the evaluation worker set through n episodes.
88
-
- `Parallel evaluation and training <https://github.com/ray-project/ray/blob/master/rllib/examples/parallel_evaluation_and_training.py>`__:
73
+
- `Parallel evaluation and training <https://github.com/ray-project/ray/blob/master/rllib/examples/evaluation/evaluation_parallel_to_training.py>`__:
89
74
Example showing how the evaluation workers and the "normal" rollout workers can run (to some extend) in parallel to speed up training.
90
75
91
76
@@ -113,23 +98,23 @@ Serving and Offline
113
98
Multi-Agent and Hierarchical
114
99
----------------------------
115
100
116
-
- `Simple independent multi-agent setup vs a PettingZoo env <https://github.com/ray-project/ray/blob/master/rllib/examples/multi_agent_independent_learning.py>`__:
101
+
- `Simple independent multi-agent setup vs a PettingZoo env <https://github.com/ray-project/ray/blob/master/rllib/examples/multi_agent_and_self_play/independent_learning.py>`__:
117
102
Setup RLlib to run any algorithm in (independent) multi-agent mode against a multi-agent environment.
118
-
- `More complex (shared-parameter) multi-agent setup vs a PettingZoo env <https://github.com/ray-project/ray/blob/master/rllib/examples/multi_agent_parameter_sharing.py>`__:
103
+
- `More complex (shared-parameter) multi-agent setup vs a PettingZoo env <https://github.com/ray-project/ray/blob/master/rllib/examples/multi_agent_and_self_play/parameter_sharing.py>`__:
119
104
Setup RLlib to run any algorithm in (shared-parameter) multi-agent mode against a multi-agent environment.
Example of how to handle variable-length or parametric action spaces.
147
-
- `Using the "Repeated" space of RLlib for variable lengths observations <https://github.com/ray-project/ray/blob/master/rllib/examples/complex_struct_space.py>`__:
132
+
- `Using the "Repeated" space of RLlib for variable lengths observations <https://github.com/ray-project/ray/blob/master/rllib/examples/_old_api_stack/complex_struct_space.py>`__:
148
133
How to use RLlib's `Repeated` space to handle variable length observations.
149
134
- `Autoregressive action distribution example <https://github.com/ray-project/ray/blob/master/rllib/examples/autoregressive_action_dist.py>`__:
150
135
Learning with auto-regressive action dependencies (e.g. 2 action components; distribution for 2nd component depends on the 1st component's actually sampled value).
@@ -185,3 +170,18 @@ Community Examples
185
170
Example of training in StarCraft2 maps with RLlib / multi-agent.
Example of optimizing mixed-autonomy traffic simulations with RLlib / multi-agent.
173
+
174
+
175
+
Blog Posts
176
+
----------
177
+
178
+
- `Attention Nets and More with RLlib’s Trajectory View API <https://medium.com/distributed-computing-with-ray/attention-nets-and-more-with-rllibs-trajectory-view-api-d326339a6e65>`__:
179
+
Blog describing RLlib's new "trajectory view API" and how it enables implementations of GTrXL (attention net) architectures.
180
+
- `Reinforcement Learning with RLlib in the Unity Game Engine <https://medium.com/distributed-computing-with-ray/reinforcement-learning-with-rllib-in-the-unity-game-engine-1a98080a7c0d>`__:
181
+
How-To guide about connecting RLlib with the Unity3D game engine for running visual- and physics-based RL experiments.
182
+
- `Lessons from Implementing 12 Deep RL Algorithms in TF and PyTorch <https://medium.com/distributed-computing-with-ray/lessons-from-implementing-12-deep-rl-algorithms-in-tf-and-pytorch-1b412009297d>`__:
183
+
Discussion on how the Ray Team ported 12 of RLlib's algorithms from TensorFlow to PyTorch and the lessons learned.
@@ -145,7 +145,7 @@ the same way as the parent's config.
145
145
Here is an example of how to create an :py:class:`~ray.rllib.utils.replay_buffers.multi_agent_replay_buffer.MultiAgentReplayBuffer` with an alternative underlying :py:class:`~ray.rllib.utils.replay_buffers.replay_buffer.ReplayBuffer`.
146
146
The :py:class:`~ray.rllib.utils.replay_buffers.multi_agent_replay_buffer.MultiAgentReplayBuffer` can stay the same. We only need to specify our own buffer along with a default call argument:
0 commit comments