一个针对强化学习和深度学习所设计的大规模分布式计算框架。

Eric Liang 6da7eff4b2 [rllib] Properly flatten 2-d observations as input to FCnet (#5733) 5 年之前
.github c33d6662ce Remove Modin from Ray wheels. (#5647) 5 年之前
bazel e2e30ca507 Ray, Tune, and RLlib support for memory, object_store_memory options (#5226) 5 年之前
ci 5f88823c49 [Serve] Rewrite Ray.Serve From Scratch (#5562) 5 年之前
doc 7131166d44 [rllib] Tracing for eager tensorflow policies with `tf.function` (#5705) 5 年之前
docker 93e103135b Update doc versions from 0.8.0.dev3 to 0.8.0.dev4. (#5585) 5 年之前
java 1b880191b0 Replace NotImplementedException with UnsupportedOperationException (#5694) 5 年之前
python d1e4b3677f bump version to 0.7.5 5 年之前
rllib 6da7eff4b2 [rllib] Properly flatten 2-d observations as input to FCnet (#5733) 5 年之前
src f4deecb5ab Fix travis error in direct_actor_transport.cc (#5710) 5 年之前
thirdparty c33d6662ce Remove Modin from Ray wheels. (#5647) 5 年之前
.bazelrc 99a2f9fab3 Scale bazel HTTP timeout by 5x (#5482) 5 年之前
.clang-format 658c14282c Remove legacy Ray code. (#3121) 6 年之前
.gitignore c33d6662ce Remove Modin from Ray wheels. (#5647) 5 年之前
.style.yapf 9a8f29e571 YAPF, take 3 (#2098) 6 年之前
.travis.yml 5f88823c49 [Serve] Rewrite Ray.Serve From Scratch (#5562) 5 年之前
BUILD.bazel b1aadd863b Fix project templates in wheel (#5714) 5 年之前
CONTRIBUTING.rst 2b7b7c7547 Add linting pre-push hook (#5154) 5 年之前
LICENSE 6201a6d1c7 [rllib] add augmented random search (#2714) 6 年之前
README.rst 4d16677a68 Fix PyPI version in readme. (#5662) 5 年之前
WORKSPACE c578be23a5 [Bazel] Modifying WORKSPACE file, so that you can make the project used as a thirdparty project (#4711) 5 年之前
build-docker.sh af463e8bb1 Find bazel even if it isn't in the PATH. (#4729) 5 年之前
build.sh 4bf7de084d Speed up TaskSpecification copy (#5709) 5 年之前
pylintrc 191909dd93 adding pylint (#233) 8 年之前
scripts f1239a7a63 Lint script link broken, also lint filter was broken for generated py files (#4133) 5 年之前
setup_hooks.sh 1a8fa5d2fa Clean up top level Ray dir (#5404) 5 年之前
setup_thirdparty.sh af463e8bb1 Find bazel even if it isn't in the PATH. (#4729) 5 年之前

README.rst

.. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/ray_header_logo.png

.. image:: https://travis-ci.com/ray-project/ray.svg?branch=master
:target: https://travis-ci.com/ray-project/ray

.. image:: https://readthedocs.org/projects/ray/badge/?version=latest
:target: http://ray.readthedocs.io/en/latest/?badge=latest

.. image:: https://img.shields.io/badge/pypi-0.7.4-blue.svg
:target: https://pypi.org/project/ray/

|


**Ray is a fast and simple framework for building and running distributed applications.**

Ray is packaged with the following libraries for accelerating machine learning workloads:

- `Tune`_: Scalable Hyperparameter Tuning
- `RLlib`_: Scalable Reinforcement Learning
- `Distributed Training `__

Install Ray with: ``pip install ray``. For nightly wheels, see the `Installation page `__.

Quick Start
-----------

Execute Python functions in parallel.

.. code-block:: python

import ray
ray.init()

@ray.remote
def f(x):
return x * x

futures = [f.remote(i) for i in range(4)]
print(ray.get(futures))

To use Ray's actor model:

.. code-block:: python


import ray
ray.init()

@ray.remote
class Counter():
def __init__(self):
self.n = 0

def increment(self):
self.n += 1

def read(self):
return self.n

counters = [Counter.remote() for i in range(4)]
[c.increment.remote() for c in counters]
futures = [c.read.remote() for c in counters]
print(ray.get(futures))


Ray programs can run on a single machine, and can also seamlessly scale to large clusters. To execute the above Ray script in the cloud, just download `this configuration file `__, and run:

``ray submit [CLUSTER.YAML] example.py --start``

Read more about `launching clusters `_.

Tune Quick Start
----------------

.. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/tune-wide.png

`Tune`_ is a library for hyperparameter tuning at any scale.

- Launch a multi-node distributed hyperparameter sweep in less than 10 lines of code.
- Supports any deep learning framework, including PyTorch, TensorFlow, and Keras.
- Visualize results with `TensorBoard `__.
- Choose among scalable SOTA algorithms such as `Population Based Training (PBT)`_, `Vizier's Median Stopping Rule`_, `HyperBand/ASHA`_.
- Tune integrates with many optimization libraries such as `Facebook Ax `_, `HyperOpt `_, and `Bayesian Optimization `_ and enables you to scale them transparently.

To run this example, you will need to install the following:

.. code-block:: bash

$ pip install ray torch torchvision filelock


This example runs a parallel grid search to train a Convolutional Neural Network using PyTorch.

.. code-block:: python


import torch.optim as optim
from ray import tune
from ray.tune.examples.mnist_pytorch import (
get_data_loaders, ConvNet, train, test)


def train_mnist(config):
train_loader, test_loader = get_data_loaders()
model = ConvNet()
optimizer = optim.SGD(model.parameters(), lr=config["lr"])
for i in range(10):
train(model, optimizer, train_loader)
acc = test(model, test_loader)
tune.track.log(mean_accuracy=acc)


analysis = tune.run(
train_mnist, config={"lr": tune.grid_search([0.001, 0.01, 0.1])})

print("Best config: ", analysis.get_best_config(metric="mean_accuracy"))

# Get a dataframe for analyzing trial results.
df = analysis.dataframe()

If TensorBoard is installed, automatically visualize all trial results:

.. code-block:: bash

tensorboard --logdir ~/ray_results

.. _`Tune`: https://ray.readthedocs.io/en/latest/tune.html
.. _`Population Based Training (PBT)`: https://ray.readthedocs.io/en/latest/tune-schedulers.html#population-based-training-pbt
.. _`Vizier's Median Stopping Rule`: https://ray.readthedocs.io/en/latest/tune-schedulers.html#median-stopping-rule
.. _`HyperBand/ASHA`: https://ray.readthedocs.io/en/latest/tune-schedulers.html#asynchronous-hyperband

RLlib Quick Start
-----------------

.. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/rllib-wide.jpg

`RLlib`_ is an open-source library for reinforcement learning built on top of Ray that offers both high scalability and a unified API for a variety of applications.

.. code-block:: bash

pip install tensorflow # or tensorflow-gpu
pip install ray[rllib] # also recommended: ray[debug]

.. code-block:: python

import gym
from gym.spaces import Discrete, Box
from ray import tune

class SimpleCorridor(gym.Env):
def __init__(self, config):
self.end_pos = config["corridor_length"]
self.cur_pos = 0
self.action_space = Discrete(2)
self.observation_space = Box(0.0, self.end_pos, shape=(1, ))

def reset(self):
self.cur_pos = 0
return [self.cur_pos]

def step(self, action):
if action == 0 and self.cur_pos > 0:
self.cur_pos -= 1
elif action == 1:
self.cur_pos += 1
done = self.cur_pos >= self.end_pos
return [self.cur_pos], 1 if done else 0, done, {}

tune.run(
"PPO",
config={
"env": SimpleCorridor,
"num_workers": 4,
"env_config": {"corridor_length": 5}})

.. _`RLlib`: https://ray.readthedocs.io/en/latest/rllib.html


More Information
----------------

- `Documentation`_
- `Tutorial`_
- `Blog`_
- `Ray paper`_
- `Ray HotOS paper`_
- `RLlib paper`_
- `Tune paper`_

.. _`Documentation`: http://ray.readthedocs.io/en/latest/index.html
.. _`Tutorial`: https://github.com/ray-project/tutorial
.. _`Blog`: https://ray-project.github.io/
.. _`Ray paper`: https://arxiv.org/abs/1712.05889
.. _`Ray HotOS paper`: https://arxiv.org/abs/1703.03924
.. _`RLlib paper`: https://arxiv.org/abs/1712.09381
.. _`Tune paper`: https://arxiv.org/abs/1807.05118

Getting Involved
----------------

- `ray-dev@googlegroups.com`_: For discussions about development or any general
questions.
- `StackOverflow`_: For questions about how to use Ray.
- `GitHub Issues`_: For reporting bugs and feature requests.
- `Pull Requests`_: For submitting code contributions.

.. _`ray-dev@googlegroups.com`: https://groups.google.com/forum/#!forum/ray-dev
.. _`GitHub Issues`: https://github.com/ray-project/ray/issues
.. _`StackOverflow`: https://stackoverflow.com/questions/tagged/ray
.. _`Pull Requests`: https://github.com/ray-project/ray/pulls