README.rst 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328
  1. .. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/ray_header_logo.png
  2. .. image:: https://readthedocs.org/projects/ray/badge/?version=master
  3. :target: http://docs.ray.io/en/master/?badge=master
  4. .. image:: https://img.shields.io/badge/Ray-Join%20Slack-blue
  5. :target: https://forms.gle/9TSdDYUgxYs8SA9e8
  6. .. image:: https://img.shields.io/badge/Discuss-Ask%20Questions-blue
  7. :target: https://discuss.ray.io/
  8. .. image:: https://img.shields.io/twitter/follow/raydistributed.svg?style=social&logo=twitter
  9. :target: https://twitter.com/raydistributed
  10. |
  11. **Ray provides a simple, universal API for building distributed applications.**
  12. Ray is packaged with the following libraries for accelerating machine learning workloads:
  13. - `Tune`_: Scalable Hyperparameter Tuning
  14. - `RLlib`_: Scalable Reinforcement Learning
  15. - `RaySGD <https://docs.ray.io/en/master/raysgd/raysgd.html>`__: Distributed Training Wrappers
  16. - `Datasets`_: Flexible Distributed Data Loading (alpha)
  17. As well as libraries for taking ML and distributed apps to production:
  18. - `Serve`_: Scalable and Programmable Serving
  19. - `Workflows`_: Fast, Durable Application Flows (alpha)
  20. There are also many `community integrations <https://docs.ray.io/en/master/ray-libraries.html>`_ with Ray, including `Dask`_, `MARS`_, `Modin`_, `Horovod`_, `Hugging Face`_, `Scikit-learn`_, and others. Check out the `full list of Ray distributed libraries here <https://docs.ray.io/en/master/ray-libraries.html>`_.
  21. Install Ray with: ``pip install ray``. For nightly wheels, see the
  22. `Installation page <https://docs.ray.io/en/master/installation.html>`__.
  23. .. _`Modin`: https://github.com/modin-project/modin
  24. .. _`Hugging Face`: https://huggingface.co/transformers/main_classes/trainer.html#transformers.Trainer.hyperparameter_search
  25. .. _`MARS`: https://docs.ray.io/en/latest/data/mars-on-ray.html
  26. .. _`Dask`: https://docs.ray.io/en/latest/data/dask-on-ray.html
  27. .. _`Horovod`: https://horovod.readthedocs.io/en/stable/ray_include.html
  28. .. _`Scikit-learn`: joblib.html
  29. .. _`Serve`: https://docs.ray.io/en/master/serve/index.html
  30. .. _`Datasets`: https://docs.ray.io/en/master/data/dataset.html
  31. .. _`Workflows`: https://docs.ray.io/en/master/workflows/concepts.html
  32. Quick Start
  33. -----------
  34. Execute Python functions in parallel.
  35. .. code-block:: python
  36. import ray
  37. ray.init()
  38. @ray.remote
  39. def f(x):
  40. return x * x
  41. futures = [f.remote(i) for i in range(4)]
  42. print(ray.get(futures))
  43. To use Ray's actor model:
  44. .. code-block:: python
  45. import ray
  46. ray.init()
  47. @ray.remote
  48. class Counter(object):
  49. def __init__(self):
  50. self.n = 0
  51. def increment(self):
  52. self.n += 1
  53. def read(self):
  54. return self.n
  55. counters = [Counter.remote() for i in range(4)]
  56. [c.increment.remote() for c in counters]
  57. futures = [c.read.remote() for c in counters]
  58. print(ray.get(futures))
  59. Ray programs can run on a single machine, and can also seamlessly scale to large clusters. To execute the above Ray script in the cloud, just download `this configuration file <https://github.com/ray-project/ray/blob/master/python/ray/autoscaler/aws/example-full.yaml>`__, and run:
  60. ``ray submit [CLUSTER.YAML] example.py --start``
  61. Read more about `launching clusters <https://docs.ray.io/en/master/cluster/index.html>`_.
  62. Tune Quick Start
  63. ----------------
  64. .. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/tune-wide.png
  65. `Tune`_ is a library for hyperparameter tuning at any scale.
  66. - Launch a multi-node distributed hyperparameter sweep in less than 10 lines of code.
  67. - Supports any deep learning framework, including PyTorch, `PyTorch Lightning <https://github.com/williamFalcon/pytorch-lightning>`_, TensorFlow, and Keras.
  68. - Visualize results with `TensorBoard <https://www.tensorflow.org/tensorboard>`__.
  69. - Choose among scalable SOTA algorithms such as `Population Based Training (PBT)`_, `Vizier's Median Stopping Rule`_, `HyperBand/ASHA`_.
  70. - Tune integrates with many optimization libraries such as `Facebook Ax <http://ax.dev>`_, `HyperOpt <https://github.com/hyperopt/hyperopt>`_, and `Bayesian Optimization <https://github.com/fmfn/BayesianOptimization>`_ and enables you to scale them transparently.
  71. To run this example, you will need to install the following:
  72. .. code-block:: bash
  73. $ pip install "ray[tune]"
  74. This example runs a parallel grid search to optimize an example objective function.
  75. .. code-block:: python
  76. from ray import tune
  77. def objective(step, alpha, beta):
  78. return (0.1 + alpha * step / 100)**(-1) + beta * 0.1
  79. def training_function(config):
  80. # Hyperparameters
  81. alpha, beta = config["alpha"], config["beta"]
  82. for step in range(10):
  83. # Iterative training function - can be any arbitrary training procedure.
  84. intermediate_score = objective(step, alpha, beta)
  85. # Feed the score back back to Tune.
  86. tune.report(mean_loss=intermediate_score)
  87. analysis = tune.run(
  88. training_function,
  89. config={
  90. "alpha": tune.grid_search([0.001, 0.01, 0.1]),
  91. "beta": tune.choice([1, 2, 3])
  92. })
  93. print("Best config: ", analysis.get_best_config(metric="mean_loss", mode="min"))
  94. # Get a dataframe for analyzing trial results.
  95. df = analysis.results_df
  96. If TensorBoard is installed, automatically visualize all trial results:
  97. .. code-block:: bash
  98. tensorboard --logdir ~/ray_results
  99. .. _`Tune`: https://docs.ray.io/en/master/tune.html
  100. .. _`Population Based Training (PBT)`: https://docs.ray.io/en/master/tune-schedulers.html#population-based-training-pbt
  101. .. _`Vizier's Median Stopping Rule`: https://docs.ray.io/en/master/tune-schedulers.html#median-stopping-rule
  102. .. _`HyperBand/ASHA`: https://docs.ray.io/en/master/tune-schedulers.html#asynchronous-hyperband
  103. RLlib Quick Start
  104. -----------------
  105. .. image:: https://github.com/ray-project/ray/raw/master/doc/source/images/rllib-wide.jpg
  106. `RLlib`_ is an open-source library for reinforcement learning built on top of Ray that offers both high scalability and a unified API for a variety of applications.
  107. .. code-block:: bash
  108. pip install tensorflow # or tensorflow-gpu
  109. pip install "ray[rllib]"
  110. .. code-block:: python
  111. import gym
  112. from gym.spaces import Discrete, Box
  113. from ray import tune
  114. class SimpleCorridor(gym.Env):
  115. def __init__(self, config):
  116. self.end_pos = config["corridor_length"]
  117. self.cur_pos = 0
  118. self.action_space = Discrete(2)
  119. self.observation_space = Box(0.0, self.end_pos, shape=(1, ))
  120. def reset(self):
  121. self.cur_pos = 0
  122. return [self.cur_pos]
  123. def step(self, action):
  124. if action == 0 and self.cur_pos > 0:
  125. self.cur_pos -= 1
  126. elif action == 1:
  127. self.cur_pos += 1
  128. done = self.cur_pos >= self.end_pos
  129. return [self.cur_pos], 1 if done else 0, done, {}
  130. tune.run(
  131. "PPO",
  132. config={
  133. "env": SimpleCorridor,
  134. "num_workers": 4,
  135. "env_config": {"corridor_length": 5}})
  136. .. _`RLlib`: https://docs.ray.io/en/master/rllib.html
  137. Ray Serve Quick Start
  138. ---------------------
  139. .. image:: https://raw.githubusercontent.com/ray-project/ray/master/doc/source/serve/logo.svg
  140. :width: 400
  141. `Ray Serve`_ is a scalable model-serving library built on Ray. It is:
  142. - Framework Agnostic: Use the same toolkit to serve everything from deep
  143. learning models built with frameworks like PyTorch or Tensorflow & Keras
  144. to Scikit-Learn models or arbitrary business logic.
  145. - Python First: Configure your model serving declaratively in pure Python,
  146. without needing YAMLs or JSON configs.
  147. - Performance Oriented: Turn on batching, pipelining, and GPU acceleration to
  148. increase the throughput of your model.
  149. - Composition Native: Allow you to create "model pipelines" by composing multiple
  150. models together to drive a single prediction.
  151. - Horizontally Scalable: Serve can linearly scale as you add more machines. Enable
  152. your ML-powered service to handle growing traffic.
  153. To run this example, you will need to install the following:
  154. .. code-block:: bash
  155. $ pip install scikit-learn
  156. $ pip install "ray[serve]"
  157. This example runs serves a scikit-learn gradient boosting classifier.
  158. .. code-block:: python
  159. import pickle
  160. import requests
  161. from sklearn.datasets import load_iris
  162. from sklearn.ensemble import GradientBoostingClassifier
  163. from ray import serve
  164. serve.start()
  165. # Train model.
  166. iris_dataset = load_iris()
  167. model = GradientBoostingClassifier()
  168. model.fit(iris_dataset["data"], iris_dataset["target"])
  169. @serve.deployment(route_prefix="/iris")
  170. class BoostingModel:
  171. def __init__(self, model):
  172. self.model = model
  173. self.label_list = iris_dataset["target_names"].tolist()
  174. async def __call__(self, request):
  175. payload = await request.json()["vector"]
  176. print(f"Received flask request with data {payload}")
  177. prediction = self.model.predict([payload])[0]
  178. human_name = self.label_list[prediction]
  179. return {"result": human_name}
  180. # Deploy model.
  181. BoostingModel.deploy(model)
  182. # Query it!
  183. sample_request_input = {"vector": [1.2, 1.0, 1.1, 0.9]}
  184. response = requests.get("http://localhost:8000/iris", json=sample_request_input)
  185. print(response.text)
  186. # Result:
  187. # {
  188. # "result": "versicolor"
  189. # }
  190. .. _`Ray Serve`: https://docs.ray.io/en/master/serve/index.html
  191. More Information
  192. ----------------
  193. - `Documentation`_
  194. - `Tutorial`_
  195. - `Blog`_
  196. - `Ray 1.0 Architecture whitepaper`_ **(new)**
  197. - `Ray Design Patterns`_ **(new)**
  198. - `RLlib paper`_
  199. - `RLlib flow paper`_
  200. - `Tune paper`_
  201. *Older documents:*
  202. - `Ray paper`_
  203. - `Ray HotOS paper`_
  204. .. _`Documentation`: http://docs.ray.io/en/master/index.html
  205. .. _`Tutorial`: https://github.com/ray-project/tutorial
  206. .. _`Blog`: https://medium.com/distributed-computing-with-ray
  207. .. _`Ray 1.0 Architecture whitepaper`: https://docs.google.com/document/d/1lAy0Owi-vPz2jEqBSaHNQcy2IBSDEHyXNOQZlGuj93c/preview
  208. .. _`Ray Design Patterns`: https://docs.google.com/document/d/167rnnDFIVRhHhK4mznEIemOtj63IOhtIPvSYaPgI4Fg/edit
  209. .. _`Ray paper`: https://arxiv.org/abs/1712.05889
  210. .. _`Ray HotOS paper`: https://arxiv.org/abs/1703.03924
  211. .. _`RLlib paper`: https://arxiv.org/abs/1712.09381
  212. .. _`RLlib flow paper`: https://arxiv.org/abs/2011.12719
  213. .. _`Tune paper`: https://arxiv.org/abs/1807.05118
  214. Getting Involved
  215. ----------------
  216. - `Forum`_: For discussions about development, questions about usage, and feature requests.
  217. - `GitHub Issues`_: For reporting bugs.
  218. - `Twitter`_: Follow updates on Twitter.
  219. - `Slack`_: Join our Slack channel.
  220. - `Meetup Group`_: Join our meetup group.
  221. - `StackOverflow`_: For questions about how to use Ray.
  222. .. _`Forum`: https://discuss.ray.io/
  223. .. _`GitHub Issues`: https://github.com/ray-project/ray/issues
  224. .. _`StackOverflow`: https://stackoverflow.com/questions/tagged/ray
  225. .. _`Meetup Group`: https://www.meetup.com/Bay-Area-Ray-Meetup/
  226. .. _`Twitter`: https://twitter.com/raydistributed
  227. .. _`Slack`: https://forms.gle/9TSdDYUgxYs8SA9e8