Ray Tune is a Python library for experiment execution and hyperparameter tuning at any scale.
With Ray Tune, tune your favorite machine learning framework (scikit-learn, XGBoost, StatsForecast, and more) by running state of the art algorithms such as Population Based Training (PBT) and HyperBand. Tune also integrates with a wide range of additional hyperparameter optimization tools, including Ax, Optuna, and more.
Hyperparameter optimization helps you quickly increase your model performance. Take advantage of cutting-edge optimization algorithms to reduce the cost of fine-tuning and optimize training schedules.
Run your existing code in parallel across a cluster with just a few code snippets. Run and scale many trials in parallel, as easily as you coding on your laptop.
Ray Tune makes it easy to speed up hyperparameter searches by scaling out a Ray cluster. Speed up a single trial by running it across compute instances, or run more trials at a time by adding more nodes.
Ray Tune removes boilerplate from your code training workflow, supporting logs results to tools such as Weights & Biases, MLflow, and TensorBoard and multiple storage options for experiment results (NFS, cloud storage), while also being highly customizable.