Hyperparameter tuning is a common technique to optimize machine learning models based on hyperparameters, or configurations that are not learned during model training. Tuning these configurations can dramatically improve model performance. However, hyperparameter tuning can be computationally expensive, slow, and unintuitive even for experts.
Databricks Runtime 5.4 and 5.4 ML (Azure | AWS) introduce new features which help to scale and simplify hyperparameter tuning. These features support tuning for ML in Python, with an emphasis on scalability via Apache Spark and automated tracking via MLflow.
MLflow: tracking tuning workflows
Hyperparameter tuning creates complex workflows involving testing many hyperparameter settings, generating lots of models, and iterating on an ML pipeline. To simplify tracking and reproducibility for tuning workflows, we use MLflow, an open source platform to help manage the complete machine learning lifecycle. Learn more about MLflow in the MLflow docs and the recent Spark+AI Summit 2019 talks on MLflow.
Our integrations encourage some best practices for organizing runs and tracking for hyperparameter tuning. At a high level, we organize runs as follows, matching the structure used by tuning itself:
Tuning | MLflow runs | MLflow logging |
Hyperparameter tuning algorithm | Parent run | Metadata, e.g., numFolds for CrossValidator |
Fit & evaluate model with hyperparameter setting #1 | Child run 1 | Hyperparameters #1, evaluation metric #1 |
Fit & evaluate model with hyperparameter setting #2 | Child run 2 | Hyperparameters #2, evaluation metric #2 |
... | ... | ... |
To learn more, check out this talk on “Best Practices for Hyperparameter Tuning with MLflow” from the Spark+AI Summit 2019.
Managed MLflow is now generally available on Databricks, and the two integrations we discuss next leverage managed MLflow by default when the MLflow library is installed on the cluster.
Apache Spark MLlib + MLflow integration
Apache Spark MLlib users often tune hyperparameters using MLlib’s built-in tools CrossValidator
and TrainValidationSplit
. These use grid search to try out a user-specified set of hyperparameter values; see the Spark docs on tuning for more info.
Databricks Runtime 5.3 and 5.3 ML and above support automatic MLflow tracking for MLlib tuning in Python.
With this feature, PySpark CrossValidator
and TrainValidationSplit
will automatically log to MLflow, organizing runs in a hierarchy and logging hyperparameters and the evaluation metric. For example, calling CrossValidator.fit()
will log one parent run. Under this run, CrossValidator
will log one child run for each hyperparameter setting, and each of those child runs will include the hyperparameter setting and the evaluation metric. Comparing these runs in the MLflow UI helps with visualizing the effect of tuning each hyperparameter.
https://www.youtube.com/watch?v=DFn3hS-s7OA
In Databricks Runtime 5.3 and 5.3 ML, automatic tracking is not enabled by default. To turn automatic tracking on, set the Spark Configuration spark.databricks.mlflow.trackMLlib.enabled
to “true”. With the 5.4 releases, automatic tracking is enabled by default.
Check out the docs (AWS | Azure) to get started!
Distributed Hyperopt + MLflow integration
Hyperopt is a popular open-source hyperparameter tuning library with strong community support (600,000+ PyPI downloads, 3300+ stars on Github as of May 2019). Data scientists use Hyperopt for its simplicity and effectiveness. Hyperopt offers two tuning algorithms: Random Search and the Bayesian method Tree of Parzen Estimators, which offers improved compute efficiency compared to a brute force approach such as grid search. However, distributing Hyperopt previously did not work out of the box and required manual setup.
In Databricks Runtime 5.4 ML, we introduce an implementation of Hyperopt powered by Apache Spark. Using a new Trials
class SparkTrials
, you can easily distribute a Hyperopt run without making any changes to the current Hyperopt APIs. You simply need to pass in the SparkTrials
class when applying the hyperopt.fmin()
function (see the example code below). In addition, all tuning experiments, along with their hyperparameters and evaluation metrics, are automatically logged to MLflow in Databricks. With this feature, we aim to improve efficiency, scalability, and simplicity for hyperparameter tuning workflows.
Check out the docs (Azure | AWS) to get started!
# New SparkTrials class which distributes tuning
<b>spark_trials = SparkTrials(parallelism=24)</b>
fmin(
fn=train, # Method to train and evaluate your model
space=search_space, # Defines space of hyperparameters
algo=tpe.suggest, # Search algorithm: Tree of Parzen Estimators
max_evals=8, # Number of hyperparameter settings to try
show_progressbar=False,
trials=<b>spark_trials</b>)
The results can be visualized using tools such as parallel coordinates plots. In the plot below, we can see that the Deep Learning models with the best (lowest) losses were trained using medium to large batch sizes, small to medium learning rates, and a variety of momentum settings. Note that this plot was made by hand via plotly
, but MLflow will provide native support for parallel coordinates plots in the near future.
At Databricks, we embrace open source communities and APIs. We are working with the Hyperopt community to contribute this Spark-powered implementation to open source Hyperopt. Stay tuned.
Get started!
To learn more about hyperparameter tuning in general:
- Don’t miss our upcoming webinar Automated Hyperparameter Tuning, Scaling, and Tracking on Databricks for a deeper dive and live demos – on Thursday June 20th.
- Check out these talks from the Spark+AI Summit 2019:
- “Best Practices for Hyperparameter Tuning with MLflow” by Joseph Bradley
- “Advanced Hyperparameter Optimization for Deep Learning with MLflow” by Maneesh Bhide
To learn more about MLflow, check out these resources:
- MLflow website
- MLflow documentation
- Spark+AI Summit 2019 talks on MLflow, including the MLflow keynote
To start using these specific features, check out the following doc pages and their embedded example notebooks. Try them out with the new Databricks Runtime 5.4 ML release.
- For MLlib use cases, look at the MLlib + Automated MLflow Tracking docs (AWS | Azure).
- For single-machine Python ML use cases (e.g., scikit-learn, single-machine TensorFlow), look at the Distributed Hyperopt + Automated MLflow Tracking docs (Azure | AWS).
- For non-MLlib distributed ML use cases (e.g., HorovodRunner), look at MLflow’s examples on adding tracking to Hyperopt and other tools.