We are excited to announce the release of Databricks Runtime 5.4 ML (Azure | AWS). This release includes two Public Preview features to improve data science productivity, optimized storage in AWS for developing distributed applications, and a number of Python library upgrades.
To get started, you simply select the Databricks Runtime 5.4 ML from the drop-down list when you create a new cluster in Databricks.
Public Preview: Distributed Hyperopt + Automated MLflow Tracking
Hyperparameter tuning is a common technique to optimize machine learning models based on hyperparameters, or parameters that are not learned during model training. However, one major challenge with hyperparameter tuning is that it can be both computationally expensive and slow.
Hyperopt is a popular open-source hyperparameter tuning library with strong community support (600,000+ PyPI downloads, 3300+ stars on Github as of May 2019). Data scientists like Hyperopt for its simplicity and effectiveness. Hyperopt offers two tuning algorithms: Random Search and the Bayesian method Tree of Parzen Estimators, which offers improved compute efficiency compared to a brute force approach such as grid search. However, distributing Hyperopt previously did not work out of the box and required manual setup.
In Databricks Runtime 5.4 ML, we introduce an implementation of Hyperopt powered by Apache Spark. Using a new Trials class
SparkTrials, you can easily distribute a Hyperopt run without making any changes to the current Hyperopt APIs. You simply need to pass in the
SparkTrials class when applying the
hyperopt.fmin function (see the example code below). In addition, all tuning experiments, along with the tuned hyperparameters and targeted metrics, are automatically logged to MLflow in Databricks. With this feature, we aim to improve efficiency, scalability, and simplicity when conducting hyperparameter tuning.
# New SparkTrials class which distributes tuning spark_trials = SparkTrials(parallelism=24) fmin( fn=train, # Method to train and evaluate your model space=search_space, # Defines space of hyperparameters algo=tpe.suggest, # Search algorithm: Tree of Parzen Estimators max_evals=8, # Number of hyperparameter settings to try show_progressbar=False, trials=spark_trials)
At Databricks, we embrace open source communities and APIs. We are working with the Hyperopt community to contribute this Spark-powered implementation to open source Hyperopt. Stay tuned.
Public Preview: MLlib + Automated MLflow Tracking
Databricks Runtime 5.4 and 5.4 ML supports automatic logging of MLflow runs for models trained using PySpark MLlib tuning algorithms CrossValidator and TrainValidationSplit. Before this feature, if you wanted to track PySpark MLlib cross validation or tuning in MLflow, you would have to make explicit MLflow API calls in Databricks notebooks. With MLflow-MLlib integration, when you tune hyperparameters by running
TrainValidationSplit, parameters and evaluation metrics will be automatically logged to MLflow. You can then review how the tuning affects evaluation metrics in MLflow.
Default Optimized FUSE Mount on AWS
The Databricks Runtime has a basic FUSE client for DBFS, a local view of a distributed file system installed on Databricks clusters. This feature has been very popular as it allows local access to remote storage. However, the previous implementation did not allow fast enough data access required for developing distributed deep learning applications.
In Databricks Runtime 5.4, Databricks on AWS now offers an optimized FUSE mount by default. You can now have high-performance data access during training and inference without applying init scripts. Data stored under
dbfs:/ml and accessible locally at
file:/dbfs/ml is now backed by this optimized FUSE mount. If you are running on a Databricks Runtime version prior to 5.4, you can follow our instructions to install a high-performance third-party FUSE client.
We introduced the default optimized FUSE mount for Azure Databricks in Databricks Runtime 5.3. By making it available under the same folder name, we achieved feature parity across Azure and AWS platforms.
In the upcoming months, we plan to enhance the DBFS FUSE client for data scientists who would like flexibility in how they access data.
Display HorovodRunner Training Logs
In the past we introduced HorovodRunner, a simple way to distribute Deep Learning training workloads in Databricks. Databricks Runtime 5.4 ML improves the user experience by displaying HorovodRunner training logs in Databricks Notebook cells. In order to review training logs to better understand optimization progress, you no longer have to look through executor logs under the Spark UI (Azure | AWS). Now, while the HorovodRunner jobs are being executed, training logs will be automatically collected to the driver node and displayed in the notebook cells. You can learn more in our Documentation (Azure | AWS).
Other Library Updates
We updated the following libraries in Databricks Runtime 5.4 ML:
- Pre-installed XGBoost Python package 0.80.
- r-base version bumped from 3.5.2 to 3.6.0.
- We published instructions (Azure | AWS) to install TensorFlow 1.13 and 2.0-alpha to Databricks Runtime ML
En savoir plus
- Databricks Runtime 5.4 ML release notes (Azure | AWS)
- Databricks Documentation – Machine Learning (Azure | AWS)
- Hyperparameter Tuning Overview (blog post and June 20 webinar)
- Spark AI Summit 2019 Talk: “Best Practices for Hyperparameter Tuning with MLflow” by Joseph Bradley
- Spark AI Summit 2019 Talk: “Advanced Hyperparameter Optimization for Deep Learning with MLflow” by Maneesh Bhide
- Try the example notebooks for distributed deep learning training for Azure Databricks and AWS on Databricks Runtime 5.4 ML