DOIONLINE

DOIONLINE NO - IJACSCC-IRAJ-DOIONLINE-15529

Publish In
International Journal of Advances in Computer Science and Cloud Computing (IJACSCC)-IJACSCC
Journal Home
Volume Issue
Issue
Volume-7,Issue-1  ( May, 2019 )
Paper Title
Efficient Deep Learning Hyperparameter Tuning on the Cloud
Author Name
Mercy Prasanna Ranjit, Gopinath Ganapathy
Affilition
Advanced Analytics and AI, Microsoft India Corporation Private Limited Bangalore, India Department of Computer Science, Bharathidasan University, Trichy, India
Pages
12-17
Abstract
The paper discusses how we can leverage cloud infrastructure for efficient training and hyperparameter tuning of deep neural networks on the cloud. With the introduction of Horovod framework distributed training of deep learning models has been made trivial on the cloud thereby reducing the time taken to run a single iteration, but the hyperparameter tuning exercise on high dimensional hyperparameter spaces remains a challenge. The paper experiments Bayesian Sequential Gaussian Process Optimization of hyperparameters on the cloud at different levels of concurrency for the warmup runs. Two different distributed hyper-parameter tuning approaches were experimented on the cloud – Training on multiple nodes with higher warm-up concurrency Vs Distributed Training on multiple nodes with Horovod and reduced number of warm-up runs. The results indicate that greater number of warm-up runs results in better exploration of the search space. The hyper parameter choices of every run were optimized using Bayesian optimization technique to take advantage of the learnings from previous runs. The hyper parameter tuning and distributed training with Horovod was performed using Azure Machine Learning Service for Video Activity Recognition problem using LRCN network with transfer learning from Resnet50 backbone. Keywords - Distributed Training, Horovod, Hyperparameter Tuning, Deep Learning, Bayesian Optimization, Automated Machine Learning, Neural Architecture Search
  View Paper