DeepEutaxy : Diversity in Weight Search Direction for Fixing Deep Learning Model Training through Batch Prioritization

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

View graph of relations

Related Research Unit(s)

Detail(s)

Original languageEnglish
Pages (from-to)1040-1052
Number of pages13
Journal / PublicationIEEE Transactions on Reliability
Volume70
Issue number3
Online published18 Jun 2021
Publication statusPublished - Sept 2021

Abstract

Developing a deep learning (DL) based software system is slow. One of the critical issues is to conduct many trials and errors in developing a DL model that usually serves as the major component of such a system. A major reason for this inefficiency is the progress of gradual reduction of the gap between the DL model under training and the ground truths. Prior techniques commonly focus on optimizing such errors after the errors have formed. They are insensitive to how a training dataset is provided to the DL model under training in batches, making their approaches non-proactive to deal with such errors. In this paper, we propose DeepEutaxy, the first work to repair the model convergence problem from the batch prioritization perspective. Our key insight is that increasing the diversity (i.e., dissimilarity) of corresponding weights of complex DL models before and after each training step can make the models learn faster and optimize the training errors quicker. DeepEutaxy first trains a DL model with several epochs for initialization. It then partitions and continually prioritizes the training batches for subsequent training epochs based on our novel notion of diversity between the pair of models before and after training on each batch, capturing the strength of the search direction to deal with training errors impacted by that batch. The experiment on six deep learning models over the MNIST and CIFAR-10 datasets shows that DeepEutaxy can accelerate the convergence of DL models on these two datasets with speedups of 1.75 to 8.45 and 2.67 to 15.15 times with respect to the training and test accuracies, respectively. DeepEutaxy can also be integrated into existing techniques and compare favorably with the prior art in the experiment.

Research Area(s)

  • Accuracy improvement, debug, deep learning (DL) models, efficiency, error reduction, fixing, model convergence

Bibliographic Note

Information for this record is supplemented by the author(s) concerned.