DeepEutaxy : Diversity in Weight Search Direction for Fixing Deep Learning Model Training through Batch Prioritization
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 1040-1052 |
Number of pages | 13 |
Journal / Publication | IEEE Transactions on Reliability |
Volume | 70 |
Issue number | 3 |
Online published | 18 Jun 2021 |
Publication status | Published - Sept 2021 |
Link(s)
Abstract
Developing a deep learning (DL) based software system is slow. One of the critical issues is to conduct many trials and errors in developing a DL model that usually serves as the major component of such a system. A major reason for this inefficiency is the progress of gradual reduction of the gap between the DL model under training and the ground truths. Prior techniques commonly focus on optimizing such errors after the errors have formed. They are insensitive to how a training dataset is provided to the DL model under training in batches, making their approaches non-proactive to deal with such errors. In this paper, we propose DeepEutaxy, the first work to repair the model convergence problem from the batch prioritization perspective. Our key insight is that increasing the diversity (i.e., dissimilarity) of corresponding weights of complex DL models before and after each training step can make the models learn faster and optimize the training errors quicker. DeepEutaxy first trains a DL model with several epochs for initialization. It then partitions and continually prioritizes the training batches for subsequent training epochs based on our novel notion of diversity between the pair of models before and after training on each batch, capturing the strength of the search direction to deal with training errors impacted by that batch. The experiment on six deep learning models over the MNIST and CIFAR-10 datasets shows that DeepEutaxy can accelerate the convergence of DL models on these two datasets with speedups of 1.75 to 8.45 and 2.67 to 15.15 times with respect to the training and test accuracies, respectively. DeepEutaxy can also be integrated into existing techniques and compare favorably with the prior art in the experiment.
Research Area(s)
- Accuracy improvement, debug, deep learning (DL) models, efficiency, error reduction, fixing, model convergence
Bibliographic Note
Information for this record is supplemented by the author(s) concerned.
Citation Format(s)
DeepEutaxy: Diversity in Weight Search Direction for Fixing Deep Learning Model Training through Batch Prioritization. / Zhang, Hao; Chan, W. K.
In: IEEE Transactions on Reliability, Vol. 70, No. 3, 09.2021, p. 1040-1052.
In: IEEE Transactions on Reliability, Vol. 70, No. 3, 09.2021, p. 1040-1052.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review