Improve L2-normalized Softmax with Exponential Moving Average

Xuefei Zhe, Le Ou-Yang, Hong Yan

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

2 Citations (Scopus)

Abstract

In this paper, we propose an effective training method to improve the performance of L2-normalized softmax for convolutional neural networks. Recent studies of deep learning show that by L2-normalizing the input features of softmax, the accuracy of CNN can be increased. Several works proposed novel loss functions based on the L2-normalized softmax. A common property shared by these modified normalized softmax models is that an extra set of parameters is introduced as the class centers. Although the physical meaning of this parameter is clear, few attentions have been paid to how to learn these class centers, which limits futher improvement.
In this paper, we address the problem of learning the class centers in the L2-normalized softmax. By treating the CNN training process as a time series, we propose a novel learning algorithm that combines the generally used gradient descent with the exponential moving average. Extensive experiments show that our model not only achieves better performance but also has a higher tolerance to the imbalance data.
Original languageEnglish
Title of host publication2019 International Joint Conference on Neural Networks (IJCNN)
PublisherIEEE
Number of pages7
ISBN (Electronic)978-1-7281-1985-4
DOIs
Publication statusPublished - Jul 2019
Event2019 International Joint Conference on Neural Networks (IJCNN 2019) - Budapest, Hungary
Duration: 14 Jul 201919 Jul 2019
https://www.ijcnn.org/

Publication series

NameProceedings of the International Joint Conference on Neural Networks
Volume2019-July

Conference

Conference2019 International Joint Conference on Neural Networks (IJCNN 2019)
PlaceHungary
CityBudapest
Period14/07/1919/07/19
Internet address

Bibliographical note

Information for this record is supplemented by the author(s) concerned.

Fingerprint

Dive into the research topics of 'Improve L2-normalized Softmax with Exponential Moving Average'. Together they form a unique fingerprint.

Cite this