Abstract
Many existing results on fault-tolerant algorithms focus on the single fault source situation, where a trained network is affected by one kind of weight failure. In fact, a trained network may be affected by multiple kinds of weight failure. This paper first studies how the open weight fault and the multiplicative weight noise degrade the performance of radial basis function (RBF) networks. Afterward, we define the objective function for training fault-tolerant RBF networks. Based on the objective function, we then develop two learning algorithms, one batch mode and one online mode. Besides, the convergent conditions of our online algorithm are investigated. Finally, we develop a formula to estimate the test set error of faulty networks trained from our approach. This formula helps us to optimize some tuning parameters, such as RBF width.
| Original language | English |
|---|---|
| Article number | 7442583 |
| Pages (from-to) | 1360-1372 |
| Journal | IEEE Transactions on Neural Networks and Learning Systems |
| Volume | 28 |
| Issue number | 6 |
| Online published | 28 Mar 2016 |
| DOIs | |
| Publication status | Published - Jun 2017 |
Research Keywords
- Fault tolerance
- prediction error
- radial basis function (RBF)
- regularization
Fingerprint
Dive into the research topics of 'A Regularizer Approach for RBF Networks Under the Concurrent Weight Failure Situation'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver