Abstract
Identifying reliable information from the information ocean is a natural talent of human being, which can be hardly formalized by machines. In this paper, we show how to measure the uncertainty of each sample in terms of their similarity. For applications under complex uncertainty, it is desirable that we provide a systematic way for users to identify what are reliable information. In our approach, Posterior self-information for different kinds of data sets is defined according to the Shannon entropy theory. To improve the performance of three learning algorithms under uncertainty, this method is introduced to them by weighting their loss function. On both simulation data set and several public data sets, we find our method leads to significant improvements when uncertainty is high, and preserves original performance when uncertainty can be ignored.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of 2016 SAI Computing Conference, SAI 2016 |
| Publisher | IEEE |
| Pages | 182-187 |
| ISBN (Print) | 9781467384605 |
| DOIs | |
| Publication status | Published - 29 Aug 2016 |
| Event | 2016 SAI Computing Conference, SAI 2016 - London, United Kingdom Duration: 13 Jul 2016 → 15 Jul 2016 http://www.saiconference.com |
Conference
| Conference | 2016 SAI Computing Conference, SAI 2016 |
|---|---|
| Place | United Kingdom |
| City | London |
| Period | 13/07/16 → 15/07/16 |
| Internet address |
Research Keywords
- Distance Metric Learning
- Outlier
- Posterior self-information
- Uncertainty Measurement
Fingerprint
Dive into the research topics of 'Posterior self-information based uncertainty measurement for data classification and learning'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver