TY - JOUR
T1 - Information estimators for weighted observations
AU - Hino, Hideitsu
AU - Murata, Noboru
N1 - Funding Information:
The authors would like to express our special thanks to the editor and reviewers, whose valuable comments led to improvements in the paper. Part of this work was supported by MEXT Kakenhi No. 22800067 and No. 25870811 , and Grant-in-Aid for Challenging Exploratory Research No. 23656072 .
PY - 2013/10
Y1 - 2013/10
N2 - The Shannon information content is a valuable numerical characteristic of probability distributions. The problem of estimating the information content from an observed dataset is very important in the fields of statistics, information theory, and machine learning. The contribution of the present paper is in proposing information estimators, and showing some of their applications. When the given data are associated with weights, each datum contributes differently to the empirical average of statistics. The proposed estimators can deal with this kind of weighted data. Similar to other conventional methods, the proposed information estimator contains a parameter to be tuned, and is computationally expensive. To overcome these problems, the proposed estimator is further modified so that it is more computationally efficient and has no tuning parameter. The proposed methods are also extended so as to estimate the cross-entropy, entropy, and Kullback-Leibler divergence. Simple numerical experiments show that the information estimators work properly. Then, the estimators are applied to two specific problems, distribution-preserving data compression, and weight optimization for ensemble regression.
AB - The Shannon information content is a valuable numerical characteristic of probability distributions. The problem of estimating the information content from an observed dataset is very important in the fields of statistics, information theory, and machine learning. The contribution of the present paper is in proposing information estimators, and showing some of their applications. When the given data are associated with weights, each datum contributes differently to the empirical average of statistics. The proposed estimators can deal with this kind of weighted data. Similar to other conventional methods, the proposed information estimator contains a parameter to be tuned, and is computationally expensive. To overcome these problems, the proposed estimator is further modified so that it is more computationally efficient and has no tuning parameter. The proposed methods are also extended so as to estimate the cross-entropy, entropy, and Kullback-Leibler divergence. Simple numerical experiments show that the information estimators work properly. Then, the estimators are applied to two specific problems, distribution-preserving data compression, and weight optimization for ensemble regression.
KW - Entropy estimation
KW - Information estimation
KW - Weighted data
UR - http://www.scopus.com/inward/record.url?scp=84880410580&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84880410580&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2013.06.005
DO - 10.1016/j.neunet.2013.06.005
M3 - Article
C2 - 23859828
AN - SCOPUS:84880410580
SN - 0893-6080
VL - 46
SP - 260
EP - 275
JO - Neural Networks
JF - Neural Networks
ER -