TY - GEN
T1 - Tsallis Entropy Based Labelling
AU - Goto, Kentaro
AU - Uchida, Masato
N1 - Funding Information:
ACKNOWLEDGMENT This work was supported in part by the Japan Society for the Promotion of Science through Grants-in-Aid for Scientific Research (C) (20K11800).
Publisher Copyright:
© 2020 IEEE.
PY - 2020/12
Y1 - 2020/12
N2 - In the field of supervised classification, the quality of training data is an essential aspect of accurate learning, along with the selection of a learning algorithm or parameters optimisation. To improve the quality of training data, it is necessary to reflect an annotator's idea on to which class any given instance belongs in the form of a label as flexibly and accurately as possible. However, in conventional problem settings used in machine learning, the number of labels per instance is uniformly fixed at a certain value, and it is implicitly assumed that annotators provide labels under such a constraint. Thus, in this study, we propose an annotation framework; Tsallis entropy based labelling, which models a method that dynamically selects the number of labels for every single given instance depending on the uncertainty regarding the class to which each instance belongs. Using the proposed framework, an annotator's instinctive uncertainty about classification task is expressed based on the Tsallis entropy and Tsallis self-information. In addition, the proposed framework has a well-organised mathematical structure that includes some typical annotation models. We conduct an experiment to evaluate the proposed framework and demonstrate that it outperforms another annotation model in terms of the labels accuracy; In the comparison model, the number of labels per instance is deliberately set at a fixed value for all instances. Moreover, we exemplify that the conventional single labelling scheme is not always the best option, which reveals the fact that increasing the number of labels per instance does not necessarily hinder the labels accuracy.
AB - In the field of supervised classification, the quality of training data is an essential aspect of accurate learning, along with the selection of a learning algorithm or parameters optimisation. To improve the quality of training data, it is necessary to reflect an annotator's idea on to which class any given instance belongs in the form of a label as flexibly and accurately as possible. However, in conventional problem settings used in machine learning, the number of labels per instance is uniformly fixed at a certain value, and it is implicitly assumed that annotators provide labels under such a constraint. Thus, in this study, we propose an annotation framework; Tsallis entropy based labelling, which models a method that dynamically selects the number of labels for every single given instance depending on the uncertainty regarding the class to which each instance belongs. Using the proposed framework, an annotator's instinctive uncertainty about classification task is expressed based on the Tsallis entropy and Tsallis self-information. In addition, the proposed framework has a well-organised mathematical structure that includes some typical annotation models. We conduct an experiment to evaluate the proposed framework and demonstrate that it outperforms another annotation model in terms of the labels accuracy; In the comparison model, the number of labels per instance is deliberately set at a fixed value for all instances. Moreover, we exemplify that the conventional single labelling scheme is not always the best option, which reveals the fact that increasing the number of labels per instance does not necessarily hinder the labels accuracy.
UR - http://www.scopus.com/inward/record.url?scp=85102494619&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85102494619&partnerID=8YFLogxK
U2 - 10.1109/ICMLA51294.2020.00015
DO - 10.1109/ICMLA51294.2020.00015
M3 - Conference contribution
AN - SCOPUS:85102494619
T3 - Proceedings - 19th IEEE International Conference on Machine Learning and Applications, ICMLA 2020
SP - 33
EP - 40
BT - Proceedings - 19th IEEE International Conference on Machine Learning and Applications, ICMLA 2020
A2 - Wani, M. Arif
A2 - Luo, Feng
A2 - Li, Xiaolin
A2 - Dou, Dejing
A2 - Bonchi, Francesco
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 19th IEEE International Conference on Machine Learning and Applications, ICMLA 2020
Y2 - 14 December 2020 through 17 December 2020
ER -