TY - GEN
T1 - Holistic feature extraction for automatic image annotation
AU - Sarin, Supheakmungkol
AU - Fahrmair, Michael
AU - Wagner, Matthias
AU - Kameyama, Wataru
PY - 2011
Y1 - 2011
N2 - Automating the annotation process of digital images is a crucial step towards efficient and effective management of this increasingly high volume of content. It is, nevertheless, an extremely challenging task for the research community. One of the main bottle necks is the lack of integrity and diversity of features. We solve this problem by proposing to utilize 43 image features that cover the holistic content of the image from global to subject, background, and scene. In our approach, saliency regions and background are separated without prior knowledge. Each of them together with the whole image is treated independently for feature extraction. Extensive experiments were designed to show the efficiency and effectiveness of our approach. We chose two publicly available datasets manually annotated and with the diverse nature of images for our experiments, namely, the Corel5k and ESP Game datasets. They contain 5,000 images with 260 keywords and 20,770 images with 268 keywords, respectively. Through empirical experiments, it is confirmed that by using our features with the state-of-the-art technique, we achieve superior performance in many metrics, particularly in auto-annotation.
AB - Automating the annotation process of digital images is a crucial step towards efficient and effective management of this increasingly high volume of content. It is, nevertheless, an extremely challenging task for the research community. One of the main bottle necks is the lack of integrity and diversity of features. We solve this problem by proposing to utilize 43 image features that cover the holistic content of the image from global to subject, background, and scene. In our approach, saliency regions and background are separated without prior knowledge. Each of them together with the whole image is treated independently for feature extraction. Extensive experiments were designed to show the efficiency and effectiveness of our approach. We chose two publicly available datasets manually annotated and with the diverse nature of images for our experiments, namely, the Corel5k and ESP Game datasets. They contain 5,000 images with 260 keywords and 20,770 images with 268 keywords, respectively. Through empirical experiments, it is confirmed that by using our features with the state-of-the-art technique, we achieve superior performance in many metrics, particularly in auto-annotation.
KW - Automatic image annotation
KW - Background
KW - Holistic feature extraction
KW - K nearest neighbours
KW - Saliency regions
UR - http://www.scopus.com/inward/record.url?scp=80052595697&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=80052595697&partnerID=8YFLogxK
U2 - 10.1109/MUE.2011.22
DO - 10.1109/MUE.2011.22
M3 - Conference contribution
AN - SCOPUS:80052595697
SN - 9780769544700
T3 - Proceedings of the 2011 5th FTRA International Conference on Multimedia and Ubiquitous Engineering, MUE 2011
SP - 59
EP - 66
BT - Proceedings of the 2011 5th FTRA International Conference on Multimedia and Ubiquitous Engineering, MUE 2011
T2 - 2011 5th FTRA International Conference on Multimedia and Ubiquitous Engineering, MUE 2011
Y2 - 28 June 2011 through 30 June 2011
ER -