TY - JOUR
T1 - Self-supervised learning for visual summary identification in scientific publications
AU - Yamamoto, Shintaro
AU - Lauscher, Anne
AU - Ponzetto, Simone Paolo
AU - Glavaš, Goran
AU - Morishima, Shigeo
N1 - Funding Information:
This work was supported by the Program for Leading Graduate Schools,”Graduate Program for Embodiment Informatics” of the Ministry of Education, Culture, Sports, Science and Technology (MEXT) of Japan, and JST ACCEL (JPMJAC1602). Computational resource of AI Bridging Cloud Infrastructure (ABCI) provided by National Institute of Advanced Industrial Science and Technology (AIST) was used.
Funding Information:
This work was supported by the Program for Leading Graduate Schools, ”Graduate Program for Embodiment Informatics” of the Ministry of Education, Culture, Sports, Science and Technology ?MEXT) of Japan, and JST ACCEL ?JPMJAC1602). Computational resource of AI Bridging Cloud Infrastructure ? ABCI) provided by National Institute of Advanced Industrial Science and Technology ? AIST) was used.
Publisher Copyright:
© 2021 Copyright for this paper by its authors.
PY - 2021
Y1 - 2021
N2 - Providing visual summaries of scientific publications can increase information access for readers and thereby help deal with the exponential growth in the number of scientific publications. Nonetheless, efforts in providing visual publication summaries have been few and far apart, primarily focusing on the biomedical domain. This is primarily because of the limited availability of annotated gold standards, which hampers the application of robust and high-performing supervised learning techniques. To address these problems we create a new benchmark dataset for selecting figures to serve as visual summaries of publications based on their abstracts, covering several domains in computer science. Moreover, we develop a self-supervised learning approach, based on heuristic matching of inline references to figures with figure captions. Experiments in both biomedical and computer science domains show that our model is able to outperform the state of the art despite being self-supervised and therefore not relying on any annotated training data.
AB - Providing visual summaries of scientific publications can increase information access for readers and thereby help deal with the exponential growth in the number of scientific publications. Nonetheless, efforts in providing visual publication summaries have been few and far apart, primarily focusing on the biomedical domain. This is primarily because of the limited availability of annotated gold standards, which hampers the application of robust and high-performing supervised learning techniques. To address these problems we create a new benchmark dataset for selecting figures to serve as visual summaries of publications based on their abstracts, covering several domains in computer science. Moreover, we develop a self-supervised learning approach, based on heuristic matching of inline references to figures with figure captions. Experiments in both biomedical and computer science domains show that our model is able to outperform the state of the art despite being self-supervised and therefore not relying on any annotated training data.
KW - Multimodal retrieval
KW - Scientific publication mining
KW - Visual summary identification
UR - http://www.scopus.com/inward/record.url?scp=85104043362&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85104043362&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85104043362
SN - 1613-0073
VL - 2847
SP - 5
EP - 19
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - 11th International Workshop on Bibliometric-Enhanced Information Retrieval, BIR 2021
Y2 - 1 April 2021
ER -