TY - GEN
T1 - A selective fusion module for video super resolution with recurrent architecture
AU - Gong, Zichen
AU - Hori, Toshiya
AU - Watanabe, Hiroshi
AU - Ikai, Tomohiro
AU - Chujoh, Takeshi
AU - Sasaki, Eiichi
AU - Ito, Norio
N1 - Publisher Copyright:
© 2020 SPIE CCC.
PY - 2020
Y1 - 2020
N2 - As an important subtask of video restoration, video super-resolution has attracted a lot of attention in the community as it can eventually promote a wide range of technologies, e.g., video transmission system. Recent video super resolution model1 achieves cutting-edge performance. It efficiently utilizes recurrent architecture with neural networks to gradually aggregate details from previous frames. Nevertheless, this method faces a serious drawback that it is sensitive to occlusion, blur, and large motion changes since it only takes the previous generated output as recurrent input for the super resolution model. This will lead to undesirable rapid information loss during the recurrently generating process, and performance will therefore be dramatically decreased. Our works focus on addressing the issue of rapid information loss in video super-resolution model with recurrent architecture. By producing attention maps through selective fusion module, the recurrent model can adaptively aggregate necessary details across all previously generated high-resolution (HR) frames according to their informativeness. The proposed method is useful for preserving high frequency details collected progressively from each frame while being capable of removing noisy artifacts. This significantly improves the average quality of the super resolution video.
AB - As an important subtask of video restoration, video super-resolution has attracted a lot of attention in the community as it can eventually promote a wide range of technologies, e.g., video transmission system. Recent video super resolution model1 achieves cutting-edge performance. It efficiently utilizes recurrent architecture with neural networks to gradually aggregate details from previous frames. Nevertheless, this method faces a serious drawback that it is sensitive to occlusion, blur, and large motion changes since it only takes the previous generated output as recurrent input for the super resolution model. This will lead to undesirable rapid information loss during the recurrently generating process, and performance will therefore be dramatically decreased. Our works focus on addressing the issue of rapid information loss in video super-resolution model with recurrent architecture. By producing attention maps through selective fusion module, the recurrent model can adaptively aggregate necessary details across all previously generated high-resolution (HR) frames according to their informativeness. The proposed method is useful for preserving high frequency details collected progressively from each frame while being capable of removing noisy artifacts. This significantly improves the average quality of the super resolution video.
KW - Recurrent networks
KW - Selective fusion
KW - Video super resolution
KW - Video transmission system
UR - http://www.scopus.com/inward/record.url?scp=85086627895&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85086627895&partnerID=8YFLogxK
U2 - 10.1117/12.2566219
DO - 10.1117/12.2566219
M3 - Conference contribution
AN - SCOPUS:85086627895
T3 - Proceedings of SPIE - The International Society for Optical Engineering
BT - International Workshop on Advanced Imaging Technology, IWAIT 2020
A2 - Lau, Phooi Yee
A2 - Shobri, Mohammad
PB - SPIE
T2 - International Workshop on Advanced Imaging Technology, IWAIT 2020
Y2 - 5 January 2020 through 7 January 2020
ER -