TY - GEN
T1 - Performance Analysis of Generated Predictive Frames Using PredNet Bi-directionally
AU - Sakama, Kanato
AU - Sekiguchi, Shunichi
AU - Kameyama, Wataru
N1 - Publisher Copyright:
© 2022 SPIE.
PY - 2022
Y1 - 2022
N2 - For generating motion-compensated predictive frames, which is one of the video coding processes, there has been a lot of studies on using DNN without using motion vectors. Conventional methods of generating motion-compensated predictive frames using DNN use only the source frames for prediction in the forward direction. However, in the ever-standardized video coding schemes, it has been confirmed that the bi-directional prediction, e.g., B-picture, improves coding efficiency. Thus, for generating motion-compensated predictive frames to be used in video coding, we propose to apply PredNet bidirectionally, that is a future frame generation model using DNN based on the prediction process of visual input stimuli in brain. In this paper, the performance of the predictive frames generated by the proposed method is evaluated by using MSE and SSIM compared with the prediction accuracy applying PredNet only to the forward direction. In addition, we also investigate whether the prediction accuracy of the predicted frames can be improved by increasing the amount of training frames in videos chosen from YouTube-8M. The results show the effectiveness of the proposed method in terms of less prediction error compared with the forward-only PredNet, as well as the performance increasing by more training data.
AB - For generating motion-compensated predictive frames, which is one of the video coding processes, there has been a lot of studies on using DNN without using motion vectors. Conventional methods of generating motion-compensated predictive frames using DNN use only the source frames for prediction in the forward direction. However, in the ever-standardized video coding schemes, it has been confirmed that the bi-directional prediction, e.g., B-picture, improves coding efficiency. Thus, for generating motion-compensated predictive frames to be used in video coding, we propose to apply PredNet bidirectionally, that is a future frame generation model using DNN based on the prediction process of visual input stimuli in brain. In this paper, the performance of the predictive frames generated by the proposed method is evaluated by using MSE and SSIM compared with the prediction accuracy applying PredNet only to the forward direction. In addition, we also investigate whether the prediction accuracy of the predicted frames can be improved by increasing the amount of training frames in videos chosen from YouTube-8M. The results show the effectiveness of the proposed method in terms of less prediction error compared with the forward-only PredNet, as well as the performance increasing by more training data.
KW - Bi-directional prediction
KW - DNN
KW - Motion compensation
KW - PredNet
KW - Video coding
UR - http://www.scopus.com/inward/record.url?scp=85131804130&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85131804130&partnerID=8YFLogxK
U2 - 10.1117/12.2625172
DO - 10.1117/12.2625172
M3 - Conference contribution
AN - SCOPUS:85131804130
T3 - Proceedings of SPIE - The International Society for Optical Engineering
BT - International Workshop on Advanced Imaging Technology, IWAIT 2022
A2 - Nakajima, Masayuki
A2 - Muramatsu, Shogo
A2 - Kim, Jae-Gon
A2 - Guo, Jing-Ming
A2 - Kemao, Qian
PB - SPIE
T2 - 2022 International Workshop on Advanced Imaging Technology, IWAIT 2022
Y2 - 4 January 2022 through 6 January 2022
ER -