Neural Network-Based Mapping Mining of Image Style Transfer in Big Data Systems

Hong An Li, Qiaoxue Zheng, Xin Qi*, Wenjing Yan, Zheng Wen, Na Li, Chu Tang

*この研究の対応する著者

研究成果: Article査読

12 被引用数 (Scopus)

抄録

Image style transfer can realize the mutual transfer between different styles of images and is an essential application for big data systems. The use of neural network-based image data mining technology can effectively mine the useful information in the image and improve the utilization rate of information. However, when using the deep learning method to transform the image style, the content information is often lost. To address this problem, this paper introduces L1 loss on the basis of the VGG-19 network to reduce the difference between image style and content and adds perceptual loss to calculate the semantic information of the feature map to improve the model's perceptual ability. Experiments show that the proposal in this paper improves the ability of style transfer, while maintaining image content information. The stylization of the improved model can better meet people's requirements for stylization, and the evaluation indexes of structural similarity, cosine similarity, and mutual information value have increased by 0.323%, 0.094%, and 3.591%, respectively.

本文言語English
論文番号8387382
ジャーナルComputational Intelligence and Neuroscience
2021
DOI
出版ステータスPublished - 2021

ASJC Scopus subject areas

  • コンピュータサイエンス一般
  • 神経科学一般
  • 数学一般

フィンガープリント

「Neural Network-Based Mapping Mining of Image Style Transfer in Big Data Systems」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル