Title
Reduction of Video Compression Artifacts Based on Deep Temporal Networks.
Abstract
It has been shown that deep convolutional neural networks (CNNs) reduce JPEG compression artifacts better than the previous approaches. However, the latest video compression standards have more complex artifacts than the JPEG, including the flickering which is not well reduced by the CNN-based methods developed for still images. Moreover, recent video compression algorithms include in-loop filters which reduce the blocking artifacts, and thus post-processing barely improves the performance. In this paper, we propose a temporal-CNN architecture to reduce the artifacts in video compression standards as well as in JPEG. Specifically, we exploit a simple CNN structure and introduce a new training strategy that captures the temporal correlation of the consecutive frames in videos. The similar patches are aggregated from the neighboring frames by a simple motion search method, and they are fed to the CNN, which further reduces the artifacts. Experiments show that our approach shows improvements over the conventional CNN-based methods with similar complexities for image and video compression standards, such as MPEG-2, AVC, and HEVC, with average PSNR gain of 1.27, 0.47, and 0.23 dB, respectively.
Year
DOI
Venue
2018
10.1109/ACCESS.2018.2876864
IEEE ACCESS
Keywords
Field
DocType
Advanced video coding (AVC),compression artifacts,convolutional neural networks (CNN),high efficiency video coding (HEVC),video compression
Flicker,Computer vision,Computer science,Convolutional neural network,Convolution,Transform coding,Feature extraction,JPEG,Artificial intelligence,Data compression,Encoding (memory),Distributed computing
Journal
Volume
ISSN
Citations 
6
2169-3536
1
PageRank 
References 
Authors
0.35
0
7
Name
Order
Citations
PageRank
Jae Woong Soh1266.76
Jaewoo Park2112.41
Yoonsik Kim3113.87
Byeongyong Ahn4203.37
Hyun-Seung Le591.52
Young-Su Moon6634.54
Nam Ik Cho7712106.98