Review: VRCNN — Variable-Filter-Size Residue-Learning CNN (Codec Filtering)
Achieved 4.6% Average BD-Rate Reduction Compared to HEVC Baseline, Outperforms ARCNN and VDSR
In this story, Variable-Filter-Size Residue-Learning CNN (VRCNN), by University of Science and Technology of China, is reviewed. VRCNN is used as in-loop or out-of-loop filtering to improve the image quality after decoding so as to increase the coding efficiency. It is published in 2017 MMM with more than 70 citations. (Sik-Ho Tsang @ Medium)
Outline
- VRCNN Network Architecture
- Experimental Results
1. VRCNN Network Architecture
- VRCNN is modified from ARCNN.
- The second layer of ARCNN ( fixed 7×7 filters) is replaced with the combination of 5×5 and 3×3 filters. The outputs of different-sized filters are concatenated to be fed into the next layer.
- Variable filter size in the third layer is also adopted that performs “restoration” of features. The fixed 1×1 filters in ARCNN are replaced by combination of 3×3 and 1×1 filters.
- The variable size filter technique is originated from GoogLeNet / Inception-v1.
- The first and the last layers of VRCNN do not use variable filter size, because these two layers perform feature extraction and final reconstruction, respectively.
- Besides, different from ARCNN, residue learning (concept of ResNet) is also adopted since, in the case of artifact reduction, the input (before filtering) and the output (after filtering) shall be similar to the other to a large extent, therefore, learning the difference between them can be easier and more robust.
- Mean Square Error (MSE) is used as loss function.
- Unlike Sample Adaptive Offset (SAO) filter used in HEVC, VRCNN no needs extra signalling bits for parameters.
- VRCNN can be made in-loop or out-of-loop.
2. Experimental Results
- 4.6% average Y BD-Rate reduction compared to HEVC baseline.
- Up to 7.6% Y BD-Rate reduction for RaceHorses sequence.
- Authors also trained the ARCNN and VDSR.
- ARCNN obtain 0.2% increase in Y BD-rate which means even worse.
- VDSR can achieve 3.8% average Y BD-rate reduction only.
- VRCNN and ARCNN are both 4-layer.
- VRCNN is slightly slower because in the second and third layers there are filters of different sizes, causing some troubles for parallel computing.
- Memory cost is an important issue for post-processing especially at the decoder side.
- VRCNN is much shallower than VDSR and also has much less parameters than ARCNN.
Reference
[2017 MMM] [VRCNN]
A Convolutional Neural Network Approach for Post-Processing in HEVC Intra Coding
My Previous Reviews
Image Classification [LeNet] [AlexNet] [Maxout] [NIN] [ZFNet] [VGGNet] [Highway] [SPPNet] [PReLU-Net] [STN] [DeepImage] [SqueezeNet] [GoogLeNet / Inception-v1] [BN-Inception / Inception-v2] [Inception-v3] [Inception-v4] [Xception] [MobileNetV1] [ResNet] [Pre-Activation ResNet] [RiR] [RoR] [Stochastic Depth] [WRN] [Shake-Shake] [FractalNet] [Trimps-Soushen] [PolyNet] [ResNeXt] [DenseNet] [PyramidNet] [DRN] [DPN] [Residual Attention Network] [DMRNet / DFN-MR] [IGCNet / IGCV1] [MSDNet] [ShuffleNet V1] [SENet] [NASNet] [MobileNetV2]
Object Detection [OverFeat] [R-CNN] [Fast R-CNN] [Faster R-CNN] [MR-CNN & S-CNN] [DeepID-Net] [CRAFT] [R-FCN] [ION] [MultiPathNet] [NoC] [Hikvision] [GBD-Net / GBD-v1 & GBD-v2] [G-RMI] [TDM] [SSD] [DSSD] [YOLOv1] [YOLOv2 / YOLO9000] [YOLOv3] [FPN] [RetinaNet] [DCN]
Semantic Segmentation [FCN] [DeconvNet] [DeepLabv1 & DeepLabv2] [CRF-RNN] [SegNet] [ParseNet] [DilatedNet] [DRN] [RefineNet] [GCN] [PSPNet] [DeepLabv3] [LC] [FC-DenseNet] [IDW-CNN] [SDN]
Biomedical Image Segmentation [CUMedVision1] [CUMedVision2 / DCAN] [U-Net] [CFS-FCN] [U-Net+ResNet] [MultiChannel] [V-Net] [3D U-Net] [M²FCN] [SA] [QSA+QNT] [3D U-Net+ResNet]
Instance Segmentation [SDS] [Hypercolumn] [DeepMask] [SharpMask] [MultiPathNet] [MNC] [InstanceFCN] [FCIS]
Super Resolution [SRCNN] [FSRCNN] [VDSR] [ESPCN] [RED-Net] [DRCN] [DRRN] [LapSRN & MS-LapSRN] [SRDenseNet]
Human Pose Estimation [DeepPose] [Tompson NIPS’14] [Tompson CVPR’15] [CPM]
Codec Post-Processing [ARCNN] [Lin DCC’16] [IFCNN] [Li ICME’17] [VRCNN]