Two-Stage Filter Response Normalization Network For Real Image Denoising
Two-Stage Filter Response Normalization Network For Real Image Denoising
RESEARCH NOTE
Abstract In this paper, we propose a two-stage network for real image denoising with filter response nor-
malization, named as two-stage filter response normalization network (TFRNet). In TFRNet, we propose
a filter response normalization(FRN) block to extract features and accelerate the training of the network.
TFRNet consists of two stages, at each stage of which we use the encoder-decoder structure based on
U-Net. We also use the coordinate attention block(CAB), double channel downsampling module, double
skip connection module, and convolutional (Conv) block in our TFRNet. With the help of these modules,
TFRNet provides excellent results on both SIDD and DND datasets for real image denoising.
2.3 Attention
Fig. 1 Coordinate attention from [3]
Attention mechanisms have been shown to be ef-
fective in low-level computer vision tasks. In image
network is used with several original-resolution blocks denoising tasks, attention is mainly divided into spa-
consisting of several channel attention blocks without tial attention and channel attention. MPRNet uses
any downsampling operations, in order to preserve the channel attention, and in [11], both channel attention
desired fine texture in the final output image. and spatial attention were used to improve the image
In addition, MPRNet incorporates a SAM between denoising accuracy.
every two stages; this suppresses useless information Coordinate attention, which is shown in Fig. 1, is
and then transforms useful information to the next used in high-level computer vision tasks. It embeds
stage. Therefore we also use the SAM in our TFRNet. pixel positional information into channel attention.
MPRNet achieves strong performance in image Therefore, it captures dependences with precise
restoration tasks. However, the network structure of positional information while inheriting the benefits of
MPRNet is very complicated because of the many channel attention and achieves excellent results in im-
channel attention blocks in the last stage. There- age classification tasks. Because of such benefits, we
fore, MPRNet is computationally intensive and time- add a CAB at the beginning of each stage in our TFR-
consuming. Net.
澺濆濂澔澶激濃澷澿
X2 is used as the result of TFRNet. The Charbonnier 澮
澷濃濂濊 澺濆濂 濈激濉 澷濃濂濊 濆澹激濉
loss is used as the metric of the loss function and we
optimize the end-to-end TFRNet as
2
∑
澷濣濢濪 澶激濃澷澿
Loss = Lch (Xi , Y ) (2) 澷濃濂濊
i=1 澮
澷濃濂濊 濆澹激濉 澷濃濂濊 濆澹激濉