Preprint
Article

This version is not peer-reviewed.

R-PreNet: Deraining Network Based on Image Background Prior

A peer-reviewed version of this preprint was published in:
Applied Sciences 2023, 13(21), 11970. https://doi.org/10.3390/app132111970

Submitted:

26 September 2023

Posted:

27 September 2023

You are already at the latest version

Abstract
Single image deraining (SID) has shown its importance in many advanced computer vision tasks. Though many CNN based image deraining methods have been proposed, how to effectively remove raindrops while maintaining background structure remains a challenge that needs to be overcome. Most of the deraining work focuses on removing rain streaks, but in heavy rain images, the dense accumulation of rainwater or the rain curtain effect significantly interferes with the effective removal of rain streaks, and often introduces some artifacts that make the scene more blurry. In this paper, we propose a new network structure R-PReNet for single image deraining with good background structure maintaining. This framework fully utilizes the cyclic recursive structure of PReNet. Moreover, we introduce residual channel prior (RCP) and feature fusion modules for better deraining performance by focusing on background feature information. Compared with the previous methods, our method has significantly improvement effect on the rainstorm image with the artifacts removing and good visual detail restoring.
Keywords: 
;  ;  

1. Introduction

Rain is a common weather phenomenon which causes adverse effects on the visual quality of images and affecting the performance of subsequent image processing tasks, such as object recognition [1], object detection [2], autonomous driving and video surveillance [3,4,5], etc. Hence, removing rain streaks from images with rain has become an important and meaningful research topic, and has also received attention in recent years [6,7,8]. Single image deraining refers to restoring a clean and rainless image scene from a single image with rain. However, due to the complex combination of background information and raindrop information, how to simultaneously remove raindrops and protect the background remains a challenging issue. We found in the experiment that the PreNet deraining network model [6] can reconstruct a relatively clear rain free image, but in the test of rainstorm data set, the background structure of the reconstructed image corresponding to the rainstorm image has also been damaged to some extent, that is, the introduction of artifacts, and this destruction of the image background will sometimes lead to serious problems, such as fuzzy or missing traffic signs may lead to serious accidents in automatic driving. In order to address this problem, this paper introduce an additional image background prior to protect the background structure, so that a clearer and correct reconstruction of rainless images can be obtained in the case of rainstorm, as shown in Figure 1.
In this article, we explore the effective reconstruction problem of complex combinations of background and raindrops, and propose a new algorithm called R-PReNet that can effectively remove raindrops and protect background information. This algorithm fully utilizes the cyclic recursive structure and raindrop energy of PReNet. In addition, this article introduces residual channel prior (RCP) [9,10,11,12] in the model to achieve background structure protection. In addition, this article also proposes the use of the 'Squeeze Excitation' residual module (SE ResBlock) [13] to extract deep features of RCP, and the interactive fusion feature module (IFM) [12] to fully utilize RCP information, achieving high-quality rainless image reconstruction.
Our contributions are summarized as follows:
This article replicates and tests the PreNet deraining network on three popular image deraining datasets (Rain100H [15], Rain100L [15], Rain14000 [16]) and real datasets (Practical_by_Yang [15]), and studies the results of deraining.
This article explores the effectiveness of residual channel prior (RCP) for background protection and proposes an image deraining network structure based on RCP. Numerous experiments have shown that our method outperforms the original method on commonly used rainfall datasets, restoring visually clean images and good details.
We propose an RCP extraction module and an interactive fusion module (IFM) for RCP extraction and guidance, respectively, to obtain deep features of RCP and guide the network to recover more background details.
The rest of this article is as follows. In Section 2, we briefly reviewed the relevant research on image deraining methods. In Section 3, we proposed an overall R-PreNet deraining network based on image background prior, and elaborated on the RCP residual channel prior and IFM fusion methods in detail. In the fourth section, we presented our experimental results and comparisons. The conclusion is given in Section 5.

3. Proposed Work

In this section, we introduce the overall network architecture of the algorithms in this paper. We first describe the proposed residue channel prior (RCP) implementation details. Then We show the structure of progressive recurrent network (PReNet) as a backbone network. Finally, we propose methods for fusing high-dimensional features of the RCP.

3.1. Residue-Progressive Recurrent Network

As shown in Figure 2, R-PReNet consists of two main parts: (i) the RCP feature extraction and fusion module, and (ii) the progressive recurrent network. We first extract and fuse image features and RCP features from the rainy image. Then we concatenate the fused features with image features. The components of our method are described in detail in the following sections.

3.2. Residue Channel Prior (RCP)

The occurrence of rain streaks is usually modeled as a linear combination of a background layer and a rain layer [16,24,36,37]. Based on this model, Li et al [9] demonstrated that subtracting the minimum color channel from the maximum color channel will generate a rain-free image. Rain streaks are colorless (white or grey) and appear at the same location in different RGB color channels. So subtracting the minimum color channel from the maximum color channel will cancel the appearance of rain streaks as in Figure 3.
The colored-image intensity of a rainy image is defined [9] as:
I ~ x =   τ   ρ r s x L σ + T   τ B π
where L = ( L r , L g , L b ) T is the color vector of luminance and B = ( B r , B g , B b ) T is the color vector of background reflection.
L = L r + L g + L b   ,   B = B r + B g + B b
In the model (Eq. (2)), the first term is the rain streak term and the second term is the background term. σ = L /   L and π = B /   B are defined the chromaticities of L and B. T is the exposure time and τ is the time for the raindrop to pass through pixel x. ρ r s consists of the refraction coefficients of the raindrop, the specular reflection coefficients, and the internal reflection coefficients. We assume that ρ r s is wavelength independent, which implies that the raindrop is colorless.
As a consequence, we need to cancel the light chromaticity σ in the rain-streak term in Eq. (2) to generate a residual channel without rain streaks. To do so, we use any existing color constancy algorithm [38] to estimate σ, and then apply the following normalization step to the input image.
I x = I ~ ( x ) σ = I r s x i + I b g ( x )
where, i = ( 1 , 1 , 1 ) T , I r s = τ ρ r s L , I b g = ( T τ ) B / σ .
Vector division is done element-wise. Note that when we normalize the image, we cancel not only the light chromaticity but also the color effect of spectral sensitivity.Hence, according to the previous equation and a rainy image I, the residual channel is defined as:
I r e s x = I M x I m ( x )
where,
I M x = max { I r x , I g x , I b x }
I m x = min { I r x , I g x , I b x }
I r e s is the residual channel of the image I, which has no rain streaks.

3.3. RCP High-Dimensional Feature Extraction

Although the operation of subtracting a color channel from another color channel in the image space is useful and the structural information of the RCP is clearer than the rainy image, it can be destructive to the background image because of information loss. Therefore, we move the operation that utilizes the RCP structural information to the feature domain. We propose the RCP feature extraction module to extract the high dimensional features of RCP.
Based on the Squeeze-and-Excitation (SE) block proposed by Hu et al [14], which focuses on channel relationships to construct informative features, this residual block adaptively recalibrates the channel feature responses by explicitly modeling the interdependencies between channels. Since the RCP module is a module that uses color channels to interact with each other, in order to reduce the noise in the initial features and enrich the semantic information of the features, we use the SE-ResBlock structure shown in Figure 4 to extract the high dimensional features Fp of the RCP.

3.4. Interactive Fusion Features

We extract the high dimensional features of RCP, but it is still a challenging task that fully utilize the RCP features to guide the model.
A simple solution is directly concatenating RCP features with image features, but this is ineffective for guiding model deraining and may cause feature interference. To address this problem, we propose an interactive fusion module (IFM) [39] consisting of two branches (rainy image features and prior features) to progressively combine features. As shown in Figure 5, two 3 × 3 kernel-sized convolutions are performed to map the rainy image features Fo and RCP features Fp to F o ^   and F p ^ .
Next, the similarity map S between F o ^ and F p ^ is computed using element multiplication:
S = S i g m o i d ( F o ^ F p ^ )
The background information of the rainy image corrupted by the rain streaks is enhanced using the similarity map S. In addition, since the background of the RCP is similar to the rainy image, the similarity map S can also highlight the feature information in the priori features, which further enhances the structure of the priori features.

3.5. Progressive Recurrent Network

The progressive recurrent network consists of the following four parts: (i) a convolutional layer f i n receives network inputs, (ii) a recurrent layer f r e c u r r e n t   propagates cross-stage feature dependencies, (iii) several residual blocks f r e s extracts the deep representation, and (ii) a convolutional layer f o u t outputs deconvolutional results. Where f i n takes as input the current estimation x t 1 , the rainy image y, and the concatenation of the background fusion prior features G. The recurrent layer we implement using convolutional Long Short-Term Memory (LSTM) because LSTM has experience advantages in image deraining, through which cross-stage feature dependencies can be propagated to facilitate rain streaks removal.
x t 0.5 = f i n ( x t 1 , y , G )
s t = f r e c u r r e n t ( s t 1 , x t 0.5 )
x t = f o u t ( f r e s ( s t ) )
where f i n , f r e s and f o u t are stage-invariant, the network parameters are reused in different stages. The recurrent layer f r e c u r r e n t takes x t 0.5 and the recurrent state s t 1 as inputs to stage t-1. By unfolding PreNet [6] with T recurrent stages, the deep representation of rain streak removal is favored by recurrent state propagation. The rain removal results from the intermediate stages of the network structure show that the accumulation of storm streaks can be gradually eliminated.

3.6. Loss Function

We employ negative SSIM loss [40] as our objective function. For a model with T stages, we have T outputs, x 1 , x 2 , …, x T , and we apply supervision only to the final output x T . The negative SSIM loss is:
L = S S I M ( x T ,   x g t )
where   x g t is the corresponding ground-truth clean image.

4. Experiments

Our model was trained on Ubuntu OS, NVIDIA GeForce GTX 3080Ti GPU using Pytorch framework in Python environment with 12GB of RAM. To validate the effectiveness of our model, we evaluated our method on three popular image-deraining synthetic datasets (Rain100H, Rain100L, Rain14000) and a real rainy images dataset (Practical_by_Yang) to evaluate our approach:
Figure 6. Image deraining results tested in both synthetic and real datasets. The first column is the rainy image, the second column is the real rain-free image on the synthetic dataset (no example images on the real dataset), the third column is the deraining result of the PReNet algorithm, and the fourth column is the deraining result of the R-PReNet algorithm of this paper. It can be seen that R-PReNet can reconstruct the rain-free image with clearer background structure and reduce the introduction of artifacts.
Figure 6. Image deraining results tested in both synthetic and real datasets. The first column is the rainy image, the second column is the real rain-free image on the synthetic dataset (no example images on the real dataset), the third column is the deraining result of the PReNet algorithm, and the fourth column is the deraining result of the R-PReNet algorithm of this paper. It can be seen that R-PReNet can reconstruct the rain-free image with clearer background structure and reduce the introduction of artifacts.
Preprints 86227 g006aPreprints 86227 g006bPreprints 86227 g006c

4.1. Experimental Setup

4.1.1. A Datasets

In this paper, we mainly use synthetic datasets and real datasets for evaluation. The synthetic image datasets include (1) Rain100L, where 200 pairs of images are used for training and 100 pairs of images are used for testing; (2) Rain100H has 200 synthetic images used for training and 100 images used for testing; and (3) Rain14000, which composed of training and test images with a ratio of 12600:1400 split. The real dataset consists of (1) the Practical_by_Yang dataset with 147 images without ground-truth; and (2) real rainy images from certain movie and television productions.

4.1.2. B Evaluation Indicators

In our experiments, for images with ground truth, we can evaluate each method by two commonly used quantitative metrics, the peak signal-to-noise ratio (PSNR) [41] and the structural similarity index (SSIM) [40]. For the images without ground truth (i.e., real dataset), we provide some visual results.

4.2. Ablation Study

4.2.1. A Effectiveness on RCP module

The first ablation study evaluates the performance of R-PReNet with experimental results with and without the RCP module, we train and test the networks with and without the RCP module and the baseline algorithms of JORDER [28] and RESCAN [31] on the datasets Rain100L, Rain100H, and Rain14000, respectively, and Table 1 shows the performance of the above algorithms on the quantitative results in PSNR and SSIM. Both quantitative and visual results show that the recurrent network with RCP module outperforms the network without RCP module and the baseline algorithm.

4.2.2. B Effectiveness on IFM module

To investigate the effectiveness of the feature fusion module, we compare two different network architectures: (a) with the RCP module, but the RCP high-dimensional features are directly connected with the rainy image features into the network, and (b) with the RCP module and the IFM module, which uses interactive fusion to combine the RCP high-dimensional features and the rainy image features together into the network. We train and test the networks with and without the IFM module and the baseline algorithms of JORDER [28], RESCAN [31], and PreNet [6] on the datasets Rain100L, Rain100H, and Rain14000, respectively, and Table 2 shows the quantitative results of the above algorithms in PSNR and SSIM. Both quantitative and visual results are that the recurrent network with IFM module outperforms the network without IFM module and the baseline network.

5. Conclusion

In this paper, we propose a progressive recurrent deraining network based on background protection. The experiments show that this algorithm can remove rain streaks and protect background information at the same time. In the pre-processing stage of the rainy image, we first extract the residual channel from the rainy image, and the extracted residual channel does not contain rain streaks, while the residual channel is used to extract high-dimensional features, and then we interactively fuse the extracted features with the rainy image features and then input into the progressive recurrent network. The input to the network of each generation is composed of the fused features, the reconstructed image of the previous generation, and the original rainy image. After generations of progressive recursion, the final rain-free image is produced. Comprehensive experimental evaluations show that our method outperforms the original algorithm on both synthetic and real rainy images.

Acknowledgments

This work was supported by the National Natural Science Foundation of China under Grant 61305040 and the Higher Education Science Research Project of Shaanxi Higher Education Society of China under Grant XGH21062.

References

  1. Josi, M. Alehdaghi, R. M. O. Cruz and E. Granger, "Multimodal Data Augmentation for Visual-Infrared Person ReID with Corrupted Data," 2023 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 2023, pp. 1-10. [CrossRef]
  2. S. S. Chaturvedi, L. Zhang and X. Yuan, "Pay "Attention" to Adverse Weather: Weather-aware Attention-based Object Detection," 2022 26th International Conference on Streaks Recognition (ICPR), Montreal, QC, Canada, 2022, pp. 4573-4579. [CrossRef]
  3. J. Xiao, H. Long, R. Li and F. Li, "Research on Methods of Improving Robustness of Deep Learning Algorithms in Autonomous Driving," 2022 IEEE International Conference on Advances in Electrical Engineering and Computer Applications (AEECA), Dalian, China, 2022, pp. 644-647. [CrossRef]
  4. H. Tyagi, V. Kumar and G. Kumar, "A Review Paper on Real-Time Video Analysis in Dense Environment for Surveillance System," 2022 International Conference on Fourth Industrial Revolution Based Technology and Practices (ICFIRTP), Uttarakhand, India, 2022, pp. 171-183. [CrossRef]
  5. Z. Zhang, W. Lu, W. Sun, X. Min, T. Wang and G. Zhai, "Surveillance Video Quality Assessment Based on Quality Related Retraining," 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 2022, pp. 4278-4282. [CrossRef]
  6. D. Ren, W. Zuo, Q. Hu, P. Zhu and D. Meng, "Progressive Image Deraining Networks: A Better and Simpler Baseline," 2019 IEEE/CVF Conference on Computer Vision and Streaks Recognition (CVPR), Long Beach, CA, USA, 2019, pp. 3932-3941. [CrossRef]
  7. Y. Wei et al., "DerainCycleGAN: Rain Attentive CycleGAN for Single Image Deraining and Rainmaking," in IEEE Transactions on Image Processing, vol. 30, pp. 4788-4801, 2021. [CrossRef]
  8. H. Huang, A. Yu and R. He, "Memory Oriented Transfer Learning for Semi-Supervised Image Deraining," 2021 IEEE/CVF Conference on Computer Vision and Streaks Recognition (CVPR), Nashville, TN, USA, 2021, pp. 7728-7737. [CrossRef]
  9. Ruoteng Li, Robby T Tan, and Loong-Fah Cheong. Robust optical flow in rainy scenes. In Proceedings of the European Conference on Computer Vision (ECCV), pages 288–304, 2018. 2, 4.
  10. Ruoteng Li, Robby T Tan, and Loong-Fah Cheong. All in one bad weather removal using architectural search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Streaks Recognition, pages 3175–3185, 2020. 2.
  11. Ruoteng Li, Robby T Tan, Loong-Fah Cheong, Angelica I Aviles-Rivero, Qingnan Fan, and Carola-Bibiane Schonlieb.Rainflow: Optical flow under rain streaks and rain veiling effect. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7304–7313, 2019. 2.
  12. Q. Yi, J. Li, Q. Dai, F. Fang, G. Zhang and T. Zeng, "Structure-Preserving Deraining with Residue Channel Prior Guidance," 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 2021, pp. 4218-4227. [CrossRef]
  13. X. Zhong, O. Gong, W. Huang, L. Li and H. Xia, "Squeeze-and-Excitation Wide Residual Networks in Image Classification," 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 2019, pp. 395-399. [CrossRef]
  14. Hu, Jie et al. “Squeeze-and-Excitation Networks.” 2018 IEEE/CVF Conference on Computer Vision and Streaks Recognition (2017): 7132-7141.
  15. Yang W, Tan R T, Feng J, et al. Deep joint rain detection and removal from a single image[C]//Proceedings of the IEEE Conference on Computer Vision and Streaks Recognition. 2017: 1357-1366.
  16. Fu X , Huang J , Zeng D , et al. Removing Rain from Single Images via a Deep Detail Network[C]// 2017 IEEE Conference on Computer Vision and Streaks Recognition (CVPR). IEEE Computer Society, 2017.
  17. He K, Sun J, Tang X (2010) Guided image filtering. European Conf. on Comput. vision, pp 1–14.
  18. Xu J, Zhao W, Liu P, Tang X (2012) Removing rain and snow in a single image using guided filter. IEEE Int. Conf. on Comput. Sci. and Automation Eng, pp 304–307.
  19. Zheng X, Liao Y, Guo W, Fu X, Ding X (2013) Single-image-based rain and snow removal using multi-guided filter. Neural Inform. Process, pp. 258–265. [CrossRef]
  20. Ding X, Chen L, Zheng X, Huang Y, Zeng D (2016) Single image rain and snow removal via guided L0 smoothing filter. Multimed Tools Appl 75(5):2697–2712. [CrossRef]
  21. Kim JH, Lee C, Sim JY, Kim (2013) Single-image deraining using an adaptive nonlocal means filter. IEEE Int. Conf. on Image Process, pp 914–917.
  22. Kang L, Lin C, Fu Y (2012) Automatic single-image-based rain streaks removal via image decomposition. IEEE Trans Image Process 24(4):1742–1755. [CrossRef]
  23. Kang L, Lin C, Lin C, Lin Y (2012) Self-learning-based rain streak removal for image/video. IEEE Int Symp Circuits Syst 57(1):1871–1874. [CrossRef]
  24. Luo Y, Xu Y, Ji H (2015) Removing rain from a single image via discriminative sparse coding. IEEE Int. Conf. on Comput. Vision, pp 3397–3405.
  25. Li Y, Tan RT, Guo X, Lu J, Brown MS (2016) Rain streak removal using layer priors. IEEE Conf. on Comput. vision and streaks recognition.
  26. Gu S, Meng D, Zuo W, Zhang L (2017) Joint convolutional analysis and synthesis sparse representation for single image layer separation. IEEE Int. Conf. on Comput. Vision, pp 1717–1725.
  27. Wei W, Meng D, Zhao Q, Xu Z, Wu Y (2019) Semi-supervised transfer learning for image rain removal. In Proc. of the IEEE Conf. on Comput. vision and streaks recognition, pp 3877–3886.
  28. Yang W, Tan RT, Feng J, Liu J, Guo Z, Yan S (2017) Deep joint rain detection and removal from a single image. IEEE Conf. on Comput. vision and streaks recognition, pp 1685–1694.
  29. Mu P, Chen J, Liu R, Fan X, Luo Z (2018) Learning bilevel layer priors for single image rain streaks removal. IEEE Signal Process Lett 26(2):307–31. [CrossRef]
  30. Eigen D,Krishnan D,Fergus R.Restoring an image taken through a window covered with dirt or rain [C]//Proc of IEEE International Conference on Computer Vision.Washington DC:IEEE Computer Society, 2013: 633-640.
  31. Li Xia,Wu Jianlong,Lin Zhouchen, et al. Recurrent squeeze-and-excitation context aggregation net for single image deraining[C]//Proc of the 15th European Conference on Computer Vision. Cham: Springer, 2018: 262-277.
  32. Rajeev Yasarla, Vishwanath A Sindagi, and Vishal M Patel.Syn2real transfer learning for image deraining using gaussian processes. In IEEE/CVF Conference on Computer Vision and Streaks Recognition, pages 2726–2736, 2020.
  33. Jin Xin,Chen Zhibo,Lin Jianxin, et al. Unsupervised single image deraining with self-supervised constraints[C]//Proc of IEEE International Conference on Image Processing.Piscataway, NJ: IEEE Press,2019: 2761-2765.
  34. Zhu H , Peng X , Zhou J T , et al. Singe Image Rain Removal with Unpaired Information: A Differentiable Programming Perspective[C]// 2019:9332-9339.
  35. R. Li, R. T. Tan, and L.-F. Cheong. Robust optical flow in rainy scenes. In The European Conference on Computer Vision (ECCV), September 2018.
  36. W. Yang, R. T. Tan, J. Feng, J. Liu, S. Yan, and Z. Guo.Joint rain detection and removal from a single image with contextualized deep networks. IEEE transactions on streaks analysis and machine intelligence, 2019. [CrossRef]
  37. L. W. Kang, C. W. Lin, and Y . H. Fu. Automatic single-image-based rain streaks removal via image decomposition.IEEE Transactions on Image Processing, 21(4):1742–1755,April 2012. [CrossRef]
  38. Cheng, D., Prasad, D.K., Brown, M.S.: Illuminant estimation for color constancy: why spatial-domain methods work and the role of the color distribution. JOSA A 31(5), 1049–1058 ,2014.
  39. Yuchen Hu, Nana Hou, Chen Chen, Eng Siong Chng. Interactive Feature Fusion for End-to-End Noise-Robust Speech Recognition. ICASSP 2022: 6292-6296.
  40. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P . Simoncelli.Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing,13(4):600–612, 2004. 3, 5. [CrossRef]
  41. Q. Huynh-Thu and M. Ghanbari. Scope of validity of psnr in image/video quality assessment. Electronics letters, 44(13):800–801, 2008. 6. [CrossRef]
Figure 1. Image deraining in the real world. PReNet [6] and R-PReNet were trained on RainTrainH. This image shows that R-PReNet can effectively remove rain streaks while retaining better background textures and maintaining the basic tone of the original image.
Figure 1. Image deraining in the real world. PReNet [6] and R-PReNet were trained on RainTrainH. This image shows that R-PReNet can effectively remove rain streaks while retaining better background textures and maintaining the basic tone of the original image.
Preprints 86227 g001
Figure 2. The overall structure of Residue-progressive recurrent network (R-PReNet), where (a) shows the overall network framework of R-PReNet; (b) shows progressive recurrent network composition in R-PReNet, where f i n is a convolutional layer with ReLU, f r e s is a recursive ResBlocks, f o u t is a convolutional layer, f r e c u r r e n t is a convolutional LSTM, © is a connectivity layer; (c) is the RCP fusion feature module.
Figure 2. The overall structure of Residue-progressive recurrent network (R-PReNet), where (a) shows the overall network framework of R-PReNet; (b) shows progressive recurrent network composition in R-PReNet, where f i n is a convolutional layer with ReLU, f r e s is a recursive ResBlocks, f o u t is a convolutional layer, f r e c u r r e n t is a convolutional LSTM, © is a connectivity layer; (c) is the RCP fusion feature module.
Preprints 86227 g002
Figure 3. RCP Extraction Module.
Figure 3. RCP Extraction Module.
Preprints 86227 g003
Figure 4. SE-ResBlock Module.
Figure 4. SE-ResBlock Module.
Preprints 86227 g004
Figure 5. Interactive Fusion Feature Module.
Figure 5. Interactive Fusion Feature Module.
Preprints 86227 g005
Table 1. Performance comparison of synthetic datasets on network structure with and without RCP module.
Table 1. Performance comparison of synthetic datasets on network structure with and without RCP module.
Methods. PReNet R-PReNet JORDER [28] RESCAN [31]
Rain100H 29.46/0.899 30.76/0.916 26.54/0.835 28.88/0.866
Rain100L 37.48/0.979 38.87/0.984 36.61/0.974 ---
Rain14000 32.60/0.946 33.03/0.963 --- ---
Table 2. Performance comparison of synthetic datasets with and without IFM module network structure.
Table 2. Performance comparison of synthetic datasets with and without IFM module network structure.
Methods PReNet R-PreNet
(no IFM)
R-PreNet JORDER [28] RESCAN [31]
Rain100H 29.46/0.899 29.86/0.901 30.76/0.916 26.54/0.835 28.88/0.866
Rain100L 37.48/0.979 37.67/0.967 38.87/0.984 36.61/0.974 ---
Rain14000 32.60/0.946 32.89/0.954 33.03/0.963 --- ---
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated