Publications-Theses
Article View/Open
Publication Export
-
Google ScholarTM
NCCU Library
Citation Infomation
Related Publications in TAIR
題名 透過最小-最大均值池化濾波器與多頭自注意力卷積神經網路之影像雜訊移除研究
Image Denoising Using Min-Max Mean Pooling Filters and Multi-Head Self-Attention Convolutional Neural Networks作者 陸妍諭
Lu, Yen-Yu貢獻者 張宏慶
Jang, Hung-Chin
陸妍諭
Lu, Yen-Yu關鍵詞 多頭自注意力神經網路
卷積神經網路
深度學習
均值濾波器
椒鹽雜訊
池化濾波
multi-head self-attention neural network
convolutional neural network
deep learning
mean filter
salt-and-pepper noise
pooling-based filtering日期 2025 上傳時間 1-Jul-2025 15:06:44 (UTC+8) 摘要 數位影像在傳輸的過程中,可能會受到電磁干擾和攝影元件受損,導致影像受到脈衝雜訊干擾而損壞,如何有效修復遭受極值脈衝雜訊(椒鹽雜訊)干擾的影像,對於提升數位影像品質極為重要。本文提出一種使用自適應分析視窗最小-最大均值池化的多頭自注意力卷積神經網路,去除在傳輸過程中產生的椒鹽雜訊;首先,估測影像的雜訊密度,若為輕度雜訊干擾之影像,乾淨像素足夠,使用多頭自注意力卷積神經網路,計算輸入序列中不同位置像素的權重,並擷取長距離像素間的相關性,預測適合重建的臨域像素;相對的,若是中、高雜訊密度干擾時,受損像素較多,則透過多層的最小值和最大值均值池化濾波器,分別計算最大池化後的影像和最小池化後的兩類影像;最後將處理後的影像重新組合,並進行均值濾波處理;在高雜訊密度的環境中,若分析視窗沒有乾淨像素可以參考,則擴大分析視窗的尺寸,增加鄰域未受干擾像素引入的機率,修復影像中的受損像素。經由實驗結果證明:本文所提出的濾波器可以在各種雜訊密度有效重建受到雜訊干擾的影像,重建效果也優於許多極為先進(state-of-the-art)的演算法。
During the transmission of digital images, electromagnetic interference and sensor defects can result in impulse noise corruption, leading to severe image degradation. Effectively restoring images contaminated by extreme impulse noise, such as salt-and-pepper noise, is crucial for improving digital image quality. This paper proposes a multi-head self-attention neural network enhanced by adaptive min-max mean pooling to remove salt-and-pepper noise introduced during transmission. First, the noise density of the corrupted image is estimated. Suppose the image is subjected to low-level noise and contains a sufficient number of clean pixels; a multi-head self-attention mechanism is employed to compute the attention weights of pixels at different positions in the input sequence. This mechanism captures long-range dependencies to predict suitable neighborhood pixels for reconstruction. In contrast, under moderate to high noise density, where a larger portion of pixels are corrupted, the method utilizes multiple layers of min and max mean pooling filters to compute two feature maps separately based on maximum and minimum pooling operations. The processed images are then fused and further refined using mean filtering. In high-noise scenarios where no clean pixels are available within a small analysis window, the window size is adaptively enlarged to increase the likelihood of including undisturbed neighboring pixels, thereby improving the reconstruction of damaged regions. Experimental results demonstrate that the proposed filtering method can effectively restore noise-corrupted images across various noise levels, outperforming several state-of-the-art algorithms regarding reconstruction quality.參考文獻 [1] R. C. Gonzalez, R. E. Woods, and S. L. Eddins, Digital Image Processing, 2nd ed. New York, NY, USA: Prentice-Hall, 2002. [2] J. D. Gibson and A. Bovik, Handbook of Image and Video Processing, 1st ed. San Diego, CA, USA: Academic Press, 2000. [3] H. Hwang and R. A. Haddad, "Adaptive median filters: new algorithms and results," IEEE Trans. Image Process., vol. 4, no. 4, pp. 499–502, Apr. 1995. [4] S.-J. Ko and Y. H. Lee, "Center weighted median filters and their applications to image enhancement," IEEE Trans. Circuits Syst., vol. 38, no. 9, pp. 984–993, Sep. 1991. [5] Y. Dong and S. Xu, "A new directional weighted median filter for removal of random-valued impulse noise," IEEE Signal Process. Lett., vol. 14, no. 3, pp. 193–196, Mar. 2007. [6] T. Chen and H. R. Wu, "Adaptive impulse detection using center-weighted median filters," IEEE Signal Process. Lett., vol. 8, no. 1, pp. 1–3, Jan. 2001. [7] C.-T. Lu and T.-C. Chou, "Denoising of salt-and-pepper noise corrupted image using modified directional-weighted-median filter," Pattern Recognit. Lett., vol. 33, no. 10, pp. 1287–1295, Jul. 2012. [8] S. Esakkirajan, T. Veerakumar, A. N. Subramanyam, and C. H. PremChand, "Removal of high density salt and pepper noise through modified decision based unsymmetric trimmed median filter," IEEE Signal Process. Lett., vol. 18, no. 5, pp. 287–290, May 2011. [9] T. Chen, K.-K. Ma, and L.-H. Chen, "Tri-state median filter for image denoising," IEEE Trans. Image Process., vol. 8, no. 12, pp. 1834–1838, Dec. 1999. [10] P. Satti, N. Sharma, and B. Garg, "Min-max average pooling based filter for impulse noise removal," IEEE Signal Process. Lett., vol. 27, pp. 1475–1479, 2020. [11] Y.-Y. Lu and H.-C. Jang, "Removal of salt-and-pepper impulse noise using adaptive analysis windows with min-max mean pooling," in Proc. 36th IPPR Conf. Comput. Vis. Graph. Image Process., 2023. [12] L. Zhai, L. Dong, Y. Liu, S. Fu, F. Wang, and Y. Li, "Adaptive hybrid threshold shrinkage using singular value decomposition," J. Electron. Imaging, vol. 28, no. 6, pp. 063002, Nov. 2019. [13] S. H. Chan, T. Zickler, and Y. M. Lu, "Monte Carlo non-local means: Random sampling for large-scale image filtering," IEEE Trans. Image Process., vol. 23, no. 8, pp. 3711–3725, Aug. 2014. [14] E. Luo, S. H. Chan, and T. Q. Nguyen, "Adaptive image denoising by mixture adaptation," IEEE Trans. Image Process., vol. 25, no. 10, pp. 4489–4503, Oct. 2016. [15] S. Anwar, F. Porikli, and C. P. Huynh, "Category-specific object image denoising," IEEE Trans. Image Process., vol. 26, no. 11, pp. 5506–5518, Nov. 2017. [16] J. Xu, L. Zhang, and D. Zhang, "External prior guided internal prior learning for real-world noisy image denoising," IEEE Trans. Image Process., vol. 27, no. 6, pp. 2996–3010, Jun. 2018. [17] M. S. M. Zain, A. K. Junoh, and A. Abdurrazzaq, "Hybrid singular value decomposition based alpha trimmed mean-median filter in eliminating high density salt and pepper noise from grayscale image," Multimed. Tools Appl., vol. 83, pp. 62895–62913, 2024. [18] J. Gao, L. Li, X. Ren, Q. Chen, and Y. M. Abdul-Abbass, "An effective method for salt and pepper noise removal based on algebra and fuzzy logic function," Multimed. Tools Appl., vol. 83, pp. 9547–9576, 2024. [19] Q. Liu, X. Li, and J. Yang, "Optimum codesign for image denoising between type-2 fuzzy identifier and matrix completion denoiser," IEEE Trans. Fuzzy Syst., vol. 30, no. 1, pp. 287–292, Jan. 2022. [20] X. Zhang, X. Li, Z. Tang, S. Zhang, and S. Xie, "Noise removal in embedded image with bit approximation," IEEE Trans. Knowl. Data Eng., vol. 34, no. 3, pp. 1359–1369, Mar. 2022. [21] N. N. Hien, D. N. H. Thanh, U. Erkan, and J. M. R. S. Tavares, "Image noise removal method based on thresholding and regularization techniques," IEEE Access, vol. 10, pp. 71584–71597, 2022. [22] N. Aslam, M. K. Ehsan, Z. U. Rehman, M. Hanif, and G. Mustafa, "A modified form of different applied median filter for removal of salt & pepper noise," Multimed. Tools Appl., vol. 82, pp. 7479–7490, 2023. [23] S. Zhang and M. A. Karim, "A new impulse detector for switching median filters," IEEE Signal Process. Lett., vol. 9, no. 11, pp. 360–363, Nov. 2002. [24] C.-T. Lu, Y.-Y. Chen, L.-L. Wang, and C.-F. Chang, "Removal of salt-and-pepper noise in corrupted image using three-values-weighted approach with variable-size window," Pattern Recognit. Lett., vol. 80, pp. 188–199, Sep. 2016. [25] D. Liu, B. Wen, J. Jiao, X. Liu, Z. Wang, and T. S. Huang, "Connecting image denoising and high-level vision tasks via deep learning," IEEE Trans. Image Process., vol. 29, pp. 3695–3706, 2020. [26] Z. Liu, W. Q. Ya, and M. L. Yang, "Image denoising based on a CNN model," in Proc. Int. Conf. Control, Autom. Robot., 2018, pp. 389–393. [27] A. V. Miclea, R. Terebes, I. Ilea, and M. Borda, "Hyperspectral image classification using combined spectral-spatial denoising and deep learning techniques," in Proc. IEEE Int. Conf. Autom., Qual. Testing, Robot. (AQTR), 2018. [28] K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, "Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising," IEEE Trans. Image Process., vol. 26, no. 7, pp. 3142–3155, Jul. 2017. [29] X. Li, Q. Gao, Y. Lu, and D. Sun, "Convolutional neural network denoising method based on multisize features," in Proc. IEEE Int. Conf. Intell. Human-Machine Syst. Cybern., 2018, pp. 146–149. [30] K. Zhang, W. Zuo, and L. Zhang, "FFDNet: Toward a fast and flexible solution for CNN-based image denoising," IEEE Trans. Image Process., vol. 27, no. 9, pp. 4608–4622, Sep. 2018. [31] T. Wang, M. Sun, and K. Hu, "Dilated deep residual network for image denoising," in Proc. IEEE Int. Conf. Tools Artif. Intell., 2017, pp. 1272–1279. [32] K. Isogawa, T. Shiodera, and T. Takeguchi, "Deep shrinkage convolutional neural network for adaptive noise reduction," IEEE Signal Process. Lett., vol. 25, no. 2, pp. 224–228, Feb. 2018. [33] E. Kandić, A. Akagic, and M. Bohlouli, "Exploring convolutional autoencoder efficacy in noise removal for image processing and computer vision: A study using the MNIST dataset," in Proc. 10th Int. Conf. Control, Decision Inf. Technol. (CoDIT), Vallette, Malta, 2024, pp. 1275–1280. [34] L. Deng, "The MNIST database of handwritten digit images for machine learning research," IEEE Signal Process. Mag., vol. 29, no. 6, pp. 141–142, Nov. 2012. [35] C.-T. Lu, R.-H. Chen, L.-L. Wang, and J.-A. Lin, "Image enhancement using convolutional neural network to identify similar patterns," IET Image Process., vol. 14, no. 17, pp. 3880–3889, 2020. [36] C.-T. Lu, H.-J. Hsu, and L.-L. Wang, "Image denoising using DLNN to recognize the direction of pixel variation," Signal Image Video Process., vol. 15, pp. 1247–1256, 2021. [37] A. A. Rafiee and M. Farhang, "A deep convolutional neural network for salt-and-pepper noise removal using selective convolutional blocks," Appl. Soft Comput., vol. 145, 2023. [38] C.-T. Lu, L.-L. Wang, J.-H. Shen, and J. A. Lin, "Image enhancement using deep-learning fully connected neural network mean filter," J. Supercomput., vol. 77, pp. 3144–3164, 2021. [39] Q. Yang, P. Yan, M. K. Kalra, and G. Wang, "CT image denoising with perceptive deep neural networks," arXiv preprint, arXiv:1702.07019, 2017. [40] K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," arXiv preprint arXiv:1409.1556, 2014. [41] E. Kang and J. C. Ye, "Framelet denoising for low-dose CT using deep learning," in Proc. IEEE Int. Symp. Biomed. Imaging (ISBI), 2018, pp. 311–314. [42] H. Yuan, J. Jia, and Z. Zhu, "SIPID: A deep learning framework for sinogram interpolation and image denoising in low-dose CT reconstruction," in Proc. IEEE Int. Symp. Biomed. Imaging (ISBI), 2018, pp. 1521–1524. [43] Q. Yang, P. Yan, Y. Zhang, H. Yu, Y. Shi, X. Mou, M. K. Kalra, Y. Zhang, L. Sun, and G. Wang, "Low-dose CT image denoising using a generative adversarial network with Wasserstein distance and perceptual loss," IEEE Trans. Med. Imaging, vol. 37, no. 6, pp. 1348–1357, Jun. 2018. [44] M. A. Marnissi, "Revolutionizing thermal imaging: GAN-based vision transformers for image enhancement," in Proc. IEEE Int. Conf. Image Process. (ICIP), Kuala Lumpur, Malaysia, 2023, pp. 2735–2739. [45] S. I. Jang, T. Pan, Y. Li, P. Heidari, and J. Chen, "Spach Transformer: Spatial and channel-wise transformer based on local and global self-attentions for PET image denoising," IEEE Trans. Med. Imaging, vol. 43, no. 6, pp. 2036–2049, Jun. 2024. [46] Z. Li, J. Zhang, S. Wei, Y. Gao, C. Cao, and Z. Wu, "TPAFNet: Transformer-driven pyramid attention fusion network for 3D medical image segmentation," IEEE J. Biomed. Health Inform., vol. 28, no. 11, pp. 6803–6814, Nov. 2024. [47] M. A. Marnissi and A. Fathallah, "GAN-based vision transformer for high-quality thermal image enhancement," in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), Vancouver, BC, Canada, 2023, pp. 817–825. [48] Z. Gao, Z. Tong, K. Q. Lin, J. Chen, and M. Z. Shou, "Bootstrapping sparseformers from vision foundation models," in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Seattle, WA, USA, 2024, pp. 17710–17721. [49] S. Lu, W. Zhang, H. Zhao, H. Liu, N. Wang, and H. Li, "Anomaly detection for medical images using heterogeneous auto-encoder," IEEE Trans. Image Process., vol. 33, pp. 2770–2782, 2024. [50] M. S. I. Sajol and A. S. M. J. Hasan, "Benchmarking CNN and cutting-edge transformer models for brain tumor classification through transfer learning," in Proc. IEEE 12th Int. Conf. Intell. Syst. (IS), Varna, Bulgaria, 2024, pp. 1–6. [51] M. Songade, S. Arya, and S. M. Moorthi, "Single band NIR-to-RGB image colorization using attention guided conditional GAN," in Proc. IEEE Space, Aerosp. Defence Conf. (SPACE), Bangalore, India, 2024, pp. 407–410. [52] M. H. Kolekar, S. Bose, and A. Pai, "SARain-GAN: Spatial attention residual UNet based conditional generative adversarial network for rain streak removal," IEEE Access, vol. 12, pp. 43874–43888, 2024. [53] D. Zhang, N. Tang, and Y. Qu, "Joint motion deblurring and super-resolution for single image using diffusion model and GAN," IEEE Signal Process. Lett., vol. 31, pp. 736–740, 2024. [54] D. Martin, C. Fowlkes, D. Tal, and J. Malik, "A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics," in Proc. 8th Int. Conf. Comput. Vis., vol. 2, Jul. 2001, pp. 416–423. [55] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, et al., "PyTorch: An imperative style, high-performance deep learning library," in Advances in Neural Information Processing Systems, vol. 32, 2019, pp. 8026–8037. 描述 碩士
國立政治大學
資訊科學系
112753205資料來源 http://thesis.lib.nccu.edu.tw/record/#G0112753205 資料類型 thesis dc.contributor.advisor 張宏慶 zh_TW dc.contributor.advisor Jang, Hung-Chin en_US dc.contributor.author (Authors) 陸妍諭 zh_TW dc.contributor.author (Authors) Lu, Yen-Yu en_US dc.creator (作者) 陸妍諭 zh_TW dc.creator (作者) Lu, Yen-Yu en_US dc.date (日期) 2025 en_US dc.date.accessioned 1-Jul-2025 15:06:44 (UTC+8) - dc.date.available 1-Jul-2025 15:06:44 (UTC+8) - dc.date.issued (上傳時間) 1-Jul-2025 15:06:44 (UTC+8) - dc.identifier (Other Identifiers) G0112753205 en_US dc.identifier.uri (URI) https://nccur.lib.nccu.edu.tw/handle/140.119/157813 - dc.description (描述) 碩士 zh_TW dc.description (描述) 國立政治大學 zh_TW dc.description (描述) 資訊科學系 zh_TW dc.description (描述) 112753205 zh_TW dc.description.abstract (摘要) 數位影像在傳輸的過程中,可能會受到電磁干擾和攝影元件受損,導致影像受到脈衝雜訊干擾而損壞,如何有效修復遭受極值脈衝雜訊(椒鹽雜訊)干擾的影像,對於提升數位影像品質極為重要。本文提出一種使用自適應分析視窗最小-最大均值池化的多頭自注意力卷積神經網路,去除在傳輸過程中產生的椒鹽雜訊;首先,估測影像的雜訊密度,若為輕度雜訊干擾之影像,乾淨像素足夠,使用多頭自注意力卷積神經網路,計算輸入序列中不同位置像素的權重,並擷取長距離像素間的相關性,預測適合重建的臨域像素;相對的,若是中、高雜訊密度干擾時,受損像素較多,則透過多層的最小值和最大值均值池化濾波器,分別計算最大池化後的影像和最小池化後的兩類影像;最後將處理後的影像重新組合,並進行均值濾波處理;在高雜訊密度的環境中,若分析視窗沒有乾淨像素可以參考,則擴大分析視窗的尺寸,增加鄰域未受干擾像素引入的機率,修復影像中的受損像素。經由實驗結果證明:本文所提出的濾波器可以在各種雜訊密度有效重建受到雜訊干擾的影像,重建效果也優於許多極為先進(state-of-the-art)的演算法。 zh_TW dc.description.abstract (摘要) During the transmission of digital images, electromagnetic interference and sensor defects can result in impulse noise corruption, leading to severe image degradation. Effectively restoring images contaminated by extreme impulse noise, such as salt-and-pepper noise, is crucial for improving digital image quality. This paper proposes a multi-head self-attention neural network enhanced by adaptive min-max mean pooling to remove salt-and-pepper noise introduced during transmission. First, the noise density of the corrupted image is estimated. Suppose the image is subjected to low-level noise and contains a sufficient number of clean pixels; a multi-head self-attention mechanism is employed to compute the attention weights of pixels at different positions in the input sequence. This mechanism captures long-range dependencies to predict suitable neighborhood pixels for reconstruction. In contrast, under moderate to high noise density, where a larger portion of pixels are corrupted, the method utilizes multiple layers of min and max mean pooling filters to compute two feature maps separately based on maximum and minimum pooling operations. The processed images are then fused and further refined using mean filtering. In high-noise scenarios where no clean pixels are available within a small analysis window, the window size is adaptively enlarged to increase the likelihood of including undisturbed neighboring pixels, thereby improving the reconstruction of damaged regions. Experimental results demonstrate that the proposed filtering method can effectively restore noise-corrupted images across various noise levels, outperforming several state-of-the-art algorithms regarding reconstruction quality. en_US dc.description.tableofcontents 謝辭 I 摘要 II ABSTRACT III 目錄 IV 圖目錄 VI 表目錄 IX 第一章 緒論 1 1.1 研究動機與目的 1 1.2 文獻探討 1 1.2.1 傳統濾波方法的改進與應用 1 1.2.2選擇性濾波技術的創新研究 3 1.2.3深度學習技術在影像修復中的應用 4 1.2.4深度學習技術在醫學影像中的應用 5 1.3章節結構 6 第二章 數位影像的雜訊移除方法 7 2.1數位影像表示法 7 2.2常見的影像雜訊 8 2.2.1脈衝雜訊 8 2.2.2高斯雜訊 9 2.2.3均勻雜訊 9 2.2.4雷萊雜訊 10 2.3灰階影像的雜訊生成 11 2.4檢測影像中的極值像素與鏡射方法 11 2.4.1偵測雜訊像素 12 2.4.2鏡射 13 2.5數位影像雜訊濾波演算法 13 2.5.1中值濾波器 14 2.5.2均值濾波器 14 2.5.3切換式自適性權重均值濾波器 15 2.5.4脈衝雜訊偵測用於開關中值濾波器 16 2.5.5中間值權重中位數濾波器 17 2.5.6方向權重中值濾波器 18 2.5.7修正型之方向權重中值濾波 19 2.5.8修正型非對稱剪枝中值濾波器 21 2.5.9可變視窗與像素機率調適之方法 22 2.6深度學習於影像雜訊濾波演算法 24 2.6.1卷積自動編碼器去雜訊中值濾波器 24 2.6.2使用卷積神經網路進行影像雜訊移除方法 25 2.6.3使用DLNN辨識像素變化方向的雜訊移除方法 26 2.6.4使用選擇性卷積塊之雜訊移除方法 27 2.6.5使用深度學習全連接神經網路均值濾波器進行影像增強 28 第三章 結合最小-最大均值池化濾波與多頭自注意力卷積神經網路之影像重建系統 29 3.1 不同雜訊密度的影像修復演算法 31 3.2 卷積神經網路介紹 32 3.3 多頭自注意力卷積神經網路的訓練影像蒐集方法 34 3.4 多頭自注意力卷積神經網路的雜訊移除方法 41 3.5 多頭自注意力卷積神經網路的設計架構 42 3.6 多頭自注意力卷積神經網路修復受損像素 44 3.7 最小-最大均值池化濾波 47 3.8 自適應分析視窗最小-最大均值池化與多頭自注意力卷積神經網路濾波器 50 第四章 實驗結果 51 4.1 實驗環境 51 4.2 PSNR效能評估 52 4.3 MSSIM效能評估 59 4.4 BOAT影像遭受固定式脈衝雜訊干擾之實驗結果 68 4.5 BABOON影像遭受固定式脈衝雜訊干擾之實驗結果 80 4.6 CAMERAMAN影像遭受固定式脈衝雜訊干擾之實驗結果 92 4.7 LAKE影像遭受固定式脈衝雜訊干擾之實驗結果 104 第五章 結論與未來研究方向 116 5.1 結論 116 5.2 未來研究方向 116 參考文獻 118 附錄一 符號對照表 125 附錄二 論文發表列表 128 zh_TW dc.format.extent 12720780 bytes - dc.format.mimetype application/pdf - dc.source.uri (資料來源) http://thesis.lib.nccu.edu.tw/record/#G0112753205 en_US dc.subject (關鍵詞) 多頭自注意力神經網路 zh_TW dc.subject (關鍵詞) 卷積神經網路 zh_TW dc.subject (關鍵詞) 深度學習 zh_TW dc.subject (關鍵詞) 均值濾波器 zh_TW dc.subject (關鍵詞) 椒鹽雜訊 zh_TW dc.subject (關鍵詞) 池化濾波 zh_TW dc.subject (關鍵詞) multi-head self-attention neural network en_US dc.subject (關鍵詞) convolutional neural network en_US dc.subject (關鍵詞) deep learning en_US dc.subject (關鍵詞) mean filter en_US dc.subject (關鍵詞) salt-and-pepper noise en_US dc.subject (關鍵詞) pooling-based filtering en_US dc.title (題名) 透過最小-最大均值池化濾波器與多頭自注意力卷積神經網路之影像雜訊移除研究 zh_TW dc.title (題名) Image Denoising Using Min-Max Mean Pooling Filters and Multi-Head Self-Attention Convolutional Neural Networks en_US dc.type (資料類型) thesis en_US dc.relation.reference (參考文獻) [1] R. C. Gonzalez, R. E. Woods, and S. L. Eddins, Digital Image Processing, 2nd ed. New York, NY, USA: Prentice-Hall, 2002. [2] J. D. Gibson and A. Bovik, Handbook of Image and Video Processing, 1st ed. San Diego, CA, USA: Academic Press, 2000. [3] H. Hwang and R. A. Haddad, "Adaptive median filters: new algorithms and results," IEEE Trans. Image Process., vol. 4, no. 4, pp. 499–502, Apr. 1995. [4] S.-J. Ko and Y. H. Lee, "Center weighted median filters and their applications to image enhancement," IEEE Trans. Circuits Syst., vol. 38, no. 9, pp. 984–993, Sep. 1991. [5] Y. Dong and S. Xu, "A new directional weighted median filter for removal of random-valued impulse noise," IEEE Signal Process. Lett., vol. 14, no. 3, pp. 193–196, Mar. 2007. [6] T. Chen and H. R. Wu, "Adaptive impulse detection using center-weighted median filters," IEEE Signal Process. Lett., vol. 8, no. 1, pp. 1–3, Jan. 2001. [7] C.-T. Lu and T.-C. Chou, "Denoising of salt-and-pepper noise corrupted image using modified directional-weighted-median filter," Pattern Recognit. Lett., vol. 33, no. 10, pp. 1287–1295, Jul. 2012. [8] S. Esakkirajan, T. Veerakumar, A. N. Subramanyam, and C. H. PremChand, "Removal of high density salt and pepper noise through modified decision based unsymmetric trimmed median filter," IEEE Signal Process. Lett., vol. 18, no. 5, pp. 287–290, May 2011. [9] T. Chen, K.-K. Ma, and L.-H. Chen, "Tri-state median filter for image denoising," IEEE Trans. Image Process., vol. 8, no. 12, pp. 1834–1838, Dec. 1999. [10] P. Satti, N. Sharma, and B. Garg, "Min-max average pooling based filter for impulse noise removal," IEEE Signal Process. Lett., vol. 27, pp. 1475–1479, 2020. [11] Y.-Y. Lu and H.-C. Jang, "Removal of salt-and-pepper impulse noise using adaptive analysis windows with min-max mean pooling," in Proc. 36th IPPR Conf. Comput. Vis. Graph. Image Process., 2023. [12] L. Zhai, L. Dong, Y. Liu, S. Fu, F. Wang, and Y. Li, "Adaptive hybrid threshold shrinkage using singular value decomposition," J. Electron. Imaging, vol. 28, no. 6, pp. 063002, Nov. 2019. [13] S. H. Chan, T. Zickler, and Y. M. Lu, "Monte Carlo non-local means: Random sampling for large-scale image filtering," IEEE Trans. Image Process., vol. 23, no. 8, pp. 3711–3725, Aug. 2014. [14] E. Luo, S. H. Chan, and T. Q. Nguyen, "Adaptive image denoising by mixture adaptation," IEEE Trans. Image Process., vol. 25, no. 10, pp. 4489–4503, Oct. 2016. [15] S. Anwar, F. Porikli, and C. P. Huynh, "Category-specific object image denoising," IEEE Trans. Image Process., vol. 26, no. 11, pp. 5506–5518, Nov. 2017. [16] J. Xu, L. Zhang, and D. Zhang, "External prior guided internal prior learning for real-world noisy image denoising," IEEE Trans. Image Process., vol. 27, no. 6, pp. 2996–3010, Jun. 2018. [17] M. S. M. Zain, A. K. Junoh, and A. Abdurrazzaq, "Hybrid singular value decomposition based alpha trimmed mean-median filter in eliminating high density salt and pepper noise from grayscale image," Multimed. Tools Appl., vol. 83, pp. 62895–62913, 2024. [18] J. Gao, L. Li, X. Ren, Q. Chen, and Y. M. Abdul-Abbass, "An effective method for salt and pepper noise removal based on algebra and fuzzy logic function," Multimed. Tools Appl., vol. 83, pp. 9547–9576, 2024. [19] Q. Liu, X. Li, and J. Yang, "Optimum codesign for image denoising between type-2 fuzzy identifier and matrix completion denoiser," IEEE Trans. Fuzzy Syst., vol. 30, no. 1, pp. 287–292, Jan. 2022. [20] X. Zhang, X. Li, Z. Tang, S. Zhang, and S. Xie, "Noise removal in embedded image with bit approximation," IEEE Trans. Knowl. Data Eng., vol. 34, no. 3, pp. 1359–1369, Mar. 2022. [21] N. N. Hien, D. N. H. Thanh, U. Erkan, and J. M. R. S. Tavares, "Image noise removal method based on thresholding and regularization techniques," IEEE Access, vol. 10, pp. 71584–71597, 2022. [22] N. Aslam, M. K. Ehsan, Z. U. Rehman, M. Hanif, and G. Mustafa, "A modified form of different applied median filter for removal of salt & pepper noise," Multimed. Tools Appl., vol. 82, pp. 7479–7490, 2023. [23] S. Zhang and M. A. Karim, "A new impulse detector for switching median filters," IEEE Signal Process. Lett., vol. 9, no. 11, pp. 360–363, Nov. 2002. [24] C.-T. Lu, Y.-Y. Chen, L.-L. Wang, and C.-F. Chang, "Removal of salt-and-pepper noise in corrupted image using three-values-weighted approach with variable-size window," Pattern Recognit. Lett., vol. 80, pp. 188–199, Sep. 2016. [25] D. Liu, B. Wen, J. Jiao, X. Liu, Z. Wang, and T. S. Huang, "Connecting image denoising and high-level vision tasks via deep learning," IEEE Trans. Image Process., vol. 29, pp. 3695–3706, 2020. [26] Z. Liu, W. Q. Ya, and M. L. Yang, "Image denoising based on a CNN model," in Proc. Int. Conf. Control, Autom. Robot., 2018, pp. 389–393. [27] A. V. Miclea, R. Terebes, I. Ilea, and M. Borda, "Hyperspectral image classification using combined spectral-spatial denoising and deep learning techniques," in Proc. IEEE Int. Conf. Autom., Qual. Testing, Robot. (AQTR), 2018. [28] K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, "Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising," IEEE Trans. Image Process., vol. 26, no. 7, pp. 3142–3155, Jul. 2017. [29] X. Li, Q. Gao, Y. Lu, and D. Sun, "Convolutional neural network denoising method based on multisize features," in Proc. IEEE Int. Conf. Intell. Human-Machine Syst. Cybern., 2018, pp. 146–149. [30] K. Zhang, W. Zuo, and L. Zhang, "FFDNet: Toward a fast and flexible solution for CNN-based image denoising," IEEE Trans. Image Process., vol. 27, no. 9, pp. 4608–4622, Sep. 2018. [31] T. Wang, M. Sun, and K. Hu, "Dilated deep residual network for image denoising," in Proc. IEEE Int. Conf. Tools Artif. Intell., 2017, pp. 1272–1279. [32] K. Isogawa, T. Shiodera, and T. Takeguchi, "Deep shrinkage convolutional neural network for adaptive noise reduction," IEEE Signal Process. Lett., vol. 25, no. 2, pp. 224–228, Feb. 2018. [33] E. Kandić, A. Akagic, and M. Bohlouli, "Exploring convolutional autoencoder efficacy in noise removal for image processing and computer vision: A study using the MNIST dataset," in Proc. 10th Int. Conf. Control, Decision Inf. Technol. (CoDIT), Vallette, Malta, 2024, pp. 1275–1280. [34] L. Deng, "The MNIST database of handwritten digit images for machine learning research," IEEE Signal Process. Mag., vol. 29, no. 6, pp. 141–142, Nov. 2012. [35] C.-T. Lu, R.-H. Chen, L.-L. Wang, and J.-A. Lin, "Image enhancement using convolutional neural network to identify similar patterns," IET Image Process., vol. 14, no. 17, pp. 3880–3889, 2020. [36] C.-T. Lu, H.-J. Hsu, and L.-L. Wang, "Image denoising using DLNN to recognize the direction of pixel variation," Signal Image Video Process., vol. 15, pp. 1247–1256, 2021. [37] A. A. Rafiee and M. Farhang, "A deep convolutional neural network for salt-and-pepper noise removal using selective convolutional blocks," Appl. Soft Comput., vol. 145, 2023. [38] C.-T. Lu, L.-L. Wang, J.-H. Shen, and J. A. Lin, "Image enhancement using deep-learning fully connected neural network mean filter," J. Supercomput., vol. 77, pp. 3144–3164, 2021. [39] Q. Yang, P. Yan, M. K. Kalra, and G. Wang, "CT image denoising with perceptive deep neural networks," arXiv preprint, arXiv:1702.07019, 2017. [40] K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," arXiv preprint arXiv:1409.1556, 2014. [41] E. Kang and J. C. Ye, "Framelet denoising for low-dose CT using deep learning," in Proc. IEEE Int. Symp. Biomed. Imaging (ISBI), 2018, pp. 311–314. [42] H. Yuan, J. Jia, and Z. Zhu, "SIPID: A deep learning framework for sinogram interpolation and image denoising in low-dose CT reconstruction," in Proc. IEEE Int. Symp. Biomed. Imaging (ISBI), 2018, pp. 1521–1524. [43] Q. Yang, P. Yan, Y. Zhang, H. Yu, Y. Shi, X. Mou, M. K. Kalra, Y. Zhang, L. Sun, and G. Wang, "Low-dose CT image denoising using a generative adversarial network with Wasserstein distance and perceptual loss," IEEE Trans. Med. Imaging, vol. 37, no. 6, pp. 1348–1357, Jun. 2018. [44] M. A. Marnissi, "Revolutionizing thermal imaging: GAN-based vision transformers for image enhancement," in Proc. IEEE Int. Conf. Image Process. (ICIP), Kuala Lumpur, Malaysia, 2023, pp. 2735–2739. [45] S. I. Jang, T. Pan, Y. Li, P. Heidari, and J. Chen, "Spach Transformer: Spatial and channel-wise transformer based on local and global self-attentions for PET image denoising," IEEE Trans. Med. Imaging, vol. 43, no. 6, pp. 2036–2049, Jun. 2024. [46] Z. Li, J. Zhang, S. Wei, Y. Gao, C. Cao, and Z. Wu, "TPAFNet: Transformer-driven pyramid attention fusion network for 3D medical image segmentation," IEEE J. Biomed. Health Inform., vol. 28, no. 11, pp. 6803–6814, Nov. 2024. [47] M. A. Marnissi and A. Fathallah, "GAN-based vision transformer for high-quality thermal image enhancement," in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), Vancouver, BC, Canada, 2023, pp. 817–825. [48] Z. Gao, Z. Tong, K. Q. Lin, J. Chen, and M. Z. Shou, "Bootstrapping sparseformers from vision foundation models," in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Seattle, WA, USA, 2024, pp. 17710–17721. [49] S. Lu, W. Zhang, H. Zhao, H. Liu, N. Wang, and H. Li, "Anomaly detection for medical images using heterogeneous auto-encoder," IEEE Trans. Image Process., vol. 33, pp. 2770–2782, 2024. [50] M. S. I. Sajol and A. S. M. J. Hasan, "Benchmarking CNN and cutting-edge transformer models for brain tumor classification through transfer learning," in Proc. IEEE 12th Int. Conf. Intell. Syst. (IS), Varna, Bulgaria, 2024, pp. 1–6. [51] M. Songade, S. Arya, and S. M. Moorthi, "Single band NIR-to-RGB image colorization using attention guided conditional GAN," in Proc. IEEE Space, Aerosp. Defence Conf. (SPACE), Bangalore, India, 2024, pp. 407–410. [52] M. H. Kolekar, S. Bose, and A. Pai, "SARain-GAN: Spatial attention residual UNet based conditional generative adversarial network for rain streak removal," IEEE Access, vol. 12, pp. 43874–43888, 2024. [53] D. Zhang, N. Tang, and Y. Qu, "Joint motion deblurring and super-resolution for single image using diffusion model and GAN," IEEE Signal Process. Lett., vol. 31, pp. 736–740, 2024. [54] D. Martin, C. Fowlkes, D. Tal, and J. Malik, "A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics," in Proc. 8th Int. Conf. Comput. Vis., vol. 2, Jul. 2001, pp. 416–423. [55] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, et al., "PyTorch: An imperative style, high-performance deep learning library," in Advances in Neural Information Processing Systems, vol. 32, 2019, pp. 8026–8037. zh_TW
