The International Arab Journal of Information Technology (IAJIT)

..............................
..............................
..............................


Incorporating Triple Attention and Multi-scale Pyramid Network for Underwater Image Enhancement

Clear images are a prerequisite of high-level underwater vision tasks, but images captured underwater are often degraded due to absorption and scattering of light. To solve this issue, traditional methods have shown some success, but often generate unwanted artifacts for knowledge priori dependency. In contrast, learning-based approaches can produce more refined results. Most popular methods are based on an encoder-decoder configuration for simply learning the nonlinear transformation of input and output images, so their ability to capture details is limited. In addition, the significant pixel-level features and multi- scale features are often overlooked. Accordingly, we propose a novel and efficient network that incorporates triple attention and a multi-scale pyramid with an encoder-decoder architecture. Specifically, a triple attention module that captures the channel- pixel-spatial features is used as the transformation of the encoder-decoder module to focus on the fog region; then, a multi-scale pyramid module designed for refining the preliminary defog results are used to improve the restoration visibility. Extensive experiments on the EUVP and UFO-120 datasets corroborate that the proposed method outperforms the state-of-the-art methods in quantitative metrics Peak Signal-to-Noise Ratio (PSNR), Structural Similarity (SSIM), Patch-based Contrast Quality Index (PCQI) and visual quality.

  1. Ancuti C., Ancuti C., Vleeschouwer C., and Bekaert P., “Color Balance and Fusion for Underwater Image Enhancement,” IEEE Transactions on Image Processing, vol. 27, no. 1, pp. 379-393, 2018.
  2. Chen X., Zhang P., Quan L., Yi C., and Lu C., “Underwater Image Enhancement based on Deep Learning and Image Formation Model,” arXiv preprint arXiv:2101.00991, 2021.
  3. Fabbri C., Islam M., and Sattar J., “Enhancing Underwater Imagery using Generative Adversarial Networks,” in Proceedings of IEEE International Conference on Robotics and Automation, Brisbane, pp. 7159-7165, 2018.
  4. Gaude G. and Borkar S., “Fish Detection and Tracking for Turbid Underwater Video,” in Proceedings of the International Conference on Intelligent Computing and Control Systems, Madurai, pp. 326-331, 2019.
  5. Guo Q., Xue L., Tang R., and Guo L., “Underwater Image Enhancement Based on the Dark Channel Prior and Attenuation Compensation,” Journal of Ocean University of China, vol. 16, no. 5, pp. 757-765, 2017.
  6. Haulsee D., Breece M., Miller D., Wetherbee B., Fox D., and Oliver M., “Habitat Selection of a Coastal Shark Species Estimated from an Autonomous Underwater Vehicle,” Marine Ecology Progress Series, vol. 528, pp. 277-288, 2015.
  7. Hore A. and Ziou D., “Image Quality Metrics: PSNR vs. SSIM,” in Proceedings of the International Conference on Learning Representations, Istanbul, pp.2366-2369, 2010.
  8. Islam M., Luo P., and Sattar J., “Simultaneous Enhancement and Super-Resolution of Underwater Imagery for Improved Visual Perception,” in Proceedings of the Robotics: Science and Systems, Corvalis, 2020.
  9. Islam M., Xia Y., and Sattar J., “Fast Underwater Image Enhancement for Improved Visual Perception,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 3227-3234, 2020.
  10. Jaffe J., “Computer Modeling and the Design of Optimal Underwater Imaging Systems,” IEEE Journal of Oceanic Engineering, vol. 15, no. 2, pp. 101-111, 1990.
  11. Jamadandi A., and Mudenagudi U., “Exemplar-based Underwater Image Enhancement Augmented by Wavelet Corrected Transforms,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, pp. 11-17, 2019.
  12. Li C., Guo C., Ren W., Cong R., Hou J., Kwong S., and Tao D., “An Underwater Image Enhancement Benchmark Dataset and Beyond,” IEEE Transactions on Image Processing, vol. 29, pp. 4376-4389, 2019.
  13. Loshchilov I. and Hutter F., “SGDR: Stochastic Gradient Descent with Warm Restarts,” in Proceedings of the International Conference on Learning Representations, San Juan, 2016.
  14. McGlamery B., “A Computer Model for Underwater Camera Systems,” Ocean Optics VI. International Society for Optics and Photonics, vol. 208, pp. 221-231, 1980.
  15. Mei K., Jiang A., Li J., and Wang M., “Progressive Feature Fusion Network for Realistic Image Dehazing,” in Proceedings of the Asian Conference on Computer Vision, Cham, pp. 203-215, 2018.
  16. Mnih V., Heess N., Graves A., and Kavukcuoglu K., “Recurrent Models of Visual Attention,” in Proceedings of the Advances in Neural Information Processing Systems, Montreal, pp. 2204-2212, 2014.
  17. Naik A., Swarnakar A., and Mittal K., “Shallow-UWnet: Compressed Model for Underwater Image Enhancement,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, vol. 35, no. 1, pp. 15853-15854, 2021.
  18. Ortiz A., Simó M., and Oliver G., “A Vision System for an Underwater Cable Tracker,” Machine Vision and Applications. vol. 13, no. 3, 129-140, 2002.
  19. Panetta K., Gao C., and Agaian S., “Human-Visual-System-Inspired Underwater Image Quality Measures,” IEEE Journal of Oceanic Engineering, vol. 41, no. 3, pp. 541-551, 2016.
  20. Peng Y. and Cosman P., “Underwater Image Restoration Based on Image Blurriness and Light Absorption,” IEEE Transactions on Image Processing, vol. 26, no. 4, pp. 1579-1594, 2017.
  21. Qin X., Wang Z., Bai Y., Xie X., and Jia H., “FFA-Net: Feature Fusion Attention Network for Single Image Dehazing,” in Proceedings of AAAI Conference on Artificial Intelligence, New York, pp. 11908-11915, 2020.
  22. Song W., Wang Y., Huang D., and Tjondronegoro D., “A Rapid Scene Depth Estimation Model Based on Underwater Light Attenuation Prior for Underwater Image Restoration,” in Proceedings of Pacific Rim Conference on Multimedia, Hefei, pp. 678-688, 2018.
  23. Teixeira B., Silva H., Matos A., and Silva E., “Deep Learning Approaches Assessment for Underwater Scene Understanding and Egomotion Estimation,” in Proceedings of IEEE OCEANS, Seattle, pp. 1-9, 2019.
  24. Uplavikar P., Wu Z., and Wang Z., “All-in-One Underwater Image Enhancement Using Domain-Adversarial Learning,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seoul Long Beach, pp. 1-8, 2019.
  25. Wang F., Jiang M., Qian C., Yang S., Li C., Zhang H., Wang X., and Tang X., “Residual Attention Network for Image Classification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, pp. 3156-3164, 2017.
  26. Wang S., Ma K., Yeganeh H., Wang Z., and Lin W., “A Patch-Structure Representation Method for Quality Assessment of Contrast Changed Images,” IEEE Signal Processing Letters, vol. 22, no. 12, pp. 2387-2390, 2015.
  27. Wang Z., Bovik A., Sheikh H., and Simoncelli E., “Image Quality Assessment from Error Visibility to Structural Similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600-612, 2004.
  28. Woo S., Park J., Lee J., and Kweon I., “CBAM: Convolutional Block Attention Module,” in Proceedings of the European Conference on Computer Vision, Munich, pp. 3-19, 2018.
  29. Yin S., Wang Y., and Yang Y., “A Novel Image-Dehazing Network with a Parallel Attention Block,” Pattern Recognition, vol. 102, pp. 107255, 2020.
  30. Zamir S., Arora A., Khan S., Hayat M., Khan F., Yang M., and Shao L., “Learning Enriched Features for Real Image Restoration and Enhancement,” in Proceedings of the European Conference on Computer Vision, Glasgow, pp. 492-511, 2020.
  31. Zhang H. and Patel V., “Densely Connected Pyramid Dehazing Network,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, pp. 3194-3203, 2018.
  32. Zhao H., Gallo O., Frosio I., and Kautz J., “Loss Functions for Image Restoration with Neural Networks,” IEEE Transactions on Computational Imaging, vol. 3, no. 1, pp. 47-57, 2016.
  33. Zhao H., Shi J., Qi X., Wang X., and Jia J., “Pyramid Scene Parsing Network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, pp. 2881-2890, 2017.
  34. Zhou W. and Jia J., “Training Convolutional Neural Network for Sketch Recognition on Large-Scale Dataset,” The International Arab Journal of Information Technology, vol. 17, no. 1, pp. 82-89, 2020.