Enhancing underwater video from consecutive frames while preserving temporal consistency

[thumbnail of Open Access]
Preview
Text (Open Access) - Published Version
· Available under License Creative Commons Attribution.
· Please see our End User Agreement before downloading.
| Preview
Available under license: Creative Commons Attribution

Please see our End User Agreement.

It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing.

Add to AnyAdd to TwitterAdd to FacebookAdd to LinkedinAdd to PinterestAdd to Email

Hu, K. orcid id iconORCID: https://orcid.org/0000-0001-7181-9935, Meng, Y. orcid id iconORCID: https://orcid.org/0000-0001-6901-8282, Liao, Z. orcid id iconORCID: https://orcid.org/0009-0006-4686-3436, Tang, L. orcid id iconORCID: https://orcid.org/0009-0003-2401-8520 and Ye, X. orcid id iconORCID: https://orcid.org/0000-0002-1628-2060 (2025) Enhancing underwater video from consecutive frames while preserving temporal consistency. Journal of Marine Science and Engineering, 13 (1). 127. ISSN 2077-1312 doi: 10.3390/jmse13010127

Abstract/Summary

Current methods for underwater image enhancement primarily focus on single-frame processing. While these approaches achieve impressive results for static images, they often fail to maintain temporal coherence across frames in underwater videos, which leads to temporal artifacts and frame flickering. Furthermore, existing enhancement methods struggle to accurately capture features in underwater scenes. This makes it difficult to handle challenges such as uneven lighting and edge blurring in complex underwater environments. To address these issues, this paper presents a dual-branch underwater video enhancement network. The network synthesizes short-range video sequences by learning and inferring optical flow from individual frames. It effectively enhances temporal consistency across video frames through predicted optical flow information, thereby mitigating temporal instability within frame sequences. In addition, to address the limitations of traditional U-Net models in handling complex multiscale feature fusion, this study proposes a novel underwater feature fusion module. By applying both max pooling and average pooling, this module separately extracts local and global features. It utilizes an attention mechanism to adaptively adjust the weights of different regions in the feature map, thereby effectively enhancing key regions within underwater video frames. Experimental results indicate that when compared with the existing underwater image enhancement baseline method and the consistency enhancement baseline method, the proposed model improves the consistency index by 30% and shows a marginal decrease of only 0.6% in enhancement quality index, demonstrating its superiority in underwater video enhancement tasks.

Altmetric Badge

Item Type Article
URI https://reading-clone.eprints-hosting.org/id/eprint/120398
Identification Number/DOI 10.3390/jmse13010127
Refereed Yes
Divisions Science > School of Mathematical, Physical and Computational Sciences > Department of Computer Science
Publisher MDPI
Download/View statistics View download statistics for this item

Downloads

Downloads per month over past year

University Staff: Request a correction | Centaur Editors: Update this record

Search Google Scholar