Learning Hierarchical Semantic Information for Efficient Low-Light Image Enhancement

Publisher:
Institute of Electrical and Electronics Engineers (IEEE)
Publication Type:
Conference Proceeding
Citation:
Proceedings of the International Joint Conference on Neural Networks, 2023, 2023-June, pp. 1-8
Issue Date:
2023-01-01
Full metadata record
Low-light environments can cause a variety of complex degradation problems, which result in poor visibility in images. As a classical vision task, low-light image enhancement has attracted an increasing interest in the research community. However, the existing methods tend to require a large number of parameters, making them difficult to implement and optimize, especially on resource-constrained devices. In this paper, we mainly focus on the lightweight of the method and propose a novel end-to-end two-stage CNN-ViT architecture (HSINet) to learn hierarchical semantic information (HSI) from low-light images efficiently. The HSINet consists of two stages: the first stage is a CNN-based low-level semantic (LS) Stage, and the second stage is ViT-based high-level semantic (HS) Stage. The LS Stage contains an efficient multi-scale convolution block, MLS Block, for low-level semantic information extraction. The HS stage, on the other hand, aims to learn the high-level semantic features via ViT's excellent global-learning capability. We propose a hierarchical Swin Transformer-based block, HS Block, to gradually enlarge Swin Transformer's window size as the network becomes deeper, to learn hierarchical high-level semantic information. Benefiting from the efficient architecture, our model only contains 0.6M parameters, far fewer than the existing SOTAs. We evaluated the method on three challenging benchmark datasets: LOL, VE-LOL, and MIT-Adobe FiveK, using three popular evaluation metrics. The quantitative and qualitative results both show that the proposed method not only outperforms the state of the arts in terms of PSNR, SSIM, LPIPS, and visual effects, but also with better efficiency.
Please use this identifier to cite or link to this item: