Tokenize Image Patches: Global Context Fusion for Effective Haze Removal in Large Images
Abstract
Global contextual information and local detail features are essential for haze removal tasks. Deep learning models perform well on small, low-resolution images, but they encounter difficulties with large, high-resolution ones due to GPU memory limitations. As a compromise, they often resort to image slicing or downsampling. The former diminishes global information, while the latter discards high-frequency details. To address these challenges, we propose DehazeXL, a haze removal method that effectively balances global context and local feature extraction, enabling end-to-end modeling of large images on mainstream GPU hardware. Additionally, to evaluate the efficiency of global context utilization in haze removal performance, we design a visual attribution method tailored to the characteristics of haze removal tasks. Finally, recognizing the lack of benchmark datasets for haze removal in large images, we have developed an ultra-high-resolution haze removal dataset (8KDehaze) to support model training and testing. It includes 10000 pairs of clear and hazy remote sensing images, each sized at 8192 times 8192 pixels. Extensive experiments demonstrate that DehazeXL can infer images up to 10240 times 10240 pixels with only 21 GB of memory, achieving state-of-the-art results among all evaluated methods. The source code and experimental dataset are available at https://github.com/CastleChen339/DehazeXL.
Community
In this paper, we propose DehazeXL, an end-to-end haze removal method that effectively integrates global information interaction with local feature extraction, enabling efficient processing of large images while minimizing GPU memory usage. To evaluate the efficiency of global context utilization in haze removal performance, we designed a visual attribution method for haze removal tasks. Furthermore, we develop an ultra-high-resolution haze removal dataset 8KDehaze. Quantitative and qualitative results demonstrate that the proposed DehazeXL outperforms state-of-the-art haze removal techniques in terms of accuracy and inference speed across multiple high-resolution datasets. The results of attribution analysis highlight the significance of global information in the context of image dehazing tasks.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- An Efficient and Mixed Heterogeneous Model for Image Restoration (2025)
- Crafting Query-Aware Selective Attention for Single Image Super-Resolution (2025)
- Content-Aware Transformer for All-in-one Image Restoration (2025)
- Enhanced Semantic Extraction and Guidance for UGC Image Super Resolution (2025)
- MaskAttn-UNet: A Mask Attention-Driven Framework for Universal Low-Resolution Image Segmentation (2025)
- Learning Hazing to Dehazing: Towards Realistic Haze Generation for Real-World Image Dehazing (2025)
- Semantic-Guided Global–Local Collaborative Networks for Lightweight Image Super-Resolution (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper