LDP
Date
Authors
Yang, Hao
Pan, Liyuan
Yang, Yan
Hartley, Richard
Liu, Miaomiao
Journal Title
Journal ISSN
Volume Title
Publisher
Access Statement
Abstract
Recovering sharp images from dual-pixel (DP) pairs with disparity-dependent blur is a challenging task. Existing blur map-based deblurring methods have demonstrated promising results. In this paper, we propose, to the best of our knowledge, the first framework that introduces the contrastive language-image pre-training framework (CLIP) to accurately estimate the blur map from a DP pair unsu-pervisedly. To achieve this, we first carefully design text prompts to enable CLIP to understand blur-related geo-metric prior knowledge from the DP pair. Then, we pro-pose a format to input a stereo DP pair to CLIP without any fine-tuning, despite the fact that CLIP is pre-trained on monocular images. Given the estimated blur map, we intro-duce a blur-prior attention block, a blur-weighting loss, and a blur-aware loss to recover the all-in-focus image. Our method achieves state-of-the-art performance in extensive experiments (see Fig. 1).
Description
Keywords
Citation
Collections
Source
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
Type
Book Title
Entity type
Publication