Learning to Detect Mirrors from Videos via Dual Correspondences
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Title of host publication | Proceedings - 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023 |
Publisher | Institute of Electrical and Electronics Engineers, Inc. |
Pages | 9109-9118 |
ISBN (electronic) | 9798350301298 |
ISBN (print) | 9798350301304 |
Publication status | Published - 2023 |
Publication series
Name | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
---|---|
Volume | 2023-June |
ISSN (Print) | 1063-6919 |
ISSN (electronic) | 2575-7075 |
Conference
Title | 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023) |
---|---|
Location | Vancouver Convention Center |
Place | Canada |
City | Vancouver |
Period | 18 - 22 June 2023 |
Link(s)
DOI | DOI |
---|---|
Document Link | |
Link to Scopus | https://www.scopus.com/record/display.uri?eid=2-s2.0-85169567378&origin=recordpage |
Permanent Link | https://scholars.cityu.edu.hk/en/publications/publication(d9234377-7941-426c-bc47-db1db94b52a6).html |
Abstract
Detecting mirrors from static images has received significant research interest recently. However, detecting mirrors over dynamic scenes is still under-explored due to the lack of a high-quality dataset and an effective method for video mirror detection (VMD). To the best of our knowledge, this is the first work to address the VMD problem from a deep-learning-based perspective. Our observation is that there are often correspondences between the contents inside (reflected) and outside (real) of a mirror, but such correspondences may not always appear in every frame, e.g., due to the change of camera pose. This inspires us to propose a video mirror detection method, named VMD-Net, that can tolerate spatially missing correspondences by considering the mirror correspondences at both the intra-frame level as well as inter-frame level via a dual correspondence module that looks over multiple frames spatially and temporally for correlating correspondences. We further propose a first large-scale dataset for VMD (named VMD-D), which contains 14,987 image frames from 269 videos with corresponding manually annotated masks. Experimental results show that the proposed method outperforms SOTA methods from relevant fields. To enable real-time VMD, our method efficiently utilizes the backbone features by removing the redundant multi-level module design and gets rid of post-processing of the output maps commonly used in existing methods, making it very efficient and practical for real-time video-based applications. Code, dataset, and models are available at https://jiaying.link/cvpr2023-vmd/ © 2023 IEEE.
Research Area(s)
- Low-level vision
Citation Format(s)
Learning to Detect Mirrors from Videos via Dual Correspondences. / Lin, Jiaying; Tan, Xin; Lau, Rynson W.H.
Proceedings - 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023. Institute of Electrical and Electronics Engineers, Inc., 2023. p. 9109-9118 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition; Vol. 2023-June).
Proceedings - 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023. Institute of Electrical and Electronics Engineers, Inc., 2023. p. 9109-9118 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition; Vol. 2023-June).
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review