Bringing Old Films Back to Life
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Title of host publication | Proceedings - 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition |
Subtitle of host publication | CVPR 2022 |
Publisher | Institute of Electrical and Electronics Engineers, Inc. |
Pages | 17673-17682 |
Number of pages | 10 |
ISBN (electronic) | 9781665469463 |
ISBN (print) | 978-1-6654-6947-0 |
Publication status | Published - 2022 |
Publication series
Name | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
---|---|
ISSN (Print) | 1063-6919 |
ISSN (electronic) | 2575-7075 |
Conference
Title | 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022) |
---|---|
Location | Hybrid |
Place | United States |
City | New Orleans |
Period | 19 - 24 June 2022 |
Link(s)
DOI | DOI |
---|---|
Document Link | |
Link to Scopus | https://www.scopus.com/record/display.uri?eid=2-s2.0-85141805976&origin=recordpage |
Permanent Link | https://scholars.cityu.edu.hk/en/publications/publication(7b713898-ebbb-4c67-874d-912c4310a7c0).html |
Abstract
We present a learning-based framework, recurrent transformer network (RTN), to restore heavily degraded old films. Instead of performing frame-wise restoration, our method is based on the hidden knowledge learned from adjacent frames that contain abundant information about the occlusion, which is beneficial to restore challenging artifacts of each frame while ensuring temporal coherency. Moreover, contrasting the representation of the current frame and the hidden knowledge makes it possible to infer the scratch position in an unsupervised manner, and such defect localization generalizes well to real-world degradations. To better resolve mixed degradation and compensate for the flow estimation error during frame alignment, we propose to leverage more expressive transformer blocks for spatial restoration. Experiments on both synthetic dataset and real-world old films demonstrate the significant superiority of the proposed RTN over existing solutions. In addition, the same framework can effectively propagate the color from keyframes to the whole video, ultimately yielding compelling restored films.
Research Area(s)
- Computational photography, Low-level vision
Bibliographic Note
Research Unit(s) information for this publication is provided by the author(s) concerned.
Citation Format(s)
Bringing Old Films Back to Life. / Wan, Ziyu; Zhang, Bo; Chen, Dongdong et al.
Proceedings - 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition: CVPR 2022. Institute of Electrical and Electronics Engineers, Inc., 2022. p. 17673-17682 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition).
Proceedings - 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition: CVPR 2022. Institute of Electrical and Electronics Engineers, Inc., 2022. p. 17673-17682 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition).
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review