Abstract
Hair editing is an interesting and challenging problem in computer vision and graphics. Many existing methods require well-drawn sketches or masks as conditional inputs for editing, however these interactions are neither straight-forward nor efficient. In order to free users from the tedious interaction process, this paper proposes a new hair editing interaction mode, which enables manipulating hair attributes individually or jointly based on the texts or reference images provided by users. For this purpose, we encode the image and text conditions in a shared embedding space and propose a unified hair editing framework by leveraging the powerful image text representation capability of the Contrastive Language-Image Pre-Training (CLIP) model. With the carefully designed network structures and loss functions, our framework can perform high-quality hair editing in a disentangled manner. Extensive experiments demonstrate the superiority of our approach in terms of manipulation accuracy, visual realism of editing results, and irrelevant attribute preservation.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022 |
| Publisher | IEEE |
| Pages | 18051-18060 |
| Number of pages | 10 |
| ISBN (Electronic) | 978-1-6654-6946-3 |
| ISBN (Print) | 978-1-6654-6947-0 |
| DOIs | |
| Publication status | Published - 2022 |
| Event | 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022) - Hybrid, New Orleans, United States Duration: 19 Jun 2022 → 24 Jun 2022 https://cvpr2022.thecvf.com/ |
Publication series
| Name | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
|---|---|
| ISSN (Print) | 1063-6919 |
| ISSN (Electronic) | 2575-7075 |
Conference
| Conference | 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022) |
|---|---|
| Place | United States |
| City | New Orleans |
| Period | 19/06/22 → 24/06/22 |
| Internet address |
Bibliographical note
Research Unit(s) information for this publication is provided by the author(s) concerned.Research Keywords
- Image and video synthesis and generation
- Vision + language
- Vision applications and systems