SAILOR: Synergizing Radiance and Occupancy Fields for Live Human Performance Capture

Zheng DONG, Ke XU, Yaoan GAO, Qilin SUN, Hujun BAO, Weiwei XU*, Rynson W.H. LAU

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

3 Citations (Scopus)

Abstract

Immersive user experiences in live VR/AR performances require a fast and accurate free-view rendering of the performers. Existing methods are mainly based on Pixel-aligned Implicit Functions (PIFu) or Neural Radiance Fields (NeRF). However, while PIFu-based methods usually fail to produce photorealistic view-dependent textures, NeRF-based methods typically lack local geometry accuracy and are computationally heavy (e.g., dense sampling of 3D points, additional fine-tuning, or pose estimation). In this work, we propose a novel generalizable method, named SAILOR, to create high-quality human free-view videos from very sparse RGBD live streams. To produce view-dependent textures while preserving locally accurate geometry, we integrate PIFu and NeRF such that they work synergistically by conditioning the PIFu on depth and then rendering view-dependent textures through NeRF. Specifically, we propose a novel network, named SRONet, for this hybrid representation. SRONet can handle unseen performers without fine-tuning. Besides, a neural blending-based ray interpolation approach, a tree-based voxel-denoising scheme, and a parallel computing pipeline are incorporated to reconstruct and render live free-view videos at 10 fps on average. To evaluate the rendering performance, we construct a real-captured RGBD benchmark from 40 performers. Experimental results show that SAILOR outperforms existing human reconstruction and performance capture methods. © 2023 Copyright held by the owner/author(s). Publication rights licensed to ACM.
Original languageEnglish
Article number205
JournalACM Transactions on Graphics
Volume42
Issue number6
Online published5 Dec 2023
DOIs
Publication statusPublished - Dec 2023

Bibliographical note

Full text of this publication does not contain sufficient affiliation information. With consent from the author(s) concerned, the Research Unit(s) information for this record is based on the existing academic department affiliation of the author(s).

Funding

We thank all the anonymous reviewers for their professional and constructive comments. This work was partially supported by a GRF grant from RGC of Hong Kong (Ref.: 11205620). Weiwei Xu is supported by NSFC (No. 61732016). Qilin Sun would like to acknowledge the support from NSFC (No. 62302423). Besides, this work was supported by Ant Group, and this paper is supported by the Information Technology Center and State Key Lab of CAD&CG, Zhejiang University.

Research Keywords

  • high-quality human free-view videos
  • human performance capture
  • hybrid representation
  • occupancy and radiance fields

Fingerprint

Dive into the research topics of 'SAILOR: Synergizing Radiance and Occupancy Fields for Live Human Performance Capture'. Together they form a unique fingerprint.
  • GRF: Learning to Predict Scene Contexts

    LAU, R. W. H. (Principal Investigator / Project Coordinator), FU, H. (Co-Investigator) & FU, C. W. (Co-Investigator)

    1/01/2112/06/25

    Project: Research

Cite this