News

[03/2024] Two papers accepted to CVPR 2024 with URHand as Oral presentation.

[01/2024] One paper accepted to IJCV.

[09/2023] One paper accepted to NeurIPS 2023.

[09/2023] One paper accepted to IEEE TPAMI.

[07/2023] Two papers accepted to ICCV 2023.

[04/2023] Selected as a Finalist of 2023 Meta Research PhD Fellowship!

[02/2023] One paper accepted to CVPR 2023 as Highlight.

[01/2023] One paper accepted to ICLR 2023 as Spotlight.

[08/2022] One paper accepted to TOG (Proc. SIGGRAPH Asia 2022).

[07/2022] One paper accepted to ECCV 2022.

[08/2021] Join MMLab@NTU!

[07/2021] One paper accepted to ICCV 2021 for Oral presentation.

Publications

* denotes equal contributions


ORAL   CVPR - 2024
URHand: Universal Relightable Hands
Project Page Paper Video
URHand (a.k.a Your Hand). Our model is a high-fidelity Universal prior for Relightable Hands built upon light-stage data. It generalizes to novel viewpoints, poses, identities, and illuminations, which enables quick personalization from a phone scan. We propose the spatially varying linear lighting model for scalable relighting training.
CVPR - 2024
CityDreamer: Compositional Generative Model of Unbounded 3D Cities
Project Page Paper Video Code
CityDreamer learns to generate unbounded 3D cities from images (Google Earth imagery and OpenStreetMap).
NeurIPS - 2023
PrimDiffusion: Volumetric Primitives Diffusion for 3D Human Generation
Project Page Paper Video Code
PrimDiffusion performs the diffusion and denoising process on a set of primitives which compactly represent 3D humans. This generative modeling has explicit pose, view, and shape control, with the capability of modeling off-body topology in well-defined depth. It enables downstream tasks like 3D texture transfer and inpainting.
TPAMI - 2023
SceneDreamer: Unbounded 3D Scene Generation from 2D Image Collections
Zhaoxi Chen, Guangcong Wang, Ziwei Liu
Project Page Paper Video Code
SceneDreamer learns to generate unbounded 3D scenes from in-the-wild 2D image collections. Our method can synthesize diverse landscapes across different styles, with 3D consistency, well-defined depth, and free camera trajectory.
ICCV - 2023
SparseNeRF: Distilling Depth Ranking for Few-shot Novel View Synthesis
Project Page Paper Video Code
SparseNeRF is a simple yet effective few-shot NeRF by distilling robust local depth ranking priors from real-world inaccurate depth observations.
ICCV - 2023
SynBody: Synthetic Dataset with Layered Human Models for 3D Human Perception and Modeling
Project Page Paper Video
SynBody is a large-scale synthetic dataset with massive number of subjects and high-quality annotations. It supports various research topics, including 3D human perception, reconstruction and generation.
HIGHLIGHT   CVPR - 2023
F2-NeRF: Fast Neural Radiance Field Training with Free Camera Trajectories
Project Page Paper Code
F2-NeRF enables arbitrary input camera trajectories for novel view synthesis and only costs a few minutes for training.
SPOTLIGHT   ICLR - 2023
EVA3D: Compositional 3D Human Generation from 2D Image Collections
Project Page Paper Video Code
EVA3D is a high-quality unconditional 3D human generative model that only requires 2D image collections for training.
text2light
SIGGRAPH Asia (TOG) - 2022
Text2Light: Zero-Shot Text-Driven HDR Panorama Generation
Zhaoxi Chen, Guangcong Wang, Ziwei Liu
Project Page Paper Video Code
Text2Light can generate HDR panoramas in 4K+ resolution using free-form texts, without training on text-image pairs. The high-quality generated HDR panoramas can be directly applied to downstream tasks, e.g., light 3D scenes and immersive virtual reality.
relighting4d
ECCV - 2022
Relighting4D: Neural Relightable Human from Videos
Zhaoxi Chen, Ziwei Liu
Project Page Paper Code
Relighting4D takes only videos as input, decomposing them into geometry and reflectance in a self-supervised manner, which enables relighting of dynamic humans with free viewpoints by a physically based renderer.
adafocus
ORAL   ICCV - 2021
Adaptive Focus for Efficient Video Recognition
Paper Code
In this paper, we explore the spatial redundancy in video recognition with the aim to improve the computational efficiency. Extensive experiments on five benchmark datasets, i.e., ActivityNet, FCVID, Mini-Kinetics, Something-Something V1&V2, demonstrate that our method is significantly more efficient than the competitive baselines.
AISTATS - 2021
Understanding Robustness in Teacher-Student Setting: A New Perspective
Zhuolin Yang*, Zhaoxi Chen, Tiffany Cai, Xinyun Chen, Bo Li, Yuandong Tian*
Paper
In the case of low-rank input data, we show that student specialization still happens within the input subspace, but the teacher and student nodes could differ wildly out of the data subspace, which we conjecture leads to adversarial examples.

Experiences

Aug. 2021 - now

MMLab@NTU, Nanyang Technological University
Ph.D. student supervised by Prof. Ziwei Liu

Dec. 2020 - Jul. 2021

BNRist, Tsinghua University
Undergraduate thesis supervised by Prof. Gao Huang

May. 2020 - Oct. 2020

Secure Learning Lab, UIUC
Visiting Research Intern advised by Prof. Bo Li

Jun. 2020 - Sep. 2020

ModLab, University of Pennsylvania
Intern advised by Prof. Mark Yim

Services

Conference Reviewer

  • CVPR 2023, 2024
  • ICCV 2023
  • SIGGRAPH 2023, 2024
  • NeurIPS 2023
  • ICLR 2024
  • ICML 2024
  • ECCV 2024

Journal Reviewer

  • TOG
  • IJCV
  • TVCG
  • Neurocomputing