Welcome to

Physical Vision Group (PVG).


Our group conducts research in computer vision and machine learning, focusing on reconstructing and understanding the physical world.

We study Physical Natural World Creation, building systems that can perceive, reconstruct, and interact with the physical world. Beyond classical tasks such as appearance, content and geometry generation, we investigate deeper physical properties, like occlusion, motion, gravity, interaction, mass and sound. Our broader goal is to build realistic digital twins of the natural world, with various physical properties. Our group is part of the Multimedia and Interactive Computing Lab (MICL)  within the  College of Computing and Data Science (CCDS)  at  NTU, Singapore.

📍Openings: more details are available at  Position.

  • Phd student:   we have multiple PhD openings for Spring 2026 and Fall 2026.
  • Postdoc position:   we have two postdoc positions available.
  • Research Assistant:   we have two research assistant positions available.

Chuanxia Zheng is awarded the MSCA Fellowship

The Marie Skłodowska-Curie Actions (MSCA) are among Europe's most competitive and prestigious research and innovation fellowships

13/02/2024

NRF Fellowship, National Research Foundation

PhysWorld: Integrating Physical Properties in Natural World Creation

National Research Foundation,  PI

The grant was awarded on 12/02/2025 (S$ 3,078,720.00), and started on 01/09/2025 and will end on 30/08/2030.

The aim of PhysWorld is to create a realistic natural world adhering to physical principles, rather than dealing with only the realistic pixels as in traditional image, video, and 3D synthesis. Physics-based natural world creation is challenging because it requires a holistic interpretation of scenes and objects within it, including but not limited to appearance, geometry, materials, occlusion, motion, gravity, interaction, mass and sound.

Sony Focused Research Award

Vista4D: Feed-Forward 4D Scene Reconstruction from Any Monocular Video

Sony Research Award Program,  Co-PI, with  Andrea Vedaldi

The grant was awarded on 30/03/2025 (), and started on 01/07/2025 and will end on 30/06/2026.

The aim of Vista4D is to obtain photorealistic 4D reconstructions of dynamic scenes from a single monocular video. To achieve this, Vista4D will develop a feed-forward neural network that, given as input a single monocular video, will output directly its 4D reconstruction

MSCA Fellowship

SYN3D: Synthesizing Photorealistic 3D Scene from Zero to One or Limited Views

MSCA Postdoctoral Fellowship (UKRI Horizon Europe),  PI, with  Andrea Vedaldi

The grant was awarded on 13/02/2024 (£ 206,085.62), and started on 01/06/2024 and will end on 01/06/2026.

The aim of SYN3D is to enable the effortless creation of novel view synthesis and 3D reconstruction from limited views (from 0 to 5) in a feed-forward manner. In general, such a few input views do not contain sufficient information for compressive 3D synthesis. However, our scientific hypothesis is that 3D synthesis from limited views is still possible by learning, from billion images, the capability of hallucinating unobserved parts of 3D objects/scenes, thus obtaining plausible if not faithful reconstructions.