I'm interested in 3D vision and image processing. Much of my research is about inferring the physical world and camera (shape, motion, color, light, bokeh, etc) from images.
Sat3DGen: Comprehensive Street-Level 3D Scene Generation from Single Satellite Image Ming Qian, Zimin Xia, Changkun Liu, Shuailei Ma, Wen Wang, Zeran Ke, Bin Tan, Hang Zhang, Gui-Song Xia ICLR, 2026
paper
/
code & demo (coming soon)
Sat3DGen is a feed-forward satellite-to-3D framework that learns a structured, view-consistent NeRF-style scene from 2D satellite/street-view supervision, enabling mesh export and large-area mesh generation, surround-view video rendering, semantic-map-to-3D synthesis, and single-image DSM estimation.
Sat2Density focuses on the geometric nature of generating high-quality ground street videos conditioned on satellite images learning from collections of satellite-ground image pairs.
D-DFFNet considers the physical mechanism of defocus blur and successfully distinguishes homogeneous regions.
In addition, we propose a larger benchmark EBD that includes more DOF cases.
The results of detection on multiple public test sets look great.