Research Interests:
I am a researcher at Tencent Hunyuan, working on the HY World project. I received my Ph.D. degree from the Department of Computer Science, City University of Hong Kong (CityU), under the supervision of Prof. Rynson W.H. Lau and Prof. Gerhard Hancke. I received my B.Eng. degree from Xiamen University.
During my Ph.D. studies, I was fortunate to work with Tengfei Wang and Ziwei Liu from Shanghai AI Lab and MMLab@NTU. I was also lucky to closely collaborate with Nanxuan Zhao from Adobe Research.
I'm interested in generative AI, including world models, video generation, 3D reconstruction, and 3D asset generation.
|
Technical Report
|
HY-World 2.0 🔥Tencent HY World, April 2026 Next-gen multi-modal world model that reconstructs, generates, and simulates 3D worlds. |
|
Technical Report
|
HY-World 1.5Tencent HY World, December 2025 The first open-source world model with real-time latency, long-term memory, and interactive control. |
|
Technical Report
|
HY-World 1.0Tencent HY World, July 2025 Generate immersive, editable 3D scenes from text or a single image. |
(*equal contribution, ^intern, †corresponding author)
|
Technical Report
|
HY-World 2.0: A Multi-Modal World Model for Reconstructing, Generating, and Simulating 3D WorldsTencent HY World, April 2026 Next-gen multi-modal world model that reconstructs, generates, and simulates 3D worlds. |
|
Preprint
|
WorldPlay: Towards Long-Term Geometric Consistency for Real-Time Interactive World Modelingpreprint, December 2025 Real-time interactive world model with long-term geometric memory. |
|
Preprint
|
WorldMirror: Universal 3D World Reconstruction with Any-Prior Promptingpreprint, October 2025 Universal feed-forward 3D reconstruction — any input, any output. |
|
Preprint
|
MoCA: Mixture-of-Components Attention for Scalable Compositional 3D Generationpreprint, December 2025 Compositional 3D scene and object generation via sparse mixture-of-components attention. |
|
SIGGRAPH Asia 2025
|
StyleSculptor: Zero-Shot Style-Controllable 3D Asset Generation with Texture-Geometry Dual GuidanceProc. ACM SIGGRAPH Asia, December 2025 Training-free 3D asset generation with joint geometry and texture style guidance. |
|
SIGGRAPH Asia 2025
|
Voyager: Long-Range and World-Consistent Video Diffusion for Explorable 3D Scene GenerationACM Trans. on Graphics (Proc. ACM SIGGRAPH 2025, Journal), December 2025 Long-range 3D world exploration powered by RGB-D video diffusion. |
|
SIGGRAPH Asia 2025
|
Shape-for-Motion: Precise and Consistent Video Editing with 3D ProxyProc. ACM SIGGRAPH Asia, December 2025 Precise video object manipulation via 3D proxies and diffusion rendering. |
|
CVPR 2025
|
MAGE: Single Image to Material-Aware 3D via the Multi-View G-Buffer Estimation ModelCVPR, June 2025 Single-image, material-aware 3D reconstruction via G-buffer estimation. |
|
ICLR 2025
|
Phidias: A Generative Model for Creating 3D Content from Text, Image, and 3D Conditions with Reference-Augmented DiffusionICLR, April 2025 Retrieval-augmented 3D diffusion — generate from text, image, or existing 3D assets. |
|
SIGGRAPH 2024
|
ThemeStation: Generating Theme-Aware 3D Assets from Few ExemplarsProc. ACM SIGGRAPH, August 2024 Generate theme-consistent 3D asset galleries from just a few exemplars. |
|
AAAI 2024
|
Recasting Regional Lighting for Shadow RemovalAAAI, February 2024 Shadow removal that restores degraded textures conditioned on recovered illumination. |
|
SIGGRAPH 2023
|
Language-based Photo Color Adjustment for Graphic DesignsACM Trans. on Graphics (Proc. ACM SIGGRAPH 2023, Journal), August 2023 Language-driven photo recoloring for graphic designs. |