About Me

Hello! I am a research scientist at NVIDIA, working on physical AI and autonomous vehicles with Marco Pavone at Stanford University. I obtained my PhD degree from New York University (NYU) in 2025, working on robot perception with Chen Feng. I am also fortunate to collaborate with Saining Xie on visual-spatial intelligence. During my PhD, I did three internships at NVIDIA, working on autonomous perception and neural simulation with Anima Anandkumar (Caltech), Sanja Fidler (UofT), Jose M. Alvarez, Zhiding Yu, Chaowei Xiao (JHU), Zan Gojcjc, and Yue Wang (USC). I also spent some time at Tsinghua IIIS with Hang Zhao and Shanghai Jiao Tong University (SJTU) with Siheng Chen.

My research has been cited nearly 4,000 times (as of Nov 2025), and I am honored to be a recipient of the NVIDIA Fellowship (2024-2025), NYU Dean's PhD Fellowship, and NYU Outstanding Dissertation Awards (Finalist).

I will join Tsinghua University as an Assistant Professor in the College of AI (led by Prof. Andrew Chi-Chih Yao) in 2026.

🎓 Recruiting: Looking for postdocs, PhD students, undergraduates, interns, and visiting scholars to join my lab. Welcome to reach out!
⏳ Deadline: International PhD applications for Fall 2026 are due by January 15, 2026. Urgently seeking qualified candidates!

Research Lab - Spatial Intelligence

Our mission is to scale AI across space, time, and embodiments to address real-world challenges. Towards this end, we are pushing the frontiers of spatial intelligence through the convergence of vision, learning, and robotics. Our research agenda centers on three thrusts:

🧠 Research Thrust 1 - Spatial Cognition: How can embodied agents perceive, represent, and reason about space and time like humans do?

Sensing and Perception, Spatial Representation, Spatial Reasoning, Spatial Memory, Spatial World Model, Cognitive Mapping, Mental Manipulation

🌏 Research Thrust 2 - Spatial Computing: How to create realistic digital twins from multimodal sensory streams without human supervision?

3D Reconstruction, Neural Radiance Fields, Gaussian Splatting, Physics-based Simulation, Generative Modeling, Real2Sim2Real, Edge-Computing

🦾 Research Thrust 3 - Spatial Robotics: How can we ground cognitive intelligence in real-world robots across different morphologies?

Autonomous Navigation, Humanoid Robotics, Multi-Robot System, Field Robotics, Bio-Inspired Robotics, Robotic Design and Its Automation

💡 We believe the above three thrusts are deeply synergistic: cognitive models trained in simulation are grounded in physical robots, whose real-world experiences in turn benefit both the cognitive models and digital worlds. Our long-term vision is to build a self-sustaining, self-evolving spatial AI ecosystem where machines autonomously perceive, reason about, and transform the physical world—enabling AI Designs AI, Robots Build Robots through the seamless convergence of embodied cognition, digital universe, and physical embodiments.

💪 Towards this long-term vision, we are currently pushing the following research directions:

  • Prototyping spatially-grounded foundation models
  • Exploring human-like efficient spatial representations
  • Building open-world embodied AI simulators
  • Grounding visual-spatial intelligence in humanoid robots
  • Constructing large-scale datasets and benchmarks to broaden the practical applications of spatial AI
  • Designing specialized physical embodiments to diversify problem domains of spatial AI

📣 To advance our mission, we are building an interdisciplinary team and welcome researchers with diverse expertise, including but not limited to: (1) AI Computing (e.g., MLLMs, UMMs, VFMs, world model, generative model), (2) 3D Vision (e.g., 3DGS, NeRF, SLAM), and (3) Robotics (e.g., sim2real, humanoid robotics, field robotics, robotic design). Please send me an email, and I will get back to you if there is a good fit!

Selected Publications

(* indicate equal contribution/advising)

For full publication list, please refer to my Google Scholar Google Scholar page.

Thinking in 360°: Humanoid Visual Search in the Wild
🔥 Thinking in 360°: Humanoid Visual Search in the Wild
Technical Report, ArXiv 2025
Heyang Yu*, Yinan Han*, Xiangyu Zhang, Baiqiao Yin, Bowen Chang, Xiangyu Han, Xinhao Liu, Jing Zhang, Marco Pavone, Chen Feng*, Saining Xie*, Yiming Li*
Wanderland: Geometrically Grounded Simulation for Open-World Embodied AI
🔥 Wanderland: Geometrically Grounded Simulation for Open-World Embodied AI
Technical Report, ArXiv 2025
Xinhao Liu, Jiaqi Li, Youming Deng, Ruxin Chen, Yingjia Zhang, Yifei Ma, Li Guo, Yiming Li, Jing Zhang, Chen Feng
Flex: Towards Efficient and Effective Multi-Camera Encoding for End-to-End Driving
🔥 Flex: Towards Efficient and Effective Multi-Camera Encoding for End-to-End Driving
Technical Report, ArXiv 2025
Jiawei Yang, Ziyu Chen, Yurong You, Yan Wang, Yiming Li, Yuxiao Chen, Boyi Li, Boris Ivanovic, Marco Pavone, Yue Wang
CLM: Removing the GPU Memory Barrier for 3D Gaussian Splatting
🔥 CLM: Removing the GPU Memory Barrier for 3D Gaussian Splatting
ASPLOS 2026
Hexu Zhao, Xiwen Min, Xiaoteng Liu, Moonjun Gong, Yiming Li, Ang Li, Saining Xie, Jinyang Li, Aurojit Panda
Adversarial Exploitation of Data Diversity Improves Visual Localization
Adversarial Exploitation of Data Diversity Improves Visual Localization
ICCV 2025
Sihang Li*, Siqi Tan*, Bowen Chang, Jing Zhang, Chen Feng*, Yiming Li*
Unraveling the Effects of Synthetic Data on End-to-End Autonomous Driving
Unraveling the Effects of Synthetic Data on End-to-End Autonomous Driving
ICCV 2025
Junhao Ge, Zuhong Liu, Longteng Fan, Yifan Jiang, Jiaqi Su, Yiming Li, Zhejun Zhang, Siheng Chen
Extrapolated Urban View Synthesis Benchmark
Extrapolated Urban View Synthesis Benchmark
ICCV 2025
Xiangyu Han*, Zhen Jia*, Boyi Li, and Yan Wang, Boris Ivanovic, Yurong You, Lingjie Liu, Yue Wang, Marco Pavone, Chen Feng, Yiming Li
Memorize What Matters: Emergent Scene Decomposition from Multitraverse
Memorize What Matters: Emergent Scene Decomposition from Multitraverse
NeurIPS 2024
Yiming Li, Zehong Wang, Yue Wang, and Zhiding Yu, Zan Gojcic, Marco Pavone, Chen Feng, Jose M Alvarez
Spotlight Presentation
RadarOcc: Robust 3D Occupancy Prediction with 4D Imaging Radar
RadarOcc: Robust 3D Occupancy Prediction with 4D Imaging Radar
NeurIPS 2024
Fangqiang Ding*, Xiangyu Wen*, Yunzhou Zhu, and Yiming Li, Chris Xiaoxuan Lu
Multiagent Multitraversal Multimodal Self-Driving: Open MARS Dataset
Multiagent Multitraversal Multimodal Self-Driving: Open MARS Dataset
CVPR 2024
Yiming Li, Zhiheng Li, Nuo Chen, Moonjun Gong, Zonglin Lyu, Zehong Wang, Peili Jiang, Chen Feng
Predicting 3D Action Target from 2D Egocentric Vision for Human-Robot Interaction
Predicting 3D Action Target from 2D Egocentric Vision for Human-Robot Interaction
ICRA 2024
Irving Fang*, Yuzhong Chen*, Yifan Wang*, and Jianghan Zhang, Qiushi Zhang, Jiali Xu, Xibo He, Weibo Gao, Hao Su, Yiming Li, Chen Feng
SSCBench: A Large-Scale 3D Semantic Scene Completion Benchmark for Autonomous Driving
SSCBench: A Large-Scale 3D Semantic Scene Completion Benchmark for Autonomous Driving
IROS 2024
Yiming Li*, Sihang Li*, Xinhao Liu*, Moonjun Gong*, Kenan Li, Nuo Chen, Zijun Wang, Zhiheng Li, Tao Jiang, Fisher Yu, Yue Wang, Hang Zhao, Zhiding Yu, Chen Feng
Among Us: Adversarially Robust Collaborative Perception by Consensus
Among Us: Adversarially Robust Collaborative Perception by Consensus
ICCV 2023
Yiming Li*, Qi Fang*, Jiamu Bai, Siheng Chen, Felix Juefei-Xu, Chen Feng
VoxFormer: Sparse Voxel Transformer for Camera-based 3D Semantic Scene Completion
VoxFormer: Sparse Voxel Transformer for Camera-based 3D Semantic Scene Completion
CVPR 2023
Yiming Li, Zhiding Yu, Christopher Choy, and Chaowei Xiao, Jose M Alvarez, Sanja Fidler, Chen Feng, Anima Anandkumar
Highlight Presentation
DeepMapping2: Self-Supervised Large-Scale LiDAR Map Optimization
DeepMapping2: Self-Supervised Large-Scale LiDAR Map Optimization
CVPR 2023
Chao Chen*, Xinhao Liu*, Yiming Li, and Li Ding, Chen Feng
Metric-Free Exploration for Topological Mapping by Task and Motion Imitation in Feature Space
Metric-Free Exploration for Topological Mapping by Task and Motion Imitation in Feature Space
RSS 2023
Yuhang He*, Irving Fang*, Yiming Li, Rushi Bhavesh Shah, Chen Feng
Multi-Robot Scene Completion: Towards Task-Agnostic Collaborative Perception
Multi-Robot Scene Completion: Towards Task-Agnostic Collaborative Perception
CoRL 2022
Yiming Li*, Juexiao Zhang*, Dekun Ma, Yue Wang, Chen Feng
Egocentric Prediction of Action Target in 3D
Egocentric Prediction of Action Target in 3D
CVPR 2022
Yiming Li*, Ziang Cao*, Andrew Liang, and Benjamin Liang, Luoyao Chen, Hang Zhao, Chen Feng
Learning Distilled Collaboration Graph for Multi-Agent Perception
Learning Distilled Collaboration Graph for Multi-Agent Perception
NeurIPS 2021
Yiming Li, Shunli Ren, Pengxiang Wu, and Siheng Chen, Chen Feng, Wenjun Zhang
Fooling LiDAR Perception via Adversarial Trajectory Perturbation
Fooling LiDAR Perception via Adversarial Trajectory Perturbation
ICCV 2021
Yiming Li*, Congcong Wen*, Felix Juefei-Xu, and Chen Feng
Oral Presentation