Embodied AI: Ushering in the Next Era of Intelligent Systems
The rapid advancements in artificial intelligence, particularly with the rise of large language models (LLMs), visual-language models (VLMs), and multimodal language models (MLMs), have marked significant milestones in the field. These innovations have primarily focused on cognitive capabilities within digital realms.
However, a new frontier is emerging: Embodied AI (EAI). This interdisciplinary domain integrates principles from AI, robotics, and cognitive science to empower robots with human-like perception, planning, decision-making, and behavioral execution. EAI emphasizes the perception-action loop, where physical entities perceive their environment, formulate plans and decisions based on objectives, and leverage manipulative actions to complete tasks.
As Dr. Fei-Fei Li eloquently posited regarding spatial intelligence, vision has been a powerful driver of animal intelligence. Similarly, interaction with the environment accelerates the evolution of intelligence for animals and upright humans. We believe EAI will similarly catalyze substantial progress in AI, particularly in the realm of physical intelligence. If the advancements in LLMs, VLMs, and MLMs represent the initial phase of AI development, then EAI is poised to define its next major chapter.
This forum brings together distinguished industry leaders, scientists, engineers and even investors to delve into the pressing challenges and opportunities within Embodied AI. Discussions will encompass crucial topics such as the integration of MLMs in embodied systems, advanced robot (and humanoid) manipulation learning, innovative data capture and synthesis techniques, and achieving robust model generalization in real-world environments.
We look forward to a stimulating discussion on these transformative topics.
Time: 4:15 - 6:00 PM PT, August 7, 2025, Thursday
Moderator:
Yu Huang
CEO, Roboraction.AI, USA
Bio: Dr. Yu Huang is CEO of Roboraction.AI. Formerly he was Chief Autonomous Driving Scientist at SAIC Zone Tech, Adjust professor of Shanghai University, VP of Autonomous Driving Research at Black Sesame Technology, Chairman of Singulato USA. Yu also worked in Baidu USA, Intel (San Jose), Samsung Electronics USA and Futurewei Technology Inc., USA. Dr. Huang have more than 40 academic papers published in international conferences and journals, 18 patents issued in US and Europe and one book “System Development of Autonomous Driving” (in Chinese). He got BS degree, MS degree and Doctor degree at Xi'an Jiaotong University, Xidian University, Beijing Jiaotong University respectively. Dr. Huang also was an AvH (Alexander von Humboldt) research scholar (Germany) and postdoctoral associate of Beckman Inst., UIUC.
Speakers:
Jingyi Jin
Principal Engineer, Nvidia, USA
Bio: Dr. Jingyi Jin is a Principal Engineer in NVIDIA’s GenAI group, contributing to the development of the Cosmos World Foundation Model. Previously, she played a key role in advancing Edify 3D. Before joining NVIDIA, Jingyi was a Principal Engineer at Intel, where she led cloud software initiatives, incubated privacy-preserving machine learning technologies, and accelerated deep learning across a range of hardware architectures, including GPUs, FPGAs, Nervana, and Movidius. Earlier in her career, she worked extensively on seismic data processing and image processing. Jingyi holds a Master’s and Ph.D. in Computer Graphics from the University of Illinois at Urbana-Champaign (UIUC).
Rui Xu
Chief Operating Officer, K-Scale Labs
Rui was an early member of Amazon Lab 126 and served as the product manager for Kindle Fire TV. He later became Director of International Product at Xiaomi and General Manager of both the Hardware Platform and Desk Lamp AI Product at ByteDance. His teams have won back-to-back Best of CES awards in 2018 and 2019, along with over 100 CES media awards, 4 Red Dot Awards, and 3 Good Design Awards. Two of his AI hardware products were featured at Google I/O, one of which was presented on the main stage. Rui is currently building a general-purpose robotics startup in Silicon Valley.
Haoqi Fan
Research Scientist at Seed Edge, ByteDance, USA
Bio: Haoqi Fan is a Research Scientist at Seed Edge of ByteDance, where he leads efforts to build world foundational models. He spent seven years at Facebook AI Research (FAIR), focusing on self-supervised learning and backbone design for image and video understanding. His works won the ActivityNet Challenge at ICCV 2019 and were nominated for Best Paper at CVPR 2020. He has also co-organized several tutorials at CVPR, ICCV, and ECCV.
Jinwei Gu
Principal Research Scientist, Nvidia, USA
Dr. Jinwei Gu is currently a principal research scientist at NVIDIA, working on deep generative models, vision foundation models, world models, and the general fields of computer vision, computer graphics, and machine learning. He is also an adjunct associate professor at the Chinese University of Hong Kong. He received my Ph.D. from Columbia University in 2010, and B.S. and M.S. from Tsinghua University in 2002 and 2005.
Alex Ren
Founding Partner, Fellows Fund (early-stage AI Venture Capital), USA
Alex Ren is the Founding Partner of Fellows Fund, an AI-native venture capital firm based in the San Francisco Bay Area. With over 20 years of experience across R&D, sales, marketing, HR, and startup investment, he leads a portfolio exceeding 50 early-stage startups. Notable investments include OpusClip, an AI-powered video editing platform, as well as robotics innovators like Dyna Robotics and K-Scale Labs, reflecting his commitment to advancing AI and automation technologies.