This workshop focuses on cooperative intelligence within multi-agent embodied intelligent systems. Artificial intelligence has propelled the development of embodied AI, particularly in autonomous vehicles, robotics, and drones. However, achieving full autonomy in complex and dynamic environments remains a formidable challenge for individual agents. Cooperative intelligence offers a transformative approach that allows agents to collaborate and interact with the infrastructure to handle a wide range of tasks more efficiently. In autonomous driving, the availability of datasets and breakthrough algorithms has spurred research interest in cooperative autonomous driving. Vehicle-to-Everything (V2X) interactions, including Vehicle-to-Vehicle (V2V) and Vehicle-to-Infrastructure (V2I), empower autonomous vehicles to extend perception, increase safety, and overcome the limitations of single-vehicle autonomy, laying the groundwork for large-scale adoption. In robotics, the evolution of multi-agent systems is revolutionizing the exploration of unknown environments. These advances allow robots to efficiently assist humans in challenging, open-world tasks. In drones, aerial robot swarms collaborate to perform complex tasks such as drone shows, 3D printing, and navigating cluttered environments. Furthermore, ground-air collaboration between drones and mobile robots shows immense potential in areas such as large-scale mapping and joint search and rescue. Despite progress, challenges in coordinating multi-agent systems remain underexplored. Key hurdles include deciding what information to transmit, how to transmit, and how to fuse data across various levels like perception, prediction, and planning. Moreover, obtaining high-quality real-world datasets is difficult. Recent advances in foundational and generative models offer promising ways to overcome these obstacles. This workshop will explore opportunities, challenges, and future directions for multi-agent embodied intelligent systems in the Agentic-AI era.

We invite submissions including but not limited to the following topics:
①Foundation Models and Architectures
- ›Large Language Model-assisted Cooperative System
- ›Foundation Models for Cooperative System
- ›Reasoning and Memory in Agentic System
- ›VLA for Robotics and Autonomous Driving (AD)
②Multi-Agent Systems & Collaboration
- ›Vehicle-to-Everything (V2X): V2V, V2I, V2P, V2D
- ›Multi-agent Robotic System and Swarm Robots
- ›Swarm of Drones and Aerial Robots
- ›Cooperative Motion Prediction and Decision-Making
- ›Communication-Efficient Cooperative Perception
- ›End-to-End Cooperative Policy Learning
③Simulation and Evaluation
- ›Simulation Platform for Cooperative System
- ›Datasets and Benchmarks for Cooperative Learning
- ›Simulation and Benchmarks for Agentic Systems
- ›Sim-to-Real Transfer
④Human-Agent Interaction
- ›Explainability and Interpretability for VLA
- ›Natural Language Interaction for Embodied Agents
- ›Human-Agent Collaboration
- ›Safety, Fairness, and Ethical Alignment
Important Dates
Archival Track
Non-Archival Track
Submission Guidance
- › Archival Track Submission Portal: OpenReview
- › Non-Archival Track Submission Portal: OpenReview
- ›Submission format: Submissions must follow the CVPR 2026 template (here) and will be peer-reviewed in a double-blind manner. Accepted papers will be presented in the form of posters, with several papers being selected for spotlight sessions.Archival Track:Maximum of 8 pages (excluding references). Accepted papers will be included in the official CVPR 2026 Workshop Proceedings published by IEEE, receive a DOI, and be indexed in digital libraries. These are considered formal publications.
Deadline: March 10, 2026, 11:59 PM Anywhere on Earth (AoE).Non-Archival Track:Maximum of 4 pages (excluding references). This track is for work-in-progress or recently published research. These papers will not be included in the proceedings and will not receive a DOI, allowing you to submit the work to other venues in the future.
Deadline: April 15, 2026, 11:59 PM Anywhere on Earth (AoE).
| Time | Session | Speaker / Host | Topic / Notes |
|---|---|---|---|
| 08:30 - 08:40 | Opening Remarks | Organizing Committee | Welcome & Workshop Overview |
| 08:40 - 09:00 | Opening Keynote | Xiangbo Gao | - |
| 09:00 - 09:30 | Keynote 1 | Xiaopeng Li | Agentic AI for Smart Transportation |
| 09:30 - 10:00 | Keynote 2 | Manabu Tsukada | Cooperative Intelligence for Autonomous Driving: From V2X Communication to Human-Agent Interaction |
| 10:00 - 10:30 | Poster Session I & Coffee Break | - | - |
| 10:30 - 11:00 | Keynote 3 | Bolei Zhou | Scalable Physical AI for Sidewalk Autonomy |
| 11:00 - 11:30 | Keynote 4 | Marco Pavone | Physical AI for End-to-End Vehicle Autonomy |
| 11:30 - 12:00 | Poster Session II & Coffee Break | - | - |
| 12:00 - 14:00 | Lunch Break | - | - |
| 14:00 - 14:30 | Keynote 5 | Yanjia Huang | Web based simulation teleoperation for general manipulation |
| 14:30 - 15:00 | Keynote 6 | Angela Dai | TBD |
| 15:00 - 15:30 | Keynote 7 | Bernadette Bucher | Bridging the Interaction Gap: Grounding Low-Level Control in Large-Scale Procedural Worlds |
| 15:30 - 15:45 | Oral Presentation 1 | - | Multi-Agent Video Prediction: Self-Correcting Conditional Frames for Dynamic Scene Forecasting |
| 15:45 - 16:00 | Oral Presentation 2 | - | Mitigating Multi-Module Errors for Reliable Navigation in Dynamic Environments via Online Trajectory Refinement |
| 16:00 - 16:15 | Oral Presentation 3 | - | Bridging the Pretrain-to-Real Gap: Alignment Challenges in Deploying Generalist VLA Models for Additive Manufacturing |
| 16:15 - 16:30 | Oral Presentation 4 | - | CoSA-3D: Vision-Only Automatic 3D Annotation for Cooperative Perception |
| 16:30 - 17:00 | Poster Session IV & Coffee Break | - | - |
| 17:00 - 17:30 | Keynote 8 | Jiachen Li | TBD |
| 17:30 - 17:45 | Paper Awards Ceremony | - | Best Paper, Runner-Up, Best Demo Paper |
| 17:45 - 18:00 | Closing Remarks & Group Photo | Organizing Committee | - |
Contact: If you have any questions, please contact us at: meis-cvpr-2026@googlegroups.com or xiangbog@tamu.edu .


























