
Date: 20th of October 2025 - 2-6 pm
Location: Room 310, Session 28
Submit your paper: here
Submission deadline: 25/09/2025
Call for Contributions
We invite submissions of extended abstracts, position papers, and technical contributions (2-4 page manuscripts in the IEEE double-column format) that address the integration of foundation models with human-in-the-loop robot learning. Please submit your paper by the above link. Accepted papers will be invited to submit extended versions to a partnered special issue of the Robot Learning journal related to the workshop topic.
Awards
To incentivize high-quality participation, we will present:
- Two Best Workshop Paper Awards (150 USD each), for outstanding contributions that advance human-in-the-loop learning with foundation models.
- One special best paper award (200 USD), provided by the Robot Learning journal, will be selected from authors who want to submit an extended version to our partnered special issue.
- One Travel Grant (200 USD), to support young researchers who have trouble joining the workshop in person. The application link is included in the above paper submission form
Online Questionnaire
This questionnaire will collect topics and questions, which will be discussed in the panel discussion session. Please fill it out here.
Outline and Objectives
This interdisciplinary workshop focuses on the latest advancements in human-in-the-loop robot learning, which integrate human multi-modal input (e.g. natural language, gestures, and haptic interaction) and online feedback (e.g., rewards, corrections, and preferences) to improve robot manipulation performance, adapt to new situations and align better with human intentions. Recent breakthroughs in foundation models, such as Large Language Models (LLMs), Vision-Language Models (VLMs), and Vision-Language-Action Models (VLAs), provide unprecedented perception and reasoning capabilities. However, their effective integration into robotics remains an emerging and underexplored challenge, especially for manipulation planning and control. This workshop will explore how foundation models and human-in-the-loop learning can be synergistically integrated to effectively enhance the robot learning process through active and intuitive human participation. We will delve into three critical themes:
- How can foundation models be leveraged for adaptive and generalizable learning for robot manipulation planning and control in com plex dynamic environments where robots continuously interact with the physical world?
- What are the best practices for integrating real-time human feedback to refine learning processes and improve alignment with human intentions?
- How to design adaptive learning frameworks to ensure safety and robustness in human-robot interactions and collaborations?
Topics of interest
Topics of interest include but are not limited to:
- Human-AI collaboration for robot learning
- Human-AI hybrid intelligence
- Foundation models for robot manipulation
- Transfer learning and fine-tuning of foundation models for robotic manipulation
- Knowledge representation and reasoning in robots
- Human feedback in robot learning
- Human-in-the-loop reinforcement learning
- Learning from demonstrations and corrections
- Interactive robot manipulation learning
- Multi-task robot learning
- Architectures and frameworks for human-in-the-loop learning
- Cognitive models for robot learning
- Adaptive human-robot interaction
- Safety and robustness in human-robot collaboration
Program
Time | Description |
---|---|
14.00 – 14.15 | Welcome and Introduction by the organizers |
14.15 – 14.40 | Talk 1: Robot Generalization with the Power of Large Vision Models by Huazhe Xu |
14.40 – 15.05 | Talk 2: Intelligent Physical Agents: High-Performance Human-in-the-loop Learning for Generalist Robots by Jianlan Luo |
15.05 – 15.30 | Talk 3: Collective Shaping of Multi-Particle Systems by David Navarro-Alarcon |
15.30 – 16.00 | Poster Session, Hands-on demonstrations, and Coffee Break |
16.00 – 16.25 | Talk 4: Human-Robot-Avatar Interaction and Collaboration in Dynamic Environments using Foundational Models by Alberto Sanfeliu |
16.25 – 16.50 | Talk 5: Understanding Humans, Empowering Robots: Online Assessment of Human Psycho-Physical State and Expert Feedback for Adaptive Robot Behaviour by Marta Lagomarsino |
16:50 – 17.15 | Talk 6: Robot Assembly Task Learning from Human Guide by Dongheui Lee |
17.15 – 17.45 | Panel Discussion and Award Ceremony |
17.45 – 18.30 |
Networking and Happy Hour |
Invited Speakers

Dr. Huazhe Xu
Assistant Professor, Institute for Interdisciplinary Information Sciences (IIIS), Tsinghua University, China; Co-founder of GalaxeaAI
Talk Title: Robot Generalization with the Power of Large Vision Models

Dr. Jianlan Luo
Chief Scientist, AGIBOT, Shanghai, China
Talk Title: Intelligent Physical Agents: High-Performance Human-in-the-loop Learning for Generalist Robots
Abstract
Robot learning has advanced significantly in recent years, positioning it as an effective tool for achieving scalable, flexible robotic autonomy. However, the large-scale real-world adoption of such learning-based robotic systems remains challenging, for which they must fulfill stringent real-world performance criteria to be viable. In this talk, I will describe algorithms and principles for building high-performance robotic learning systems. I’ll start by examining a range of high-performance "robot specialist" systems. These systems are tailored to address key deployment factors such as reliability, robustness, and cycle time, which has ultimately paved the way for their industrial adoption. I will then proceed to describe mechanisms to build “robot generalist” foundation models by bootstrapping the aforementioned robot specialists. To conclude, I'll further discuss the connections between these two types of systems and methods for enabling these systems to execute complex, long-horizon tasks suitable for open-world deployment.

Prof. David Navarro-Alarcon
Associate Professor, Department of Mechanical Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China
Talk Title: Collective Shaping of Multi-Particle Systems
Abstract
This talk presents our group's results in the development of methods to autonomously manipulate multi-particle aggregates with robotic systems. We demonstrate how robots can effectively transport and shape multiple dispersed particles through coordinated pushing actions. Our solution combines two key innovations: intelligent task planning powered by Vision Language Models (VLMs), and precise trajectory execution using adaptive shape representation. The VLMs enable the system to understand tool affordances and execute non-prehensile manipulation primitives. For tracking and controlling particle aggregates, we introduce an efficient parametrization method using truncated Fourier series to represent the evolving contour of particle groups. This mathematical framework allows us to compute optimal trajectories while maintaining group cohesion during manipulation. Through live demonstrations and experimental results, we'll show how our system successfully performs complex manipulation tasks while adapting to changing particle distributions. This research advances the field of robotic manipulation and opens new possibilities for applications in materials handling, assembly operations, and even in food processing scenarios.

Prof. Alberto Sanfeliu
Full Professor of Computational Sciences and Artificial Intelligence, Universitat Politècnica de Catalunya, UPC, Barcelona, Spain.
Talk Title: Human-Robot-Avatar Interaction and Collaboration in Dynamic Environments using Foundational Models

Dr. Marta Lagomarsino
Post-Doc, Human-Robot Interfaces and Interaction (HRII) Lab, Italian Institute of Technology (IIT), Genova, Italy
Talk Title: Understanding Humans, Empowering Robots: Online Assessment of Human Psycho-Physical State and Expert Feedback for Adaptive Robot Behaviour

Prof. Dongheui Lee
Full Professor, Faculty of Electrical Engineering and Information Technology, Technische Universität Wien, Austria
Talk Title: Robot Assembly Task Learning from Human Guide
Organizers
Jianzhuang Zhao, Post-Doc, Human-Robot Interfaces and Interaction (HRII) Lab, Italian Institute of Technology (IIT), Genova, Italy
e-mail: jianzhuang.zhao@iit.it
website: https://www.iit.it/it/people-details/-/people/jianzhuang-zhao
Xing Liu, Professor, Research Center for Intelligent Robotics, School of Astronautics, Northwestern Polytechnical University, Xi’an, P.R.China
e-mail: xingliu@nwpu.edu.cn
website: https://www.researchgate.net/profile/Xing-Liu-33
Marta Lagomarsino, Post-Doc, Human-Robot Interfaces and Interaction (HRII) Lab, Italian Institute of Technology (IIT), Genova, Italy
e-mail: marta.lagomarsino@iit.it
website: https://www.iit.it/it/people-details/-/people/marta-lagomarsino
Francesco Tassi, Post-Doc, Human-Robot Interfaces and Interaction (HRII) Lab, Italian Institute of Technology (IIT), Genova, Italy
e-mail: francesco.tassi@iit.it
website: https://www.iit.it/it/people-details/-/people/francesco-tassi
Shufei Li, Post-Doc, Department of Systems Engineering, City University of Hong Kong, Hong Kong SAR, P.R. China
e-mail: shufei.li@outlook.com
website: https://scholar.google.com/citations?user=CpCQmkwAAAAJ&hl=en&oi=ao
Gustavo Jose Giardini Lahr, Assistant Professor, Hospital Israelita Albert Einstein, San Paulo, Brazil
e-mail: gustavo.lahr@einstein.br
website: https://www.einstein.br/pesquisa/instituto-do-cerebro
Chenguang Yang, Professor, Fellow, IEEE, School of Electrical Engineering, Electronics and Computer Science, University of Liverpool, Liverpool, UK
e-mail: Chenguang.Yang@liverpool.ac.uk
website: https://www.liverpool.ac.uk/people/charlie-yang
Xuguang Lan, Professor, Institute of Artificial Intelligence and Robotics, College of Artificial Intelligence, Xi’an Jiaotong University, Xi’an, P.R. China
e-mail: xglan@mail.xjtu.edu.cn
website: https://gr.xjtu.edu.cn/en/web/zeuslan
Panfeng Huang, Professor, Research Center for Intelligent Robotics, School of Astronautics, Northwestern Polytechnical University, Xi’an, P.R.China
e-mail: pfhuang@nwpu.edu.cn
website: https://www.researchgate.net/profile/Panfeng-Huang
Arash Ajoudani, Senior Researcher Tenured, Human-Robot Interfaces and Interaction (HRII) Lab, Italian Institute of Technology (IIT), Genova, Italy
e-mail: arash.ajoudani@iit.it
website: https://www.iit.it/people-details/-/people/arash-ajoudani
This workshop engages three PhD students, Elena Merlo (Italian Institute of Technology, Italy), Zihao Liu (Northwestern Polytechnical University, China), and Zifan Wang (The Hong Kong University of Science and Technology, China) who will collaboratively design promotional campaigns and coordinate hands-on demonstrations during local arrangements.
Sponsors
This workshop is supported in part by the Horizon Europe project TORNADO (grant agreement No.101189557) and in part by the National Key R&D Program of China under Grant 2022ZD0117903, and the National Natural Science Foundation of China under Grant 92370123, 62273280, and the Guangdong Major Project of Basic and Applied Basic Research under Grant 2023B0303000016.
For our workshop’s best paper award and happy hour, we are privileged and grateful to have the financial support of several confirmed sponsors:
TORNADO project funded by European Union
IEEE-RAS Technical Committee on Collaborative Automation for Flexible Manufacturing
IEEE-RAS Technical Committee on Computer & Robot Vision
IEEE-RAS Technical Committee on Robot Learning
IEEE-RAS Deformable Object Manipulation (DOM) Working Group