There has been a tremendous advancement in machine learning methods over the past ten years. These advancements are reflected in robotics research primarily as ad-hoc solutions to different problems, such as learning to grasp simple objects [1], [2], learning movement primitives from demonstrations [3], [4] and so on. Although such solutions push the frontiers of robotics research, there is still a large gap between current methods and generally capable agents that learn skills and representations of progressive complexity to solve increasingly complex problems in an environment without forgetting previously learned skills, which is the ultimate aim of artificial intelligence research. Several studies [5], [6], [7], [8] focus on learning representations in an open-ended, continual manner in order to develop general skills without manually designing reward functions or curricula. The main advantage of open-ended (or lifelong, or continual) learning is that the learning loop never stops by design; the agent autonomously generates new challenges, tasks, environments, and goals to explore and learn new skills. This opens new possibilities for learning emergent skills not foreseen in hindsight. However, these methods do not precisely fit robot learning as we cannot arbitrarily evolve the embodiment of the agent and the real world.
This workshop will focus on how to create open-ended or lifelong learning systems [9], [10] that will allow for a robot to autonomously explore its environment and learn ever-growing representations for perception and actuation. This is an underexplored area in robotics and artificial intelligence research that can contribute to the development of generally capable agents. We will discuss the necessary elements (e.g., methods, environments, datasets, embodiments) for a lifelong learning setting. We will encourage participants to take part in discussions with the following points:
We believe that approaching robot learning from a lifelong learning perspective might be a feasible approach for truly intelligent robots. Stimulating discussions with the robotics community will be very valuable to understand if such an approach is indeed worth exploring or not.
8:50-9:00 (JST, Oct 23) — 1:50-2:00 (CEST, Oct 23) — 19:50-20:00 (EDT, Oct 22) Opening remarks |
9:00-9:45 (JST, Oct 23) — 2:00-2:45 (CEST, Oct 23) — 20:00-20:45 (EDT, Oct 22) Speaker: Jeff CluneTitle: Improving Robot and Deep Reinforcement Learning via Quality Diversity, Open-Ended, and AI-Generating Algorithms Abstract: Quality Diversity (QD) algorithms are those that seek to produce a diverse set of high-performing solutions to problems. I will describe them and a number of their positive attributes. I will summarize how they enable robots, after being damaged, to adapt in 1-2 minutes in order to continue performing their mission. I will next describe our QD-based Go-Explore algorithm, which dramatically improves the ability of deep reinforcement learning algorithms to solve previously unsolvable problems wherein reward signals are sparse, meaning that intelligent exploration is required. Go-Explore solved all unsolved Atari games, including Montezuma's Revenge and Pitfall, considered by many to be a grand challenges of AI research. I will next motivate research into open-ended algorithms, which seek to innovate endlessly, and introduce our POET algorithm, which generates its own training challenges while learning to solve them, automatically creating a curricula for robots to learn an expanding set of diverse skills. Finally, I'll argue that an alternate paradigm —AI-generating algorithms (AI-GAs)— may be the fastest path to accomplishing our field's grandest ambition of creating general AI, and describe how QD and Open-Ended algorithms will be essential ingredients of AI-GAs. Bio: Jeff Clune is an Associate Professor of computer science at the University of British Columbia and a faculty member at the Vector Institute. Jeff focuses on deep learning, including deep reinforcement learning. Previously he was a research manager at OpenAI, a Senior Research Manager and founding member of Uber AI Labs (formed after Uber acquired a startup he helped lead), the Harris Associate Professor in Computer Science at the University of Wyoming, and a Research Scientist at Cornell University. He received degrees from Michigan State University (PhD, master's) and the University of Michigan (bachelor's). More on Jeff's research can be found at JeffClune.com or on Twitter (@jeffclune). |
9:45-10:30 (JST, Oct 23) — 2:45-3:30 (CEST, Oct 23) — 20:45-21:30 (EDT, Oct 22) Speaker: Stefanie Tellex |
10:30-11:15 (JST, Oct 23) — 3:30-4:15 (CEST, Oct 23) — 21:30-22:15 (EDT, Oct 22) Speaker: Jorge MendezTitle: Lifelong Robot Learning via Functional Compositionality Abstract: In order to be deployed long-term in the real world, machine learning systems must be able to handle the one thing that is constant: change. Traditional machine learning focuses on stationary tasks, and is therefore incapable of adapting to change. Instead, we need lifelong learners that can accumulate knowledge that enables them to rapidly adjust to new tasks. Ideally, pieces of this accumulated knowledge could be composed in different ways to adapt to the shifting environment. This capability would dramatically improve the performance of machine learning systems in dynamic environments: hate-speech detection models could adapt to social media trends, student feedback software could adjust to new cohorts, search-and-rescue robots could handle novel disasters. Bio: Jorge Mendez is a Postdoctoral Fellow at the Massachusetts Institute of Technology, in the Computer Science and Artificial Intelligence Laboratory (CSAIL). He obtained his Ph.D. from the University of Pennsylvania (UPenn), in the Department of Computer and Information Science and the General Robotics, Automation, Sensing, and Perception (GRASP) Lab. He received a Master's degree in Robotics from UPenn (2018) and a Bachelor's degree in Electronics Engineering from Universidad Simon Bolivar in Venezuela (2016). His primary interests lie in the creation of versatile artificially intelligent systems that learn to accumulate knowledge over their lifetimes. His research leverages a breadth of techniques from transfer learning, multitask learning, compositional learning, representation learning, and reinforcement learning to develop algorithms with applications in computer vision, robotics, and natural language. His work has been recognized with the Best Paper Award in the Lifelong Machine Learning Workshop (ICML, 2020), the third place prize of the Two Sigma Ph.D. Diversity Fellowship (2021), and the School of Engineering Postdoctoral Fellowship for Engineering Excellence from MIT (2022). |
11:15-12:00 (JST, Oct 23) — 4:15-5:00 (CEST, Oct 23) — 22:15-23:00 (EDT, Oct 22) Speaker: George KonidarisTitle: Reintegrating AI: Skills, Symbols, and the Sensorimotor Dilemma Abstract: I will address the question of how a robot should learn an abstract, task-specific representation of an environment. I will present a constructivist approach, where the computation the representation is required to support - here, planning using a given set of motor skills - is precisely defined, and then its properties are used to build the representation so that it is capable of doing so by construction. The result is a formal link between the skills available to a robot and the symbols it should use to plan with them. I will present an example of a robot autonomously learning a (sound and complete) abstract representation directly from sensorimotor data, and then using it to plan. I will also discuss ongoing work on making the resulting abstractions portable across tasks. Bio: George Konidaris is an Associate Professor of Computer Science at Brown and the Chief Roboticist of Realtime Robotics, a startup commercializing his work on hardware-accelerated motion planning. He holds a BScHons from the University of the Witwatersrand, an MSc from the University of Edinburgh, and a PhD from the University of Massachusetts Amherst. Prior to joining Brown, he held a faculty position at Duke and was a postdoctoral researcher at MIT. George is the recent recipient of an NSF CAREER award, young faculty awards from DARPA and the AFOSR, and the IJCAI-JAIR Best Paper Prize. |
12:00-13:00 (JST, Oct 23) — 5:00-6:00 (CEST, Oct 23) — 23:00-00:00 (EDT, Oct 22) Lunch Break |
13:00-14:00 (JST, Oct 23) — 6:00-7:00 (CEST, Oct 23) — 00:00-01:00 (EDT, Oct 23) Contributed talks |
14:00-14:45 (JST, Oct 23) — 7:00-7:45 (CEST, Oct 23) — 01:00-01:45 (EDT, Oct 23) Speaker: Tamim AsfourTitle: Continual Learning for Personal Humanoid Robotics Abstract: Personal humanoid robots that should coexist with humans to assist and collaborate with them should be able to continually learn from interaction with the world and their users to extend the representations needed to reason about the world. The talk will present progress towards 24/7 humanoid robots — as embodied AI systems — that learn from experience and verbalize such experience, learn object relations and acquire task representations for manipulation skills from small number of human demonstration videos. The talk will conclude with a discussion of current research directions. Bio: Tamim Asfour is full Professor of Humanoid Robotics at the Institute for Anthropomatics and Robotics at the Karlsruhe Institute of Technology (KIT), Germany. His research focuses on the engineering of high performance 24/7 humanoid robotics. In particular, he studies the mechano-informatics of humanoids as the synergetic integration of informatics, artificial intelligence, and mechatronics into complete humanoid robot systems, which are able perform versatile tasks real world. Tamim is the developer of the ARMAR humanoid robot family. He has been a visiting professor at Georgia Tech, at the Tokyo University of Agriculture and Technology, and at the National University of Singapore. He is Editor-in-Chief and Editor of the Robotics and Automation Letters (RA-L), the Founding Editor-in-Chief of the IEEE-RAS Humanoids Conference Editorial Board, the president of the Executive Board of the German Robotics Society (DGR), and the scientific spokesperson of the KIT Center "Information · Systems · Technologies" (KCIST), www.humanoids.kit.edu |
14:45-15:30 (JST, Oct 23) — 7:45-8:30 (CEST, Oct 23) — 01:45-02:30 (EDT, Oct 23) Speaker: Jun TaniTitle: Dynamic Goal-directed Planning and Goal Understanding of Robots by Extended Active Inference Abstract: This study shows that goal-directed action planning and generation in a teleological framework can be formulated by extending the active inference framework. The proposed model, which is built on a variational recurrent neural network model, is characterized by three essential features. These are that (1) goals can be specified for both static sensory states, e.g., for goal images to be reached and dynamic processes, e.g., for moving around an object, (2) the model cannot only generate goal-directed action plans, but can also understand goals through sensory observation, and (3) the model generates future action plans for given goals based on the best estimate of the current state, inferred from past sensory observations. The proposed model is evaluated by conducting experiments on a real humanoid robot performing object manipulation. I'll also briefly describe the current on-going study on dynamic replanning against interferences and incremental tutoring for recovery. Bio: Jun Tani received the D.Eng. degree from Sophia University, Tokyo in 1995. He started his research career with Sony Computer Science Lab. in 1993. He became a Team Leader of the Laboratory for Behavior and Dynamic Cognition, RIKEN Brain Science Institute, Saitama, Japan in 2001. He became a Full Professor with the Electrical Engineering Department, Korea Advanced Institute of Science and Technology, Daejeon, South Korea in 2012. He is currently a Full Professor with the Okinawa Institute of Science and Technology, Okinawa, Japan. He is also a visiting professor of The Technical University of Munich. His current research interests include cognitive neuroscience, developmental psychology, phenomenology, complex adaptive systems, and robotics. He is an author of "Exploring Robotic Minds: Actions, Symbols, and Consciousness as Self-Organizing Dynamic Phenomena." published from Oxford Univ. Press in 2016. |
15:30-16:15 (JST, Oct 23) — 8:30-9:15 (CEST, Oct 23) — 02:30-03:15 (EDT, Oct 23) Speaker: Minoru AsadaTitle: How Human Brain Develops in the Process of Self-face Recognition, Bodily Awareness and Self-awareness Abstract: In this talk, we will explain how human brain develops in the process of self-face recognition, bodily awareness and self-awareness based on fMRI studies. First, we have shown that similar to slow maturation of neural processing for face recognition, neural substrates of self-face recognition slowly developed in typically developing humans. Next, the developmental process of bodily awareness showed U-shape change from children to adults, more correctly, suppression of the left-sided IPL activity is found in adolescents. Finally, we examined the development and aging of brain deactivation using a unimanual motor task, and found that brain inhibition develops and declines through aging. Based on these studies, we discuss how to design life-long learning robots. Bio: Minoru Asada received his B.E. (1977), M.E. (1979), and Ph.D. (1982) in Control Engineering from Osaka University, Osaka, Japan. He became a full Professor of Mechanical Engineering for Computer-Controlled Machinery at Osaka University in 1995. Since April 1997, he has been a Professor at the Department of Adaptive Machine Systems at Osaka University (Suita, Japan). Since 2013, he has been the director of the division of cognitive neuroscience robotics, the Institute for Academic Initiatives (IAI), Osaka University. He was the Research Director of the JST (Japan Science and Technology Agency) ERATO (Exploratory Research for Advanced Technology) ASADA Synergistic Intelligence Project during 2005 and 2012. In 2012, the Japan Society for Promotion of Science (JSPS) named him to serve as the Research Leader for the Specially Promoted Research Project (Tokusui) on Constructive Developmental Science Based on Understanding the Process From Neuro-Dynamics to Social Interaction. |
16:15-17:00 (JST, Oct 23) — 9:15-10:00 (CEST, Oct 23) — 03:15-04:00 (EDT, Oct 23) Speaker: Yukie NagaiTitle: The Emergence of Social Cognition through Open-Ended Predictive Learning Abstract: Human children acquire social cognitive abilities through open-ended learning. They start reading the internal state of others (e.g., intention and emotion) and cooperating with others in the first few years of life despite no explicit learning of social skills. An open question is where their social abilities and social motivation come from. Bio: Yukie Nagai is a Project Professor at the International Research Center for Neurointelligence, the University of Tokyo. She received her Ph.D. in Engineering from Osaka University in 2004 and then worked at the National Institute of Information and Communications Technology, Bielefeld University, and Osaka University. Since 2019, she leads Cognitive Developmental Robotics Lab at the University of Tokyo. Her research interests include cognitive developmental robotics, computational neuroscience, and assistive technologies for developmental disorders. She has been investigating underlying neural mechanisms for social cognitive development by means of computational approaches. She was elected to "30 women in robotics you need to know about" in 2019 and "World's 50 Most Renowned Women in Robotics" in 2020. |
This workshop is supported by the TÜBİTAK (The Scientific and Technological Research Council of Turkey) ARDEB 1001 program (120E274), the Grant-in-Aid for Scientific Research (project no 22H03670), and the International Joint Research Promotion Program, Osaka University, under the project 'Developmentally and biologically realistic modeling of perspective invariant action understanding'.