Home Research Teaching Projects Publications

Human-Robot Collaboration Exploration and Development in Autonomous Systems
Intelligent Tutoring Systems and Education Software
Perception and Control
Robot Design

Inspired from developmental psychology and the way animals learn, an autonomous system that aims at life-long learning needs to adapt, decide what to learn and follow a specific developmental trajectory. Specific mechanisms of curiosity, information seeking and exploration need to be developed.

Recently, I have contributed with theoretical analysis of empirical measures of learning progress to generalize well-known methods of exploration in markov decision processes to situations with nonstationary noise. We showed that the use of empirical measures of learning rate reduce to the standard measure of learning rate commonly used in machine learning but are more robust to non-stationarities.

In many cases there are several different exploration methods that can be used, and in many cases theoretical results give similar sample complexity. For any particular case it is not clear which method to use. I suggested an online method that is able to select online the best exploration strategy.

I introduced a generic perspective on life-long learning called the strategic student problem. In many cases there are a large variety of tasks to be learned, specially in life-long learning situations, and it is necessary to decide which task can be learned, and how fast. I showed that the problem can only be solved efficiently in the particular situation of submodular costs. Due to this reason we introduced a bandit type algorithm that is able to address more complex costs but with more loose guarantees.

We started a collaboration with the Columbia University (USA), to better develop the relation between computational methods and results from neuroscience. We are currently developing biological plausible computational models of curiosity and information seeking in animals.

Relevant Publications:

Information-seeking, curiosity, and attention: computational and neural mechanisms, Jacqueline Gottlieb, Pierre-Yves Oudeyer, Manuel Lopes and Adrien Baranes. Trends in Cognitive Sciences , 2013. (pdf)

Learning Exploration Strategies in Model-Based Reinforcement Learning, Todd Hester, Manuel Lopes and Peter Stone. International Conference on Autonomous Agents and Multiagent Systems (AAMAS), Saint Paul, Minnesota, USA, 2013. (pdf)

Exploration in Model-based Reinforcement Learning by Empirically Estimating Learning Progress, Manuel Lopes, Tobias Lang, Marc Toussaint and Pierre-Yves Oudeyer. Neural Information Processing Systems (NIPS), Tahoe, USA, 2012. (pdf)

The Strategic Student Approach for Life-Long Exploration and Learning, Manuel Lopes and Pierre-Yves Oudeyer. IEEE - International Conference on Development and Learning (ICDL), 2012. (pdf)

A Developmental Roadmap for Learning by Imitation in Robots, Manuel Lopes and José Santos-Victor. IEEE Transactions in Systems Man and Cybernetic - Part B: Cybernetics, 37(2), 2007. (pdf)

Body Schema Acquisition through Active Learning, Ruben Martinez-Cantin, Manuel Lopes and Luis Montesano.IEEE - International Conference on Robotics and Automation (ICRA), Anchorage, Alaska, USA, 2010. (pdf)

Active Learning for Reward Estimation in Inverse Reinforcement Learning,Manuel Lopes, Francisco Melo and Luis Montesano. European Conference on Machine Learning (ECML/PKDD), Bled, Slovenia, 2009. (pdf)