16 results (BibTeX)

2006


Thumb md bioprint
Molecular Modeling for the BioPrint Pharmaco-informatics Platform

Berenz, V., Tillier, F., Barbosa, F., Boryeu, M., Horvath, D., Froloff, N.

2006 (poster)

[BibTex]

2006

[BibTex]


Design methodologies for central pattern generators: an application to crawling humanoids

Righetti, L., Ijspeert, A.

In Proceedings of Robotics: Science and Systems, pages: 191-198, 2006 (inproceedings)

[BibTex]

[BibTex]


Adaptive Frequency Oscillators applied to dynamic walking II. Adapting to resonant body dynamics

Buchli, J., Righetti, L., Ijspeert, A.

In Proceedings of Dynamic Walking, 2006 (inproceedings)

[BibTex]

[BibTex]


Design methodologies for central pattern generators: towards ’intelligent’ locomotion in robots

Righetti, L., Ijspeert, A.

In Proceedings of 50th anniversary summit of artificial intelligence, 2006 (inproceedings)

[BibTex]

[BibTex]


Movement generation using dynamical systems : a humanoid robot performing a drumming task

Degallier, S., Santos, C., Righetti, L., Ijspeert, A.

In 6th IEEE-RAS International Conference on Humanoid Robots, 2006, pages: 512-517, 2006 (inproceedings)

[BibTex]

[BibTex]


Adaptive frequency oscillators applied to dynamic walking I. Programmable pattern generators

Righetti, L., Buchli, J., Ijspeert, A.

In Proceedings of Dynamic Walking, 2006 (inproceedings)

[BibTex]

[BibTex]


The RobotCub project – an open framework for research in embodied cognition

Metta, G., Sandini, G., Vernon, D., Caldwell, D., Tsagarakis, N., Beira, R., Santos-Victor, J., Ijspeert, A., Righetti, L., Cappiello, G., Stellin, G., Becchi, F.

In Humanoids Workshop, Proceedings of the IEEE–RAS International Conference on Humanoid Robots, 2006 (inproceedings)

[BibTex]

[BibTex]


Adaptive dynamical systems: A promising tool for embodied artificial intelligence

Buchli, J., Righetti, L., Ijspeert, A.

In Proceedings of 50th anniversary summit of artificial intelligence, 2006 (inproceedings)

[BibTex]

[BibTex]


Dynamic Hebbian learning in adaptive frequency oscillators

Righetti, L., Buchli, J., Ijspeert, A.

Physica D, 216(2):269-281, 2006 (article)

[BibTex]

[BibTex]


Programmable central pattern generators: an application to biped locomotion control

Righetti, L., Ijspeert, A.

Proceedings of the 2006 IEEE International Conference on Robotics and Automation, 2006 (proceedings)

DOI Project Page [BibTex]

DOI Project Page [BibTex]


Statistical Learning of LQG controllers

Theodorou, E.

Technical Report-2006-1, Computational Action and Vision Lab University of Minnesota, 2006, clmc (techreport)

PDF [BibTex]

PDF [BibTex]


Approximate nearest neighbor regression in very high dimensions

Vijayakumar, S., DSouza, A., Schaal, S.

In Nearest-Neighbor Methods in Learning and Vision, pages: 103-142, (Editors: Shakhnarovich, G.;Darrell, T.;Indyk, P.), Cambridge, MA: MIT Press, 2006, clmc (inbook)

link (url) [BibTex]

link (url) [BibTex]


Learning operational space control

Peters, J., Schaal, S.

In Robotics: Science and Systems II (RSS 2006), pages: 255-262, (Editors: Gaurav S. Sukhatme and Stefan Schaal and Wolfram Burgard and Dieter Fox), Cambridge, MA: MIT Press, RSS , 2006, clmc (inproceedings)

Abstract
While operational space control is of essential importance for robotics and well-understood from an analytical point of view, it can be prohibitively hard to achieve accurate control in face of modeling errors, which are inevitable in complex robots, e.g., humanoid robots. In such cases, learning control methods can offer an interesting alternative to analytical control algorithms. However, the resulting learning problem is ill-defined as it requires to learn an inverse mapping of a usually redundant system, which is well known to suffer from the property of non-covexity of the solution space, i.e., the learning system could generate motor commands that try to steer the robot into physically impossible configurations. A first important insight for this paper is that, nevertheless, a physically correct solution to the inverse problem does exits when learning of the inverse map is performed in a suitable piecewise linear way. The second crucial component for our work is based on a recent insight that many operational space controllers can be understood in terms of a constraint optimal control problem. The cost function associated with this optimal control problem allows us to formulate a learning algorithm that automatically synthesizes a globally consistent desired resolution of redundancy while learning the operational space controller. From the view of machine learning, the learning problem corresponds to a reinforcement learning problem that maximizes an immediate reward and that employs an expectation-maximization policy search algorithm. Evaluations on a three degrees of freedom robot arm illustrate the feasability of our suggested approach.

link (url) [BibTex]

link (url) [BibTex]


Reinforcement Learning for Parameterized Motor Primitives

Peters, J., Schaal, S.

In Proceedings of the 2006 International Joint Conference on Neural Networks, pages: 73-80, IJCNN, 2006, clmc (inproceedings)

Abstract
One of the major challenges in both action generation for robotics and in the understanding of human motor control is to learn the "building blocks of movement generation", called motor primitives. Motor primitives, as used in this paper, are parameterized control policies such as splines or nonlinear differential equations with desired attractor properties. While a lot of progress has been made in teaching parameterized motor primitives using supervised or imitation learning, the self-improvement by interaction of the system with the environment remains a challenging problem. In this paper, we evaluate different reinforcement learning approaches for improving the performance of parameterized motor primitives. For pursuing this goal, we highlight the difficulties with current reinforcement learning methods, and outline both established and novel algorithms for the gradient-based improvement of parameterized policies. We compare these algorithms in the context of motor primitive learning, and show that our most modern algorithm, the Episodic Natural Actor-Critic outperforms previous algorithms by at least an order of magnitude. We demonstrate the efficiency of this reinforcement learning method in the application of learning to hit a baseball with an anthropomorphic robot arm.

link (url) DOI [BibTex]

link (url) DOI [BibTex]


Policy gradient methods for robotics

Peters, J., Schaal, S.

In Proceedings of the IEEE International Conference on Intelligent Robotics Systems, pages: 2219-2225, IROS, 2006, clmc (inproceedings)

Abstract
The aquisition and improvement of motor skills and control policies for robotics from trial and error is of essential importance if robots should ever leave precisely pre-structured environments. However, to date only few existing reinforcement learning methods have been scaled into the domains of highdimensional robots such as manipulator, legged or humanoid robots. Policy gradient methods remain one of the few exceptions and have found a variety of applications. Nevertheless, the application of such methods is not without peril if done in an uninformed manner. In this paper, we give an overview on learning with policy gradient methods for robotics with a strong focus on recent advances in the field. We outline previous applications to robotics and show how the most recently developed methods can significantly improve learning performance. Finally, we evaluate our most promising algorithm in the application of hitting a baseball with an anthropomorphic arm.

link (url) DOI [BibTex]

link (url) DOI [BibTex]