F. Mondada, E. Franzi, and &. P. Ienne, Mobile Robot Miniaturisation: A Tool for Investigation in Control AlgorithmsAn Imitation of Life, Third International Symposium on Experimental Robotics, pp.42-45, 1950.

V. Braitenberg, Vehicles: Experiments in synthetic psychology, 1986.

L. Kaelbling, Learning in embedded systems, 1993.

O. Holland and &. M. Snaith, Extending the adaptive heuristic critic and Q-learning: from facts to implications, Artificial Neural Networks, pp.599-602, 1992.

C. J. Watkins, Learning from Delayed Rewards

S. Sehad and &. C. Touzet, Reinforcement Learning and Neural Reinforcement Learning, 1994.

M. Colombetti, M. Dorigo, and G. Borghi, Behavior analysis and training-a methodology for behavior engineering, IEEE Transactions on Systems, Man and Cybernetics, Part B (Cybernetics), vol.26, issue.3, pp.365-380, 1996.
DOI : 10.1109/3477.499789

S. Mahadevan and &. J. Connell, Automatic programming of behavior-based robots using reinforcement learning, Artificial Intelligence, vol.55, issue.2-3, pp.311-365, 1991.
DOI : 10.1016/0004-3702(92)90058-6

R. S. Sutton, Reinforcement Learning Architectures for Animats, Proceedings of the First International Conference on Simulation of Adaptive Behavior, From Animals to Animats, pp.288-296, 1991.

A. G. Barto, R. S. Sutton, and &. W. Anderson, Neuronlike adaptive elements that can solve difficult learning control problems, IEEE Transactions on Systems, Man, and Cybernetics, vol.13, issue.5, pp.834-846, 1983.
DOI : 10.1109/TSMC.1983.6313077

A. G. Barto and &. P. Anandan, Pattern-recognizing stochastic learning automata, IEEE Transactions on Systems, Man, and Cybernetics, vol.15, issue.3, pp.15-360, 1985.
DOI : 10.1109/TSMC.1985.6313371

D. Rumelhart, G. Hinton, and &. R. Williams, Learning Internal Representations by Error Propagation, Parallel Distributed Processing, vol.1, pp.318-362, 1986.
DOI : 10.1016/B978-1-4832-1446-7.50035-2

L. Lin, Reinforcement Learning for Robots Using Neural Networks, -CS-93-103, 1993.

D. Ackley and &. M. Littman, Interactions Between Learning and Evolution, Artificial Life II, SFI Studies Sc. Complexity, pp.487-509, 1991.

C. Touzet and &. N. Giambiasi, Application of connectionist models to fuzzy inference systems, Parallelization in Inference Systems, 1992.
DOI : 10.1007/3-540-55425-4_16

T. Kohonen, Self-Organisation and Associative Memory, 1984.

G. A. Carpenter and &. S. Grossberg, The ART of adaptive pattern recognition by a self-organizing neural network, Proc. IEEE, 1988.
DOI : 10.1109/2.33

J. S. Albus, A New Approach to Manipulator Control: The Cerebellar Model Articulation Controller (CMAC), Journal of Dynamic Systems, Measurement, and Control, vol.97, issue.3, 1975.
DOI : 10.1115/1.3426922

R. A. Mc-callum, Using transitional proximity for faster reinforcement learning

C. Touzet and &. N. Giambiasi, The Connectionist Sequential Machine: a General Model of Sequential Networks, Australian Conf. on Neural Networks

C. Touzet, S. Sehad, and &. N. Giambiasi, Improving Reinforcement Learning of Obstacle Avoidance Behavior with Forbidden Sequences of Actions, International Conference on Robotics and Manufacturing, pp.14-16, 1995.

L. Kaelbling, M. Littman, and A. Moore, Reinforcement Learning: A Survey, Journal of Artificial Intelligence Research, vol.4, pp.237-285, 1996.

M. Dorigo and M. Colombetti, Robot shaping: developing autonomous agents through learning, Artificial Intelligence, vol.71, issue.2, pp.321-370, 1994.
DOI : 10.1016/0004-3702(94)90047-7

M. Mataric, Reward Functions for Accelerated Learning, Proc. of the 11th Intern. Conf. on Machine Learning, 1994.
DOI : 10.1016/B978-1-55860-335-6.50030-1

L. J. Lin, Self-improving reactive agents based on reinforcement learning, planning and teaching, Machine Learning, vol.8, pp.293-321, 1992.

J. and R. Millàn, Rapid, safe, and incremental learning of navigation strategies, IEEE Transactions on Systems, Man and Cybernetics, Part B (Cybernetics), vol.26, issue.3, pp.408-420, 1996.
DOI : 10.1109/3477.499792