TY - GEN
T1 - Autonomous learning of active multi-scale binocular vision
AU - Lonini, Luca
AU - Zhao, Yu
AU - Chandrashekhariah, Pramod
AU - Shi, Bertram E.
AU - Triesch, Jochen
PY - 2013/12/31
Y1 - 2013/12/31
N2 - We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.
AB - We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.
UR - http://www.scopus.com/inward/record.url?scp=84891105945&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84891105945&partnerID=8YFLogxK
U2 - 10.1109/DevLrn.2013.6652541
DO - 10.1109/DevLrn.2013.6652541
M3 - Conference contribution
AN - SCOPUS:84891105945
SN - 9781479910366
T3 - 2013 IEEE 3rd Joint International Conference on Development and Learning and Epigenetic Robotics, ICDL 2013 - Electronic Conference Proceedings
BT - 2013 IEEE 3rd Joint International Conference on Development and Learning and Epigenetic Robotics, ICDL 2013 - Electronic Conference Proceedings
T2 - 2013 IEEE 3rd Joint International Conference on Development and Learning and Epigenetic Robotics, ICDL 2013
Y2 - 18 August 2013 through 22 August 2013
ER -