Paper
30 June 1994 General learning algorithm for robot vision
Shree K. Nayar, Hiroshi Murase, Sameer A. Nene
Author Affiliations +
Abstract
The problem of vision-based robot positioning and tracking is addressed. A general learning algorithm is presented for determining the mapping between robot position and object appearance. The robot is first moved through several displacements with respect to its desired position, and a large set of object images is acquired. This image set is compressed using principal component analysis to obtain a low-dimensional subspace. Variations in object images due to robot displacements are represented as a compact parametrized manifold in the subspace. While positioning or tracking, errors in end-effector coordinates are efficiently computed from a single brightness image using the parametric manifold representation. The learning component enables accurate visual control without any prior hand-eye calibration. Several experiments have been conducted to demonstrate the practical feasibility of the proposed positioning/tracking approach and its relevance to industrial applications.
© (1994) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Shree K. Nayar, Hiroshi Murase, and Sameer A. Nene "General learning algorithm for robot vision", Proc. SPIE 2304, Neural and Stochastic Methods in Image and Signal Processing III, (30 June 1994); https://doi.org/10.1117/12.179225
Lens.org Logo
CITATIONS
Cited by 13 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Detection and tracking algorithms

Visualization

Optical tracking

Calibration

Sensors

Image processing

Principal component analysis

Back to Top