Paper
4 April 1997 Feature space trajectory representation for active vision
Michael A. Sipe, David P. Casasent, Leonard Neiberg
Author Affiliations +
Abstract
A new feature space trajectory (FST) description of 3D distorted views of an object is advanced for active vision applications. In an FST, different distorted object views are vertices in feature space. A new eigen-feature space and Fourier transform features are used. Vertices for different adjacent distorted views are connected by straight lines so that an FST is created as the viewpoint changes. Each different object is represented by a distinct FST. An object to be recognized is represented as a point in feature space; the closest FST denotes the class of the object, and the closest line segment on the FST indicates its pose. A new neural network is used to efficiently calculated distances. We discuss its uses in active vision. Apart from an initial estimate of object class and pose, the FST processor can specify where to move the sensor to: confirm close and pose, to grasp the object, or to focus on a specific object part for assembly or inspection. We advanced initial remarks on the number of aspect views needed and which aspect views are needed to represent an object.
© (1997) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Michael A. Sipe, David P. Casasent, and Leonard Neiberg "Feature space trajectory representation for active vision", Proc. SPIE 3077, Applications and Science of Artificial Neural Networks III, (4 April 1997); https://doi.org/10.1117/12.271506
Lens.org Logo
CITATIONS
Cited by 10 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Active vision

3D vision

Fourier transforms

Inspection

Neural networks

Sensors

Back to Top