Loading…

Including sensor bias in shape from motion calibration and sensor fusion

Shape from motion data fusion brings a greater degree of autonomy and sensor integration to intelligent systems in which fusion by constant linear transformations is appropriate. To illustrate this, we apply shape from motion techniques to applications involving both similar and disparate sensory in...

Full description

Saved in:
Bibliographic Details
Main Authors: Voyles, R.M., Merrow, J.D., Khosla, P.K.
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Shape from motion data fusion brings a greater degree of autonomy and sensor integration to intelligent systems in which fusion by constant linear transformations is appropriate. To illustrate this, we apply shape from motion techniques to applications involving both similar and disparate sensory information vectors. First, nearly autonomous force/torque sensor calibration is demonstrated through fusion of the individual channels of raw strain gauge data. Gathering only the raw sensor signals, the motion of the force vector (the "motion") and the calibration matrix (the "shape") are simultaneously extracted by singular value decomposition. This calibration example is provided to simply explain the mathematics. Disparate sensory information is fused in a "primordial learning" mobile robot through a similar eigenspace representation. This paper summarizes these shape from motion applications and presents an extension for simultaneously extracting sensor bias.
DOI:10.1109/MFI.1996.568505