Published May 2011 | Version public
Book Section - Chapter

Fusion of stereo vision, force-torque, and joint sensors for estimation of in-hand object location

  • 1. ROR icon California Institute of Technology
  • 2. ROR icon Jet Propulsion Lab

Abstract

This paper develops a method to fuse stereo vision, force-torque sensor, and joint angle encoder measurements to estimate and track the location of a grasped object within the hand. We pose the problem as a hybrid systems estimation problem, where the continuous states are the object 6D pose, finger contact location, wrist-to-camera transform and the discrete states are the finger contact modes with the object. This paper develops the key measurement equations that govern the fusion process. Experiments with a Barrett Hand, Bumblebee 2 stereo camera, and an ATI omega force-torque sensor validate and demonstrate the method.

Additional Information

© 2011 IEEE. The author gratefully acknowledges the support from the National Science and Engineering Research Council of Canada (NSERC).

Additional details

Identifiers

Eprint ID
74879
Resolver ID
CaltechAUTHORS:20170307-170440938

Funding

Natural Sciences and Engineering Research Council of Canada (NSERC)

Dates

Created
2017-03-08
Created from EPrint's datestamp field
Updated
2021-11-15
Created from EPrint's last_modified field