Fusion of Range and Intensity Information for View Invariant Gesture Recognition

Michael Boelstoft Holte, Thomas B. Moeslund, Preben Fihl

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

31 Citations (Scopus)

Abstract

This paper presents a system for view invariant gesture recognition. The approach is based on 3D data from a CSEM SwissRanger SR-2 camera. This camera produces both a depth map as well as an intensity image of a scene. Since the two information types are aligned, we can use the intensity image to define a region of interest for the relevant 3D data. This data fusion improves the quality of the range data and hence results in better recognition. The gesture recognition is based on finding motion primitives in the 3D data. The primitives are represented compactly and view invariant using harmonic shape context. A probabilistic Edit Distance classifier is applied to identify which gesture best describes a string of primitives. The approach is trained on data from one viewpoint and tested on data from a different viewpoint. The recognition rate is 92.9% which is similar to the recognition rate when training and testing on gestures from the same viewpoint, hence the approach is indeed view invariant.
Original languageEnglish
Title of host publicationComputer Vision and Pattern Recognition Workshops, 2008
PublisherElectrical Engineering/Electronics, Computer, Communications and Information Technology Association
Publication date2008
Pages1-7
ISBN (Print)9781424423392
Publication statusPublished - 2008
EventComputer Vision and Pattern Recognition Workshops, 2008. CVPR Workshops 2008. IEEE Computer Society Conference on - Anchorage, Alaska, Canada
Duration: 23 Jun 200828 Jun 2008

Conference

ConferenceComputer Vision and Pattern Recognition Workshops, 2008. CVPR Workshops 2008. IEEE Computer Society Conference on
Country/TerritoryCanada
CityAnchorage, Alaska
Period23/06/200828/06/2008

Cite this