ppt - Soft Computing Lab.

Download Report

Transcript ppt - Soft Computing Lab.

Gesture spotting with body-worn inertial sensors
to detect user activities
Holger Junker, Oliver Amft, Paul Lukowicz, and Gerhard Troster
Pattern Recognition, vol. 41, no. 6, pp. 2010-2024, 2008
2010. 04. 08
Jongwon Yoon
Contents
• Introduction
– Related works
– Contributions
– Terminologies
• Spotting approach
• Case studies
• Spotting implementation
– Preselection stage
– Classification stage
• Experiments
• Results
• Discussion
• Conclusion
Introduction
• Activity recognition
– Motivated by a variety of mobile and ubiquitous computing applications
• Body-mounted motion sensors for activity recognition
– Advantage : Only influenced by user activity
– Difficult to extract relevant features
• Information is often ambiguous and incomplete
• Sensors do not provide exact trajectory because of gravity and arm speed changes
• Solution
– Spotting of sporadically occurring activities
Related works
Introduction
• Wearable instrumentation for gesture recognition
–
–
–
–
Kung Fu moves (Chambers et al., 2002)
“atomic” gestures recognition (Benbasat, 2000)
House holding activities recognition (Bao, 2003)
Workshop activities recognition (Lukowicz et el., 2004)
• Spotting task
– HMM-based endpoint detection in continuous data (Deng and Tsui, 2000)
• Used HMM-based accumulation score
• Search start point using the viterbi algorithm
– HMM-based Threshold model (Lee and Kim, 1999)
• Calculates the likelihood threshold of an input pattern
– Partitioning the incoming data using an intensity analysis (Lukowicz, 2004)
Contributions
Introduction
• Two-stage gesture spotting method
– Novel method based on body-worn motion sensors
– Specifically designed towards the needs and constraints of activity recognition in
wearable and pervasive systems
•
•
•
•
Large null class
Lack of appropriate models for the null class
Large variability in the way gestures are performed
Variable gesture length
• Verification of the proposed method on two scenarios
– Comprise nearly a thousand relevant gestures
– Scenario1) Interaction with different everyday objects
• Part of a wide range of wearable systems applications
– Scenario2) Nutrition intake
• Highly specialized application motivated by the needs of a large industry dominated
health monitoring project
Terminologies
Introduction
• Motion segment
– Represents atomic, non-overlapping unit of human motion
– Characterized by their spatio-temporal trajectory
• Motion event
– Span a sequence of motion segments
• Activity
– Describes a situation that may consist of various motion events
• Signal segment
– A slice of sensor data that corresponds to a motion segment
• Candidate section
– A slice of sensor data that may contain a gesture
Spotting approach
• Naïve approach
– Performs on all possible sections in the data stream
– Computational effort problem
• Two-stage gesture spotting method
– Preselection stage
• Localize and preselect sections in the continuous signal stream
– Classification stage
• Classify candidate sections
Case studies
• Case study 1
– Spotting of diverse object interaction gestures
• Key component in a context recognition system
• May facilitate more natural human-computer interfaces
• Case study 2
– Dietary intake gestures
• Become one sensing domain of an automated dietary monitoring system
Spotting implementation
• Framework
• Relevant gestures
Motion segment partitioning
Preseselection stage
• Preselection stage
– 1) Initial partitioning of the signal stream
– 2) Identify potential selection
– 3) Candidate selection
• Partition a motion parameter into non-overlapping, meaningful segments
– Used motion parameter : Pitch and Roll of the lower arm
• Used sliding-window and bottom-up algorithm (SWAB)
– Ex) Partitioning of each buffer of length n
• Step 1) Start from the arbitrary segmentation of the signal into n/2 segments
• Step 2) Calculate the cost of merging each pair of adjacent segments
– Cost : The error of approximating the signal with its linear regression
• Step 3) Merge the lowest cost pair
Motion segment partitioning
(cont.)
Preseselection stage
• Used sliding-window and bottom-up algorithm (SWAB) (cont.)
• Extension of the segmentation algorithm
– To ensure that the algorithm provided
a good approximation
– Merge adjacent segments if their linear
regressions had similar slopes
Section similarity search
Preseselection stage
• Each motion segment endpoint is considered as potential end of a gesture
– For each endpoint, potential start points were derived from preceding motion
segment boundaries
• Confining the search space
– 1) For the actual length T of the section, Tmin ≤ T ≤ Tmax
– 2) For the number of motion segments nMS in the actual section,
NMS,min ≤ nMS ≤ NMS,max
Section similarity search
(cont.)
Preseselection stage
• Searching
– Used simple single-value features
• Min / max signal values, sum of signal samples, duration of the gesture …
– If d(fPS;Gk) smaller than a gesture-specific threshold ▶ Contain gesture Gk
• Selection of candidate sections
– Collision of two sections can be occurred
– Select sections with the smallest similarity
Classification stage
Spotting implementation
• HMM based classification
• Features
–
–
–
–
–
Pitch and roll angles from the lower / upper arm sensors
Derivative of the acceleration signal from the lower arm
The cumulative sum of the acceleration from the lower arm
Derivative of the rate of turn signal from the lower sensor
The cumulative sum of the rate of turn from the lower arm
• Model
– Single Gaussian models
– Consisted of 4-10 states
Experiments
• Experimental setting
– Five inertial sensors
– One female and three male
• Right-handed
• Aged 25-35 years
• Data sets
– No constraints to the movements of the
subjects
• To obtain data sets with a realistic zeroclass
– Eight additional similar gestures
• To enrich the diversity of movements
Evaluation metrics
Results
• Recall and Precision
• Other evaluation metrics
Preselection stage
• Precision-recall curves
• Evaluation results
Results
Classification stage
Results
• Initial testing
– Case 1 : 98.4% / Case 2 : 97.4%
• Classification of candidate sections
Extensions of the Classification
Results
• Including Zero-class model
– Case 1 : Extracted from all relevant gesture models
– Case 2 : Constructed on the basis of additional gestures that were carried out by
the subjects
• Summary of the total spotting results
Conclusion
• Similarity-based search
– Way to avoid the explicit modeling of a zero-class
• Explicit zero-class model can be added to improve the recognition
– Permits different feature sets for individual gestures
• Future work
– Additional challenges
• Differences in the size and consistency of food pieces
• Additional degrees of freedom
• Temporal aspects
– The presented spotting approach can be applied to other types of motion events