200 likes | 291 Views
Gesture spotting with body-worn inertial sensors to detect user activities Holger Junker, Oliver Amft , Paul Lukowicz , and Gerhard Troster Pattern Recognition, vol. 41, no. 6, pp. 2010-2024, 2008. 2010. 04. 08 Jongwon Yoon. Contents. Introduction Related works Contributions
E N D
Gesture spotting with body-worn inertial sensorsto detect user activitiesHolger Junker, Oliver Amft, Paul Lukowicz, and Gerhard TrosterPattern Recognition, vol. 41, no. 6, pp. 2010-2024, 2008 2010. 04. 08 Jongwon Yoon
Contents • Introduction • Related works • Contributions • Terminologies • Spotting approach • Case studies • Spotting implementation • Preselection stage • Classification stage • Experiments • Results • Discussion • Conclusion
Introduction • Activity recognition • Motivated by a variety of mobile and ubiquitous computing applications • Body-mounted motion sensors for activity recognition • Advantage : Only influenced by user activity • Difficult to extract relevant features • Information is often ambiguous and incomplete • Sensors do not provide exact trajectory because of gravity and arm speed changes • Solution • Spotting of sporadically occurring activities
Related works Introduction • Wearable instrumentation for gesture recognition • Kung Fu moves (Chambers et al., 2002) • “atomic” gestures recognition (Benbasat, 2000) • House holding activities recognition (Bao, 2003) • Workshop activities recognition (Lukowiczet el., 2004) • Spotting task • HMM-based endpoint detection in continuous data (Deng and Tsui, 2000) • Used HMM-based accumulation score • Search start point using the viterbi algorithm • HMM-based Threshold model (Lee and Kim, 1999) • Calculates the likelihood threshold of an input pattern • Partitioning the incoming data using an intensity analysis (Lukowicz, 2004)
Contributions Introduction • Two-stage gesture spotting method • Novel method based on body-worn motion sensors • Specifically designed towards the needs and constraints of activity recognition in wearable and pervasive systems • Large null class • Lack of appropriate models for the null class • Large variability in the way gestures are performed • Variable gesture length • Verification of the proposed method on two scenarios • Comprise nearly a thousand relevant gestures • Scenario1) Interaction with different everyday objects • Part of a wide range of wearable systems applications • Scenario2) Nutrition intake • Highly specialized application motivated by the needs of a large industry dominated health monitoring project
Terminologies Introduction • Motion segment • Represents atomic, non-overlapping unit of human motion • Characterized by their spatio-temporal trajectory • Motion event • Span a sequence of motion segments • Activity • Describes a situation that may consist of various motion events • Signal segment • A slice of sensor data that corresponds to a motion segment • Candidate section • A slice of sensor data that may contain a gesture
Spotting approach • Naïve approach • Performs on all possible sections in the data stream • Computational effort problem • Two-stage gesture spotting method • Preselection stage • Localize and preselect sections in the continuous signal stream • Classification stage • Classify candidate sections
Case studies • Case study 1 • Spotting of diverse object interaction gestures • Key component in a context recognition system • May facilitate more natural human-computer interfaces • Case study 2 • Dietary intake gestures • Become one sensing domain of an automated dietary monitoring system
Spotting implementation • Framework • Relevant gestures
Motion segment partitioning Preseselection stage • Preselection stage • 1) Initial partitioning of the signal stream • 2) Identify potential selection • 3) Candidate selection • Partition a motion parameter into non-overlapping, meaningful segments • Used motion parameter : Pitch and Roll of the lower arm • Used sliding-window and bottom-up algorithm (SWAB) • Ex) Partitioning of each buffer of length n • Step 1) Start from the arbitrary segmentation of the signal into n/2 segments • Step 2) Calculate the cost of merging each pair of adjacent segments • Cost : The error of approximating the signal with its linear regression • Step 3) Merge the lowest cost pair
Motion segment partitioning (cont.)Preseselection stage • Used sliding-window and bottom-up algorithm (SWAB) (cont.) • Extension of the segmentation algorithm • To ensure that the algorithm provided a good approximation • Merge adjacent segments if their linear regressions had similar slopes
Section similarity search Preseselection stage • Each motion segment endpoint is considered as potential end of a gesture • For each endpoint, potential start points were derived from preceding motion segment boundaries • Confining the search space • 1) For the actual length T of the section, Tmin≤ T ≤ Tmax • 2) For the number of motion segments nMS in the actual section, NMS,min≤ nMS≤ NMS,max
Section similarity search (cont.)Preseselection stage • Searching • Used simple single-value features • Min / max signal values, sum of signal samples, duration of the gesture … • If d(fPS;Gk) smaller than a gesture-specific threshold ▶ Contain gesture Gk • Selection of candidate sections • Collision of two sections can be occurred • Select sections with the smallest similarity
Classification stage Spotting implementation • HMM based classification • Features • Pitch and roll angles from the lower / upper arm sensors • Derivative of the acceleration signal from the lower arm • The cumulative sum of the acceleration from the lower arm • Derivative of the rate of turn signal from the lower sensor • The cumulative sum of the rate of turn from the lower arm • Model • Single Gaussian models • Consisted of 4-10 states
Experiments • Experimental setting • Five inertial sensors • One female and three male • Right-handed • Aged 25-35 years • Data sets • No constraints to the movements of the subjects • To obtain data sets with a realistic zero-class • Eight additional similar gestures • To enrich the diversity of movements
Evaluation metrics Results • Recall and Precision • Other evaluation metrics
Preselection stage Results • Precision-recall curves • Evaluation results
Classification stage Results • Initial testing • Case 1 : 98.4% / Case 2 : 97.4% • Classification of candidate sections
Extensions of the Classification Results • Including Zero-class model • Case 1 : Extracted from all relevant gesture models • Case 2 : Constructed on the basis of additional gestures that were carried out by the subjects • Summary of the total spotting results
Conclusion • Similarity-based search • Way to avoid the explicit modeling of a zero-class • Explicit zero-class model can be added to improve the recognition • Permits different feature sets for individual gestures • Future work • Additional challenges • Differences in the size and consistency of food pieces • Additional degrees of freedom • Temporal aspects • The presented spotting approach can be applied to other types of motion events