This full text paper was peer-reviewed at the direction of IEEE Instrumentation and Measurement Society prior to the acceptance and publication.
Improving Classification of Sit, Stand, and Lie in a Smartphone Human Activity Recognition System Nicole A. Capela
Edward D. Lemaire
Natalie Baddour
Mechanical Engineering, University of Ottawa, Ottawa, Canada Ottawa Hospital Research Institute Ottawa, Canada
Ottawa Hospital Research Institute Ottawa, Canada Faculty of Medicine, University of Ottawa, Ottawa, Canada
Dept. of Mechanical Engineering University of Ottawa Ottawa, Canada
Abstract— Human Activity Recognition (HAR) allows healthcare specialists to obtain clinically useful information about a person’s mobility. When characterizing immobile states with a smartphone, HAR typically relies on phone orientation to differentiate between sit, stand, and lie. While phone orientation is effective for identifying when a person is lying down, sitting and standing can be misclassified since pelvis orientation can be similar. Therefore, training a classifier from this data is difficult. In this paper, a hierarchical classifier that includes the transition phases into and out of a sitting state is proposed to improve sitstand classification. For evaluation, young (age 26 ± 8.9 yrs) and senior (age 73 ± 5.9yrs) participants wore a Blackberry Z10 smartphone on their right front waist and performed a continuous series of 16 activities of daily living. Z10 accelerometer and gyroscope data were processed with a custom HAR classifier that used previous state awareness and transition identification to classify immobile states. Immobile state classification results were compared with (WT) and without (WOT) transition identification and previous state awareness. The WT classifier had significantly greater sit sensitivity and Fscore (p<0.05) than WOT. Stand specificity and F-score for WT were significantly greater than WOT for seniors. WT sit sensitivity was greater than WOT for the young population, though not significantly. All outcomes improved for the young population. These results indicated that examining the transition period before an immobile state can improve immobile state recognition. Sit-stand classification on a continuous daily activity data set was comparable to the current literature and was achieved without the use of computationally intensive feature spaces or classifiers. Keywords—Human activity recognition; HAR; activities of daily living; ADL; smartphone; movement; mobility; accelerometer
I. INTRODUCTION Human activity recognition (HAR) using wearable sensors can offer valuable information to healthcare specialists about a person’s daily activities, thus providing insight into their mobility status and the frequency and duration of activities of daily living (ADL). High mobility classification accuracy has been achieved with systems that have multiple sensor locations; however, specialized wearable sensors are inconvenient for long term use outside a hospital setting. Single sensor systems result in higher user compliance and lower device costs, making it more practical for HAR purposes [1], [2] Smartphones are This project was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC). Smartphones were provided by BlackBerry Inc. 978-1-4799-6477-2/15/$31.00 ©2015 IEEE
ubiquitous and contain useful sensors, and many people carry them in their daily lives, making them an ideal HAR platform. While most HAR devices can accurately distinguish when a person is active or inactive, knowing when a person is standing, sitting, or lying down can give a more representative portrait of a person’s health state. Del Rosario et al. [1] proposed that the cost of misclassifying a sitting state as standing is higher than the cost of misclassifying a sitting state as lying down, since sitting and lying are both sedentary states with lower energy expenditure. When classifying immobile states, HAR devices typically rely on phone orientation to differentiate between sit, stand, and lie [3] This is generally effective for identifying when a person is lying down, but often provides poor results for differentiating between sitting and standing, since pelvis and trunk orientation can be similar when the person is upright. For this reason, sitting and standing states are often mutually misclassified when using a single sensor location [4]. Gjoreski et al. noted that the sit-stand classification issue is largely sidestepped in the literature [5], with some researchers opting not to include both classes [6] and some studies merging the two classes together [7]. Gjoreski et al. used a contextbased ensemble of classifiers and additional features designed specifically to distinguish standing and sitting in order to improve sit-stand recognition accuracy from 62% to 86%. Newly added features should be mutually uncorrelated to existing features in order to provide beneficial gains in classifier accuracy [8], which is difficult when features are derived from a single sensor or sensor location. High dimensionality feature vectors, multiple contexts, and complex classifiers can quickly become computationally intensive. Furthermore, algorithms intended for long term activity recognition must consider the trade-off between load requirements and battery life [9]. Complex models may offer only marginal improvements over simple classifiers despite significant increase in computational cost, and these improvements may disappear when the classifier is used in a real-world environment [8]. The improvements found in sophisticated models over simpler models can degrade on future samples [8], since future distributions are rarely identical to the training data distribution. A simple classifier includes the data’s most important aspects, capturing the underlying phenomena without over-fitting to the training data, and is scalable and computationally efficient [8], [10].