Professional Documents
Culture Documents
fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
1
AbstractThe increasing popularity of wearable devices in actions being performed. In healthcare, inertial sensor data
recent years means that a diverse range of physiological and func- can be used for monitoring the onset of diseases as well
tional data can now be captured continuously for applications as the efficacy of treatment options [11], [12]. For patients
in sports, wellbeing, and healthcare. This wealth of information
requires efficient methods of classification and analysis where with neurodegenerative diseases, such as Parkinsons, HAR
deep learning is a promising technique for large-scale data analyt- can be used to compile diaries of their daily activities and
ics. Whilst deep learning has been successful in implementations detect episodes such as Freezing-of-Gait events, for assessing
that utilize high performance computing platforms, its use on the patients condition [13]. Quantifying physical activity
low-power wearable devices is limited by resource constraints. through HAR can also provide invaluable information for
In this paper, we propose a deep learning methodology, which
combines features learnt from inertial sensor data together with other applications, such as evaluating the condition of patients
complementary information from a set of shallow features to with chronic obstructive pulmonary disease (COPD) [14],
enable accurate and real-time activity classification. The design of [15] or evaluating the recovery progress of patients during
this combined method aims to overcome some of the limitations rehabilitation [16], [17].
present in a typical deep learning framework where on-node Currently, smartphones, wearable devices, and Internet-of-
computation is required. To optimize the proposed method for
real-time on-node computation, spectral domain pre-processing Things (IoT) are becoming more affordable and ubiquitous.
is used before the data is passed onto the deep learning frame- Many commercial products, such as the Apple Watch, Fit-
work. The classification accuracy of our proposed deep learning bit, and Microsoft Band, and smartphone apps including
approach is evaluated against state-of-the-art methods using both Runkeeper and Strava, are already available for continuous
laboratory and real world activity datasets. Our results show the collection of physiological data. These products typically
validity of the approach on different human activity datasets,
outperforming other methods, including the two methods used contain sensors that enable them to sense the environment,
within our combined pipeline. We also demonstrate that the have modest computing resources for data processing and
computation times for the proposed method are consistent with transfer, and can be placed in a pocket or purse, worn on
the constraints of real-time on-node processing on smartphones the body, or installed at home. Accurate and meaningful
and a wearable sensor platform. interpretation of the recorded physiological data from these
Index TermsDeep Learning, Low-Power Devices, HAR, Ac- devices can be applied potentially to HAR. However, most
tiveMiles, Wearable, Internet-of-Things current commercial products only provide relatively simple
metrics, such as step count or cadence. The emergence of
I. I NTRODUCTION deep learning methodologies, which are able to extract dis-
criminating features from the data, and increased processing
D EEP learning is a paradigm of machine learning that
uses multiple processing layers to infer and extract
information from big data. Research has shown that the use of
capabilities in wearable technologies give rise to the possibility
of performing detailed data analysis in situ and in real-time.
deep learning can achieve improved performance in a range of The ability to perform more complex analysis, such as activity
applications when compared to traditional approaches [1][6]. classification on the wearable device would be advantageous
Conventional learning approaches use a set of pre-designed for the aforementioned applications.
features also known as shallow features to represent the The rest of the paper is organized as follows: In Section II,
data for a specific classification task. In image processing and we introduce the current state-of-the-art in machine learning
machine vision, shallow features such as SIFT or FAST are for HAR. Our proposed methodology is then described in
often used for landmark detection [7], whereas for time-series Section III. Datasets used for performance evaluation are
analysis, statistical parameters are used [8][11]. presented in Section IV along with detailed comparison of
Human Activity Recognition (HAR), for example, generally the different approaches. Our findings and contributions are
exploits time-series data from inertial sensors to identify the concluded in Section V.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
2
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
3
5 5 5
GyroX
GyroX
GyroX
AccX
AccX
AccX
10 10 10
15 15 15
5 5 5
GyroY
GyroY
GyroY
AccY
AccY
AccY
10 10 10
Time
Time
Time
15 15 15
5 5 5
GyroZ
GyroZ
GyroZ
AccZ
AccZ
AccZ
10 10 10
15 15 15
0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8
Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz)
GyroX
GyroX
GyroX
AccX
AccX
AccX
10 10 10
15 15 15
5 5 5
GyroY
GyroY
GyroY
AccY
AccY
AccY
10 10 10
Time
Time
Time
15 15 15
5 5 5
GyroZ
GyroZ
GyroZ
AccZ
AccZ
AccZ
10 10 10
15 15 15
0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8
Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz) Frequency (Hz)
Fig. 2. Examples of averaged spectrograms extracted from different activities of the ActiveMiles dataset. Their representations exhibit different patterns for
feature extraction and class recognition.
B. Extract Segments and sampling rate invariance. This enables the classification
After the raw signals are collected, segments of n samples to be more robust against data shifting in time and also
are extracted and forwarded along processes A and B of against changes in amplitude of the signal and sampling rate.
the pipeline. The number of samples to consider depends Moreover, frequency selection in the spectrogram domain also
on the type of application involved. Of course, increasing provides an implicit way to allow noise filtering of the data
the length of the segments can introduce an improvement in over time.
recognition accuracy, but at the same time it would cause a A spectrogram of an inertial signal x is a new representation
delay in response for real-time applications as longer segments of the signal as a function of frequency and time. Specifically,
of data need to be obtained and the boundary between different the spectrogram is the magnitude squared of the short-time
activities become less well defined. Typically, segments of 4 to Fourier transform (STFT). The procedure for computing the
10 seconds are used for HAR [6]. The reason that segments spectrogram is to divide a longer time signal into short seg-
rather than single data points are used is motivated by the ments of equal length and then compute the Fourier transform
fact that the highly fluctuating raw inertial measurements separately on each shorter segment. This can be expressed as:
makes the classification of a single data point impractical [25].
X
Therefore, segments are obtained using a sliding window ST F T {x[n]}(m, ) = X(m, ) = x[n][nm]ejn
applied individually to each axis of the sensor. n=
(1)
likewise, with signal x[n] and window w[n]. The magnitude
C. Spectrogram and Deep Learning Module squared of the STFT yields the spectrogram of the function:
In process A of Fig. 1, a set of deep features is automatically spectrogram{x(n)}(m, ) = |X(m, )|2 (2)
extracted using the proposed deep learning module. This
module takes advantage of a spectrogram representation and The resulting spectrogram is a matrix st sf where st is the
an efficient design to achieve its task. In previous work, Rav` et number of different short term, time-localized points and sf
al. [1] show the importance of using a suitable domain when is the number of frequencies considered. Therefore, the spec-
a deep learning methodology is applied to time-series data. trogram describes the changing spectra as a function of time.
Specifically, they show that the spectrogram representation In Fig. 2, we show examples of the averaged spectrograms
is essential for extracting interpretable features that capture across different activities. As we can see, their representations
the intensity differences among nearest inertial data points. exhibit different patterns. Specifically, it appears that highly
The spectrogram representation also provides a form of time variable activities exhibit higher spectrogram values along all
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
4
frequencies, instead repetitive activities, such as walking or where dw is the stride of the convolution. These convolutions
running, only show high values on specific frequencies. These produce an output layer, o, with size wpOutputF rame with
discriminative patterns can be detected by the deep learning OutputF rame = (InputF rame kw)/dw + 1 and wp, the
module, which aims to extract features and characterize activ- number of filters. The results of the convolution obtained from
ities. the x, y and z axes of an inertial sensor are summed together
without any discrimination so that the orientation invariance
Once the spectrograms have been computed, they are pro- property is maintained. This helps the proposed deep learning
cessed using the deep learning module. The design of our deep framework to be more generalizable even when variation in
learning module is aimed at overcoming some of the issues the data resulting from different sensor orientation is not well
typically present in a deep learning framework where on- represented in the dataset. The filters applied to the three axes
node computation is required. Specifically, these disadvantages share the same weights, which is important for reducing the
include: i) deep learning modules can contain redundant links number of parameters for each convolution layer.
between pairs of nodes that connect two consecutive layers of
the neural network; ii) correlations in different signal points
D. Shallow Features
are usually overlooked; and iii) a large set of layers can be
built on top of each other to extract a hierarchy of features In process B of Fig. 1, 17 predefined shallow features are
from low-level to high-level. Deep learning approaches with considered. These features, listed in Table I, are extracted
these designs tend to have high computation demands and separately from each segment of each axis, creating a vec-
are unsuitable for low-power devices being considered in this tor representation for the considered segment. This step is
paper. In our proposed approach, we reduce the computation expressed on line 5 in Algorithm 1. In our case, it takes
cost by limiting the connections between the nodes of the six input segments, a[1], a[2], a[3], g[1], g[2], g[3], representing
network and by computing the features efficiently through the respectively the accelerometers and the gyroscope data vector
use of few hidden layers. Specifically, the spectral representa- along the three axes and produces a final vector of 102 features
tions of different axes and sensors are pre-arranged so that the as output.
data represents local correlations and they can be processed
using 1D convolutional kernels with the same principle that E. Classification
CNN [26] follows. These filters are applied repeatedly to the Once both deep and shallow features have been computed
entire spectrogram and the main advantage is that the network they are merged together into a unique vector and classified
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
5
100 100
Accuracy (%)
87.9 D-FoG
90
85 86.2
85.3
88 Skoda
86
80
WISDM-v2.0
84
75 82
0.1 0.2 0.3 1 2 3 4 5
Probability of retaining node Filter Size
Fig. 3. Behavior of the proposed approach by increasing the probability of Fig. 5. Behavior of the proposed approach by increasing the size of
retaining a node in the dropout regularization, where size of the convolutional convolutional kernels in all the convolutional layers, where probability of
kernel is 2, number of levels is 2, and number of filters is 40. retaining a node is 0.1, number of levels is 1, and number of filters is 40.
100
90
88
generalization error.
87.6 87.6
86
86.2
84
IV. E XPERIMENTAL R ESULTS
82 A. Datasets
80 To evaluate the proposed system we analyze the perfor-
1 2 3
# of Levels
mance obtained on complex real world activity data, collected
from multiple users. Five public datasets are analyzed using
Fig. 4. Behavior of the proposed approach by increasing the number of 10-fold cross validation. Table II summarizes these datasets.
convolutional levels, where probability of retaining a node is 0.2, size of the
convolutional kernel is 2, and number of filters is 40. Noteworthy is the release of our dataset, ActiveMiles (avail-
able at http://hamlyn.doc.ic.ac.uk/activemiles/), which contains
unconstrained real world human activity data from 10 subjects
through a fully-connected layer and a soft-max layer, as shown collected using five different smartphones. Each subject was
by lines 20 and 21 of Algorithm 1. asked to annotate the activities they carried out during the
day using an Android app developed for this purpose. There
F. Training Process are no limitations on where the smartphone is located (i.e.
Shallow and deep features are trained together in a unified pocket, bag, or held in the hand). Annotations record the start
deep neural network. During each stage of the training, errors time, end time, and label of a continuous activity. Since each
between the target and obtained values are used in the back- smartphone uses a different brand of sensor, the final dataset
ward propagation routine to update the weights of the different will contain data that has many modalities, including different
hidden layers. Stochastic Gradient Descent (SGD) is used to sampling rates and amplitude ranges. It is one of the largest
minimize the loss function defined by the L2-norm. To further datasets in terms of number of samples with around 30 hours
improve the training procedure of the weights, we have used of labelled raw data, and it is the first database that groups
three regularizations: together data captured using different sensor configurations.
1) Weight Decay: it is a term in the weight update rule that
causes the weights to exponentially decay to zero if no B. Parameters Optimizations
other update is scheduled. It is used to avoid over-fitting. The proposed deep learning framework contains a few
2) Momentum: it is a technique for accelerating gradient hyper-parameters that must be defined before training the final
descent and attempting to move the global minimum model. An optimization process based on a grid search is
of the function. It accumulates a velocity vector in proposed to find the best values for: i) the probability of
directions of persistent reduction in the objective across retaining a node during the dropout regularization; ii) the
iterations. size of the convolutional kernels for all relative convolutional
3) Dropout: it is a technique that prevents overfitting and layers; iii) the total number of convolutional layers; and iv)
provides a way of combining many different neural the total number of filters in each convolutional layer. The
network architectures efficiently for consensus purposes. behavior of the system when these parameters are system-
At each iteration of the training, dropout temporarily atically tested is shown in Fig. 3, Fig. 4 and Fig. 5. In
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
6
TABLE II
S UMMARY OF H UMAN ACTIVITY DATASETS .
Fig. 6. Intel Edison features a dual-core Intel Atom CPU at 500MHz, wireless
connectivity, and compact physical dimensions at 35.5 25.0 3.9mm. It
is a small but powerful platform that is well suited for wearable devices.
(a) (b) (c)
Fig. 3, we can infer for datasets that have many classes Fig. 7. Graphical user interface used in the app ActiveMiles. 7(a) shows the
main view presented to the user for browsing all the recorded data of the app.
and large variability, increasing the probability of retaining In 7(b) a pie chart shows the percentage of the different activities captured
a node during dropout makes the results worse. Instead, for during the day whereas in 7(c), a histogram is used to show the METS per
datasets with fewer classes and less variability, increasing minute computed in each hour.
this probability provides better performance. The results in
Fig. 4 show instead that the proposed approach requires only toring of human activity throughout the day and a complete
a few levels in order to obtain a complete hierarchy of learnt summary of the detected activities is presented to the user,
features and increasing the number of levels does not result in as shown in Fig. 7(b). The app also provides an estimation
substantial improvement. Finally, Fig. 5 shows that the optimal of the metabolic equivalents (METS) per minute captured
size of the 1D convolutional kernels in all the convolutional for each hour of the day (Fig. 7(c)). Differently from other
layers is two or three depending, also in this case, on the energy expenditure metrics, such as Calories or Joules, METS
complexity of the datasets. For simple datasets, a filter size is a biological parameter that expresses a normalized level
of two seem to provide the best performance, while for more of physical activity that does not factor in persons weight.
complex datasets this size needs to be increased up to three. In the ActiveMiles app, METS are measured by multiplying
In our experiments, we also discover that 32 filters in each the amount of minutes spent for each activity by a relatively
convolutional layer and 80 neuron nodes in the fully-connected well defined constant value and summed together [31], [32]
layer are the optimal values. However, since they show that The second implementation of the proposed system is for the
reducing the number of filters does not affect classification Intel Edison (Intel Atom Tangier CPU, 1GB RAM, dimen-
accuracy, the number of filters can be reduced up to 10 if sions: 35.5 25.0 3.9mm). This implementation is used
there are resource limitations. to demonstrate the on-node human activity classification. For
both architectures, we use the FFTW3 library [33] to extract
C. Implementation the spectrogram and the Torch framework [34] to implement
the deep learning model.
The proposed solution has been implemented for dif-
ferent low-power platforms. The deep learning model is
trained off-line on a standard workstation (2 Intel Xeon D. Classification Results
E5-2680v2 CPU, 64GB DDR3 RAM) and then exported To demonstrate the relative merit of the proposed method,
onto these platforms. Only the feed-forward routine of the we have compared the performance of the proposed solu-
deep learning module is used for on-node classification of tion against other methods. We have considered a baseline
the data. Specifically, the proposed approach has been de- approach that is purely based on shallow features, defined
ployed as an app for Android devices and also as an em- as shallow-features-only, which corresponds to process B
bedded algorithm for the Intel Edison Development Plat- of the pipeline of our combined approach. We also consider
form, Fig. 6. The app is available on the Google Store at approaches that are purely based on deep learning, such
the following link: https://play.google.com/store/apps/details? as [1], [6] in our evaluation. In particular, the approach
id=org.imperial.activemilespro. The mobile app allows moni- in [1] corresponds to process A of our combined pipeline.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
7
TABLE III
C OMPARISON OF OUR PROPOSED SOLUTION AGAINST EXISTING METHODS IN DIFFERENT DATASET.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JBHI.2016.2633287, IEEE Journal of
Biomedical and Health Informatics
8
for time-series data classification. The proposed method is [14] L. Atallah, B. Lo, R. Ali, R. King, and G. Z. Yang, Real-time activity
designed to overcome some of the issues typically present classification using ambient and wearable sensors, IEEE Transactions
on Information Technology in Biomedicine, vol. 13, no. 6, pp. 1031
in a deep learning framework when on-node computation is 1039, Nov 2009.
required. Thanks to a pre-arrangement of the data before they [15] F. Pitta, T. Troosters, V. S. Probst, M. A. Spruit, M. Decramer,
are passed into the deep learning framework and thanks to and R. Gosselink, Quantifying physical activity in daily life with
questionnaires and motion sensors in copd, European Respiratory
a simplified computation layer the learnt features are com- Journal, vol. 27, no. 5, pp. 10401055, 2006.
puted efficiently through the use of just a few levels. The [16] K.-H. Chen, P.-C. Chen, K.-C. Liu, and C.-T. Chan, Wearable
experiments conducted show that the accuracy of the proposed sensor-based rehabilitation exercise assessment for knee osteoarthritis,
Sensors, vol. 15, no. 2, p. 4193, 2015.
method is better when compared to the current state-of-the-art [17] N. Bidargaddi, A. Sarela, L. Klingbeil, and M. Karunanithi, Detecting
approaches. Moreover, the ability of the proposed method to walking activity in cardiac rehabilitation by using accelerometer, in
generalize across different classification tasks is demonstrated Intelligent Sensors, Sensor Networks and Information, 2007. ISSNIP
2007. 3rd International Conference on, Dec 2007, pp. 555560.
using a variety of human activity datasets, including datasets [18] M. Shoaib, S. Bosch, O. D. Incel, H. Scholten, and P. J. Havinga, A
collected in unconstrained real world environments. Finally, survey of online activity recognition using mobile phones, Sensors,
we show that the computation time obtained from low-power vol. 15, no. 1, p. 2059, 2015.
[19] C. C. T. Clark, C. M. Barnes, G. Stratton, M. A. McNarry, K. A.
devices, such as smartphones, wearable devices, and IoT, is Mackintosh, and H. D. Summers, A review of emerging analytical
suitable for real-time on-node human activity recognition. techniques for objective physical activity measurement in humans,
Sports Medicine, pp. 19, 2016.
[20] M. Bachlin, M. Plotnik, D. Roggen, I. Maidan, J. M. Hausdorff,
R EFERENCES N. Giladi, and G. Troster, Wearable Assistant for Parkinsons Disease
Patients with the Freezing of Gait Symptom, Trans. Info. Tech.
[1] D. Ravi, C. Wong, B. Lo, and G.-Z. Yang, Deep Learning for Human Biomed., vol. 14, no. 2, pp. 436446, Mar. 2010.
Activity Recognition: A Resource Efficient Implementation on Low- [21] Y. Lu, Y. Wei, L. Liu, J. Zhong, L. Sun, and Y. Liu,
Power Devices, in Wearable and Implantable Body Sensor Networks Towards unsupervised physical activity recognition using smartphone
(BSN), 2016 IEEE 13th International Conference on, Jun. 2016. accelerometers, pp. 119, 2016.
[2] A. Krizhevsky, I. Sutskever, and G. E. Hinton, Imagenet classification [22] A. Ignatov and V. Strijov, Human activity recognition using
with deep convolutional neural networks, in Advances in Neural quasiperiodic time series collected from a single tri-axial accelerometer,
Information Processing Systems 25, F. Pereira, C. J. C. Burges, pp. 114, 2015.
L. Bottou, and K. Q. Weinberger, Eds. Curran Associates, Inc., 2012, [23] K. H. Walse, R. V. Dharaskar, and V. M. Thakare, A study of human
pp. 10971105. activity recognition using AdaBoost classifiers on WISDM dataset, The
[3] M. Zeng, L. T. Nguyen, B. Yu, O. J. Mengshoel, J. Zhu, P. Wu, Institute of Integrative Omics and Applied Biotechnology Journal, vol. 7,
and J. Zhang, Convolutional Neural Networks for human activity no. 2, pp. 6876, Jan. 2016.
recognition using mobile sensors, in Mobile Computing, Applications [24] C. Catal, S. Tufekci, E. Pirmit, and G. Kocabag, On the use of
and Services (MobiCASE), 2014 6th International Conference on. ensemble of classifiers for accelerometer-based activity recognition,
IEEE, Nov. 2014, pp. 197205. Applied Soft Computing, vol. 37, pp. 10181022, Dec. 2015.
[4] J. B. Yang, M. N. Nguyen, P. P. San, X. L. Li, and S. Krishnaswamy, [25] O. D. Lara and M. A. Labrador, A Survey on Human Activity
Deep Convolutional Neural Networks on Multichannel Time Series for Recognition using Wearable Sensors, IEEE Communications Surveys
Human Activity Recognition, in Proceedings of the 24th International & Tutorials, vol. 15, no. 3, pp. 11921209, 2013.
Conference on Artificial Intelligence, ser. IJCAI15. AAAI Press, [26] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, Gradient-based learning
2015, pp. 39954001. applied to document recognition, Proceedings of the IEEE, vol. 86,
[5] Y. Chen and Y. Xue, A Deep Learning Approach to Human Activity no. 11, pp. 22782324, 1998.
Recognition Based on Single Accelerometer, in Systems, Man, and [27] J. R. Kwapisz, G. M. Weiss, and S. A. Moore, Activity Recognition
Cybernetics (SMC), 2015 IEEE International Conference on. IEEE, Using Cell Phone Accelerometers, SIGKDD Explor. Newsl., vol. 12,
Oct. 2015, pp. 14881492. no. 2, pp. 7482, Mar. 2011.
[6] M. A. Alsheikh, A. Selim, D. Niyato, L. Doyle, S. Lin, and H.-P. Tan, [28] J. W. Lockhart, G. M. Weiss, J. C. Xue, S. T. Gallagher, A. B. Grosner,
Deep Activity Recognition Models with Triaxial Accelerometers, and T. T. Pulickal, Design Considerations for the WISDM Smart
Nov. 2015. Phone-based Sensor Mining Architecture, in Proceedings of the Fifth
[7] D. G. Lowe, Object recognition from local scale-invariant features, in International Workshop on Knowledge Discovery from Sensor Data,
Computer vision, 1999. The proceedings of the seventh IEEE interna- ser. SensorKDD 11. New York, NY, USA: ACM, 2011, pp. 2533.
tional conference on, vol. 2. Ieee, 1999, pp. 11501157. [29] G. M. Weiss and J. W. Lockhart, The impact of personalization on
[8] A. Bulling, U. Blanke, and B. Schiele, A Tutorial on Human Activity smartphone-based activity recognition, in AAAI Workshop on Activity
Recognition Using Body-worn Inertial Sensors, ACM Comput. Surv., Context Representation: Techniques and Languages, 2012.
vol. 46, no. 3, Jan. 2014. [30] P. Zappi, C. Lombriser, T. Stiefmeier, E. Farella, D. Roggen,
[9] T. Huynh and B. Schiele, Analyzing Features for Activity Recognition, L. Benini, and G. Troster, Activity Recognition from On-Body
in Proceedings of the 2005 Joint Conference on Smart Objects and Sensors: Accuracy-Power Trade-Off by Dynamic Sensor Selection, in
Ambient Intelligence: Innovative Context-aware Services: Usages and Wireless Sensor Networks, ser. Lecture Notes in Computer Science,
Technologies, ser. sOc-EUSAI 05. New York, NY, USA: ACM, 2005, R. Verdone, Ed. Berlin, Heidelberg: Springer Berlin Heidelberg, 2008,
pp. 159163. vol. 4913, ch. 2, pp. 1733.
[10] J. Lin, E. Keogh, S. Lonardi, and B. Chiu, A Symbolic Representation [31] B. E. Ainsworth, W. L. Haskell, S. D. Herrmann, N. Meckes, D. R.
of Time Series, with Implications for Streaming Algorithms, in Bassett Jr, C. Tudor-Locke, J. L. Greer, J. Vezina, M. C. Whitt-Glover,
Proceedings of the 8th ACM SIGMOD Workshop on Research Issues in and A. S. Leon, 2011 compendium of physical activities: a second
Data Mining and Knowledge Discovery, ser. DMKD 03. New York, update of codes and met values. Medicine and science in sports and
NY, USA: ACM, 2003, pp. 211. exercise, vol. 43, no. 8, pp. 15751581, 2011.
[11] G.-Z. Yang, Body Sensor Networks, 2nd Edition, Springer, 2014, [32] D. Ravi, B. Lo, and G.-Z. Yang, Real-time food intake classification
ISBN 978-1-4471-6374-9. and energy expenditure estimation on a mobile device, in Wearable
[12] M. Ermes, J. Prkk, J. Mntyjrvi, and I. Korhonen, Detection of daily and Implantable Body Sensor Networks (BSN), 2015 IEEE 12th
activities and sports with wearable sensors in controlled and uncon- International Conference on, Jun. 2015, pp. 16.
trolled conditions, IEEE Transactions on Information Technology in [33] M. Frigo and S. G. Johnson, The design and implementation of fftw3,
Biomedicine, vol. 12, no. 1, pp. 2026, Jan 2008. Proceedings of the IEEE, vol. 93, no. 2, pp. 216231, 2005.
[13] P. Angeles, M. Mace, M. Admiraal, E. Burdet, N. Pavese, [34] R. Collobert, S. Bengio, and J. Mariethoz, Torch: a modular machine
and R. Vaidyanathan, A Wearable Automated System to Quantify learning software library, IDIAP, Tech. Rep., 2002.
Parkinsonian Symptoms Enabling Closed Loop Deep Brain Stimulation.
Cham: Springer International Publishing, 2016, pp. 819.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.