Professional Documents
Culture Documents
Signal Processing
journal homepage: www.elsevier.com/locate/sigpro
a r t i c l e in f o
abstract
Article history:
Received 24 January 2014
Received in revised form
20 October 2014
Accepted 2 December 2014
Available online 15 January 2015
This paper addresses the problem of feature extraction for signal classification. It proposes
to build features by designing a data-driven filter bank and by pooling the timefrequency
representation to provide time-invariant features. For this purpose, our work tackles the
problem of jointly learning the filters of a filter bank with a support vector machine. It is
shown that, in a restrictive case (but consistent to prevent overfitting), the problem boils
down to a multiple kernel learning instance with infinitely many kernels. To solve such a
problem, we build upon existing methods and propose an active constraint algorithm able
to handle a non-convex combination of an infinite number of kernels. Numerical
experiments on both a braincomputer interface dataset and a scene classification
problem prove empirically the appeal of our method.
& 2015 Elsevier B.V. All rights reserved.
Keywords:
Signal classification
Filter bank
SVM
Kernel learning
1. Introduction
The problem of signal classification is now becoming
more and more ubiquitous, ranging from phoneme or
environmental signal to biomedical signal classification. As
classifiers are often built on a geometric interpretation (the
way some points are arranged in a space), the usual trend to
classify signals is first to extract features from the signals,
and then to feed the classifier with them. The classifier can
thus learn an optimal decision function based on features
extracted from some training examples. Such features are
diverse according to the given classification problem: physical perceptions (loudness), statistical moments (covariance
matrix), spectral characterization (Fourier transform), time
frequency (TF) representations (spectrograms and wavelet
decompositions) and so on. In this scope, features are chosen
so as to characterize similarities within a class and disparities
between classes to be distinguished. This property of the
n
Corresponding author.
E-mail addresses: maxime.sangnier@gmail.com (M. Sangnier),
jerome.gauthier@cea.fr (J. Gauthier),
alain.rakotomamonjy@univ-rouen.fr (A. Rakotomamonjy).
http://dx.doi.org/10.1016/j.sigpro.2014.12.028
0165-1684/& 2015 Elsevier B.V. All rights reserved.
125
Table 1
Definitions.
Symbol Definition
1
;
jj
X
K
K
Imaginary unit.
Indicator vector (its size depends on the context).
Pointwise inequalities.
Pointwise modulus.
Depending on the context, the Hadamard product between
two matrices or composition between two functions.
Convolution
n
Set of signals: X [ 1
n1 K .
Divising ring (either R or C).
Positive semidefinite matrix.
126
def
ux N l hl x 1 r l r d
8
1
20
ql
<
X
def
~
N l 4@
hl jxi
j 1A
:
j1
1rirn
39
=
5
;
1rlrd
Fig. 2. Filter bank of a Haar wavelet transform at scale 3 (in the Fourier
domain). (a) Filter h1 (N 1 8). (b) Filter h2 (N 2 8). (c) Filter h3 (N 3 4).
(d) Filter h4 (N 4 2).
Signal classification usually needs some elaborate features based on local behaviors. This is mostly due to
intraclass discrepancies, which often require nonlinearities
to be compensated. This is the role of the pooling function
, we mentioned previously. The main pitfall of the
pooling function is aimed at overcoming is the random
time shift inherent to real world signals acquisition. That is
the pet hate shared by all the pooling functions used in
this study (Table 2). Secondarily, pooling functions are also
aimed at reducing the dimension of features in order to
curb the curse of dimensionality.
The most widespread pooling function may be the
p norm (computed over the whole filtered signal), particularly with p 2. This method is quite efficient against random
time shifts and is thus often used in signal classification for
the resulting information only depends on the frequency, not
on the time evolution of the signal. However that lack of time
resolution is also the main drawback of this first pooling
function. A way to overcome it is to compute local invariant
features, for instance by partitioning the input signal into a
set of non-overlapping segments and, for each such subsignal, outputting the maximum or the mean value (max and
mean pooling). These methods have been first introduced for
CNNs [12] in order to absorb space shifts while keeping a
trade-off with the space resolution. Finally, the scattering
pooling, recently introduced by [24], has been designed for
complex valued signals. The time shift invariance comes from
both the modulus and the averaging operations. It is shown
in [25] that this kind of pooling is linked with the logarithmic
aggregation of the frequencies used to compute the Melfrequency cepstral coefficients (MFCCs).
In Table 2, definitions are given for a single signal. Yet,
these definitions are easily spread out to a set of signals
(for instance a TF representation) through : X d -X d
defined by
d
xl l 1 :
fxl gdl 1 def
For instance, when applied to a TF representation of a
signal, the 2 norm provides the marginal distribution of
the energy for the given frequency bands.
Table 2
Pooling functions. Parameters: p A N, w A N, g a Gaussian lowpass filter.
p norm
n
P
Localp norm
!1=p 1n=wc
A
jxjjp
j i1w 1
Max
n=wc
max
i 1w 1 r j r iw
Mean
Scattering
jxjg
mean
i 1w 1 r j r iw
i1
xj
i1
minimize Ju; k;
uAT
N
X
ni yi kuxi ; ux:
minimize
f A H;b A R; A RN ;u A T
1
2
J f J 2H C1T
yi f uxi b Z 1 i ;
0:
8 iA NN
Both problems (2) and (5) are hard in that they are nonconvex. Indeed, as the filtered data is consecutively pooled
and kernelized, each non-convexity in the pooling function and in the mapping induced by the kernel k results
in a non-convexity of the objective function with respect
to the IRs [29,3]. This effect can be accentuated if the IRs
are naively parameterized in a non-convex way (for
instance by the cutoff frequencies for a bandpass filter).
As a consequence, the strategy of resolution of our learning problem is a major point to design an algorithm as
efficient as possible, considering both classification accuracy and computational complexity. The reason to prefer
the wrapper strategy (2) to the direct one (5) is that the
former enables us to easily deal with infinite mappings
induced by such kernels like the Gaussian one (definition
(9)). Moreover, the wrapper strategy provides us with an
algorithm that benefits from the convexity of the SVM
problem and from the current SVM solvers efficiency (for
instance [30]) in terms of precision and of training time.
Note that problems (2) and (5) are different but if they are
solvable, then they have the same global minimas. As a
consequence, choosing the problem (2) rather than (5) is
equivalent to choosing a kind of resolution strategy.
3. Solving the problem
This section describes how to solve the FB learning
problem. First, we exhibit the few hypotheses, that
restrict the framework in which our work holds, and
then we propose an algorithm to solve the resulting
problem. The algorithm we present is an extension of
the generalized multiple kernel learning algorithm [29]
so as to handle a continuously parametrized family of
kernels. Finally, we detail how our framework interestingly enable to learn the pooling function, which was
supposed fixed before.
yi f uxi b Z1 i ;
0;
def
such that
i1
n=wc
such that
xj
f A H;b A R; A RN
f ux
!1=p
iw
P
i 1
n 4 0
i
jxjjp
j1
Definition of x
Name
127
8 i A NN
x A Kn sign f ux b ;
128
AA
def
then, by denoting u0 h ; N A A the FB of the normalized filters and by setting x ux along with
y uz, we get from (6) and (7):
8 x; z A Kn ;
kux; uz k~ 2 u0 x; u0 z;
def
z fz g A A A X d :
kx; z exp J vecx vecz J 22
!
X
exp
x z 22 :
8 x fx g A A ;
def
and
8 x fx g A A ; z fz g A A A X d :
X
def
x jz ;
kx; z vecxjvecz
2
AA
def
k x; z k x; z ;
AA
10
8 x; z A Kn ;
vecuxjvecuz
2
vecfN ~ h xg A A jvecfN ~ h zg A A
2
X 2
~ N h xjN h z :
6
AA
AA
8 x A Kd : ux ux:
def
kux; uz k~ 2 u0 x; u0 z:
The last relation means that comparing two TF representations with a Gaussian kernel is like computing a multiplicative combination of the similarities between each
frequency band in an infinite-dimensional space.
Thanks to (6), we have shown that for both kernels k
(linear and Gaussian), the problem of learning the weights
minimize J u0 ; k
A RP
such that
8 T
>
<1 1
0
>
:
FS;
11
def
Data: training dataset xi ; yi 1 r i r N .
3
4
5
6
7
8
9
10
11
12
13
14
129
2
def
D N h xi N h xj k
;
2
1 r i;j r N
K k u0 xi ; u0 xj 1 r i;j r N :
def
0 V 0 ;
8 A P; V r
0 A P;
0 40
where
T
V n Y D K Y n ;
and n is the optimal dual
variable of the SVM problem
applied to u0 xi ; yi 1 r i r N with the kernel k .
Respectively, when the kernel k is linear, the main
condition to be violated is identical with [32,6]:
V n YK Y n ;
T
n
8 x A Kn ; f un x
130
def
N
X
ni yi k u0 x; u0 xi
n
i1
N
X
ni yi k un x; un xi ;
12
i1
p
X
~ 2r r uxjr uz
r1
13
n
8 x A Kn ; f n ux
def
N
X
ni yi k 0 ux; 0 uxi
n
i1
NX
ni yi k n ux; n uxi ;
14
i1
where 0 r 1 r r r p , n is the
optimal dual
variable of
the SVM problem applied to 0 uxi ; yi 1 r i r N with
the kernel kn , n is the optimal vector of weights from the
p
MKL problem and n nr r 1 r r r p .
Learning the pooling function seems advantageous, since
the practitioner does not need to choose a particular pooling
function beforehand (in practice, it is quite difficult to have
an intuition on which pooling function will perform the
best), nevertheless we have to be careful since adding new
variables to the optimization scheme may promote overfitting. For this purpose, we propose a three-stage approach:
in a first step, we draw a linear grid of parameters for the
finite IR filters and we learn the pooling function with this
def
15
AP
V ^ r
V ;
A P;
4 0
8t A Nq :
131
q
q
sin off t off
sin on t on
Wt
2
2
hon ;off t
;
t
def
8 t A Nq :
h ; t e 8t q=2= q
def
1
2
e4 t q=2=n e =2 ;
132
Table 3
Confronted methods.
Abbreviation
Analysis
Pooling
Classifier
Remark
SVM
Max
DFT
MFCC
CNN
WKL
Band-Max
Band2
Morlet-scattering
Band-pooling
Morlet-pooling
Max
Mean
2 norm
Max
2 norm
Scattering
Learned
Learned
SVM
SVM
SVM
SVM
MLP
MKL
SVM
SVM
SVM
SVM
SVM
[12]
[6]
Proposed
Proposed
Proposed
Proposed
Proposed
method
method
method
method
method
Fig. 5. Toy dataset: accuracy with respect to the size of the training set.
Fig. 6. Toy dataset: instance of learned filter banks (each color embodies
a filter in the Fourier domain). (For interpretation of the references to
color in this figure caption, the reader is referred to the web version of
this paper.)
an audio stream by providing a semantic label. For instance, the database we are interested in is made up of 10
classes [39]: busy street, quiet street, park, open-air
market, bus, subway-train (or tube), restaurant, store/
supermarket, office and subway station (or tube station).
The database contains 10 audio recordings (30-s segments) for each scene. To make it, three different recordists recorded each scene type. They visited a wide variety
of locations in Greater London over a period of months
(Summer and Autumn 2012), with moderate weather
conditions and varying times of day and week.
In our experiment, each recording is split into 300-ms
frames, filtered and downsampled in order to reduce the
computational complexity. The goal is then to label these
subsignals in the framework of several binary classification
problems, for instance tube vs. tube station. The baseline we
consider is an SVM classifier applied to the MFCCs (MFCC).
This approach, which is often used to classify audio signals,
seems really efficient on some of the biclass problems
studied in this paper (for instance office vs. supermarket).
Results. The box plots in Fig. 7 (BCI experiment) bring
out that the naive approaches, considering either the
signals as classification features (SVM) or the magnitudes
of the Fourier transform (DFT), fail to correctly generalize
from the training dataset. This proves the appeal of datadriven methods for signal classification. Moreover, it
seems that a Morlet wavelet along with a scattering
pooling (Morlet-Scattering) achieves comparable but slightly better results than a bandpass FB with a Max pooling
(Band-Max), and really better results than bandpass filters
with an 2 norm (Band2 ). As the latter method corresponds to considering the marginal distribution of the
133
134
Fig. 7. BCI dataset: classification accuracy for several channels. (a) Channel 9. (b) Channel 12. (c) Channel 17. (d) Channel 29. (e) Channel 30.
135
Fig. 8. Scenes dataset: classification accuracy. (a) Park vs. quiet street. (b) Office vs. supermarket. (c) Office vs. quiet street.
7. Conclusions
Acknowledgments
This paper has introduced a novel approach to learn a onestage filter bank for signal classification. It took a fresh look at
learning a discriminative timefrequency transform compared
to the widespread convolutional neural networks. The method
proposed to jointly learn a filter bank with an SVM classifier.
The SVM kernels, we considered in this paper, were the linear
and the Gaussian ones. They were computed by filtering the
signals thanks to the filter bank, and by pooling the result. We
showed that, by choosing a specific family of filters, defined by
few parameters, the optimization problem is actually a multiple kernel learning problem, where the number of kernels can
be infinite. We thus provided an active constraint algorithm,
that extended existing methods and that is able to handle such
a problem by generating a sparse and finite non-necessarily
136
L; ; J u0 ; k 1T 1 T :
At an equilibrium in the multiple kernel problem
(potentially a local minimum or maximum), the KKT
conditions hold:
8
~
!
T!
>
>
> 1 1; Z c0
>
>
<!
Zc0
>
> 0; 8 A P
>
>
>
: L; ; 0:
Both first conditions (primal and dual feasibility) are
just reminders. The last one is however quite important
and can be rewritten as
J~ 1 0;
16
where
J~ : A RP J u0 ; k :
Combining the third KKT condition with (16) leads to
8
>
J~
>
>
>
<
8 A P;
>
J~
>
>
>
: Z
if 4 0
17
if 0:
def
X
A P; 4 0
J~
maximize
A RN
such that
0C1;
yT 0:
18
J~
1 T K
n Y Y n ;
19
spanning kernels:
k k e
def
AP
d
AP
def
D d xi ; xj 1 r i;j r N ;
then the main equilibrium condition to be violated is
8 A P;
J~
T
n Y D K Y n r :
References
[1] N. Saito, R. Coifman, Local discriminant bases and their applications,
J. Math. Imaging Vis. 5 (1995) 337358.
[2] D. Blei, J. McAuliffe, Supervised topic models, in: Advances in Neural
Information Processing Systems (NIPS), Curran Associates, Inc., NY,
2008.
[3] R. Flamary, D. Tuia, B. Labb, G. Camps-Valls, A. Rakotomamonjy,
Large margin filtering, IEEE Trans. Signal Process. 60 (2012) 648659.
[4] E. Jones, P. Runkle, N. Dasgupta, L. Couchman, L. Carin, Genetic
algorithm wavelet design for signal classification, IEEE Trans. Pattern
Anal. Mach. Intell. 23 (2001) 890895.
[5] D.J. Strauss, G. Steidl, W. Delb, Feature extraction by shape-adapted
local discriminant bases, Signal Process. 83 (2003) 359376.
[6] F. Yger, A. Rakotomamonjy, Wavelet kernel learning, Pattern Recognit. 44 (2011) 26142629.
[7] M. Davy, A. Gretton, A. Doucet, P.J.W. Rayner, Optimized support
vector machines for nonstationary signal classification, IEEE Signal
Process. Lett. 9 (2002) 442445.
[8] P. Honeine, C. Richard, P. Flandrin, J.-B. Pothin, Optimal selection of
timefrequency representations for signal classification: a kerneltarget alignment approach, in: IEEE International Conference on
Acoustics, Speech and Signal Processing, 2006.
[9] J. Mairal, F. Bach, J. Ponce, G. Sapiro, A. Zisserman, Supervised
dictionary learning, in: Advances in Neural Information Processing
Systems (NIPS), 2009.
[10] K. Huang, S. Aviyente, Sparse representation for signal classification,
in: Advances in Neural Information Processing Systems (NIPS), 2006.
[11] F. Rodriguez, G. Sapiro, Sparse Representation for Image Classification: Learning Discriminative and Reconstructive Non-Parametric
Dictionaries, Technical Report, University of Minnesota, 2008.
[12] Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, Gradient-based learning
applied to document recognition, IEEE Proc. 86 (11) (1998)
22782324.
[13] A. Biem, S. Katagiri, E. McDermott, G.-H. Juang, An application of
discriminative feature extraction to filter-bank-based speech recognition, IEEE Trans. Speech Audio Process. 9 (2) (2001) 96110.
[14] P. Vaidyanathan, Multirate Systems and Filter Banks, Prentice Hall,
Englewood Cliffs, New Jersey, 1993.
[15] H. Kha, H. Tuan, T. Nguyen, Efficient design of cosine-modulated
filter banks via convex optimization, IEEE Trans. Signal Process. 57
(2009) 966976.
[16] J. Gauthier, L. Duval, J.-C. Pesquet, Optimization of synthesis oversampled complex filter banks, IEEE Trans. Signal Process. 57 (2009)
38273843.
[17] L.D. Vignolo, H.L. Rufiner, D.H. Milone, J. Goddard, Evolutionary
cepstral coefficients, Appl. Soft Comput. 11 (2011) 34193428.
[18] L.D. Vignolo, H.L. Rufiner, D.H. Milone, J. Goddard, Evolutionary
splines for cepstral filterbank optimization in phoneme classification, EURASIP J. Adv. Signal Process. 2011 (2011) 114.
[19] H.-I. Suk, S.-W. Lee, A novel Bayesian framework for discriminative
feature extraction in braincomputer interfaces, IEEE Trans. Pattern
Anal. Mach. Intell. 35 (2) (2013) 286299.
[20] T. Hastie, R. Tibshirani, J. Friedman, The Elements of Statistical
Learning: Data Mining, Inference, and Prediction, Springer, New
York, 2009.
137