Professional Documents
Culture Documents
AbstractA dictionary learning problem is a matrix factorization in which the goal is to factorize a training data matrix, Y,
as the product of a dictionary, D, and a sparse coefficient matrix,
X, as follows, Y DX. Current dictionary learning algorithms
minimize the representation error subject to a constraint on D
(usually having unit column-norms) and sparseness of X. The
resulting problem is not convex with respect to the pair (D, X).
In this letter, we derive a first order series expansion formula
for the factorization, DX. The resulting objective function is
jointly convex with respect to D and X. We simply solve the
resulting problem using alternating minimization and apply some
of the previously suggested algorithms onto our new problem.
Simulation results on recovery of a known dictionary and
dictionary learning for natural image patches show that our
new problem considerably improves performance with a little
additional computational load.
Index TermsDictionary learning, sparse representation, KSVD, MOD.
I. I NTRODUCTION
PARSE and redundant representation modeling has been
shown to be a powerful and efficient tool for signal
analysis and processing [1]. The goal is to represent a given
signal as a linear combination of some given basis functions
in such a way that most of the representations coefficients
be equal to zero or have a small magnitude. More precisely,
consider the signal y Rn and the basis functions D =
[d1 , d2 , . . . , dK ] RnK . In this context, D is called a
dictionary and each of its columns is called an atom. It is
typically assumed that the dictionary is overcomplete, i.e.
K > n. A sparse coding algorithm then seeks the sparsest
representation, x RK , such that y Dx. This model has
received a lot of attention during the last decade, and a lot
of work has been done to theoretically and experimentally
investigate the efficiency of this model in different signal
processing areas [1].
One crucial problem in a sparse representation-based application is how to choose the dictionary. There are many
pre-specified dictionaries, e.g. Fourier, Gabor, Discrete Cosine
Transform (DCT), and wavelet [2]. Though being simple and
having fast computations, these non-adaptive dictionaries are
M. Sadeghi and M. Babaie-Zadeh are with the Electrical Engineering Department, Sharif University of Technology, Tehran, Iran (e-mail:
m.saadeghii@gmail.com; mbzadeh@yahoo.com).
C. Jutten is with the GIPSA-Lab, Department of Images and Signals,
University of Grenoble and Institut Universitaire de France, France (e-mail:
Christian.Jutten@inpg.fr).
This work has been partially funded by Iran National Science Foundation
(INSF) under contract number 91004600, and EU project 2012-ERC-AdG320684 CHESS.
DD,XX
Y DX2F ,
(1)
where .F is the Frobenius norm, and D and X are admissible sets of the dictionary and the coefficient matrix,
respectively. D is usually defined as the set of all dictionaries
with unit column-norms. X constrains the coefficient matrix
to have sparse columns.
Note that the above problem is not convex with respect to
the pair (D, X). Most dictionary learning algorithms attack
this problem by iteratively performing a two-stage procedure:
Starting with an initial dictionary, the following two stages are
repeated several times,
1) Sparse representation:
X(k+1) = argmin Y D(k) X2F ,
(2)
XX
2) Dictionary update:
D(k+1) = argmin Y DX(k+1) 2F .
(3)
DD
closed-form expression1 :
D(k+1) = YXT (XXT )1 ,
(4)
DD,X
(5)
).
(6)
By defining i = {j : xi (j) = 0}, (6) leads to the following
solutions:
i : xi (i ) (DTi Di )1 DTi yi ,
(7)
(8)
(10)
DD,XX
Y + D0 X0 DX0 D0 X2F .
(11)
argmin
DD,XX
(12)
(13)
DD
(14)
(k+1)
di
(k)
TABLE I
P ERCENTAGE OF SUCCESSFUL RECOVERY.
10
20
30
100
Algorithm
MOD
MDU
New MOD
MOD
MDU
New MOD
MOD
MDU
New MOD
MOD
MDU
New MOD
s=3
85.40
87.93
89.40
91.47
91.93
94.07
88.87
92.07
92.67
90.60
91.73
94.00
s=4
77.27
77.80
83.73
91.80
90.40
94.13
92.20
90.87
93.60
91.67
93.00
94.87
s=5
7.73
23.60
28.33
87.87
91.73
93.07
91.60
92.73
95.53
90.93
93.27
93.80
90
80
SNR (dB)
100
70
MOD
MDU
SGK
New MOD
New MDU
New SGK
60
50
40
30
20
10
0
0
10
15
20
25
30
35
40
(a) s = 3
100
90
TABLE II
Algorithm
MOD
MDU
SGK
s=3
2.12 (2.16)
19.47 (19.93)
3.83 (3.88)
s=4
5.42 (5.52)
41.40 (42.08)
8.91 (9.00)
s=5
10.31 (10.51)
70.32 (71.61)
15.73 (15.82)
80
70
MOD
MDU
SGK
New MOD
New MDU
New SGK
60
50
40
30
20
10
0
0
20
30
40
50
60
70
80
(b) s = 4
100
90
80
R EFERENCES
10
70
MOD
MDU
SGK
New MOD
New MDU
New SGK
60
50
40
30
20
10
0
0
20
40
60
80
100
120
(c) s = 5
Fig. 1. Percentage of successful recovery versus alternation number for all
algorithms at SNR = 30 dB.
13
11.5
MOD/SGK
MDU
New MOD
New MDU
New SGK
12.5
12
11.5
MOD/SGK
MDU
New MOD
New MDU
New SGK
11
10.5
11
RMSE
10.5
10
10
9.5
9.5
9
8.5
8
0
10
20
30
40
50
8.5
0
10
20
30
40
50