You are on page 1of 12

Computers and Chemical Engineering 30 (2006) 1502–1513

An overview of subspace identification


S. Joe Qin ∗
Department of Chemical Engineering, The University of Texas at Austin, Austin, TX 78712, USA
Received 20 February 2006; received in revised form 30 May 2006; accepted 31 May 2006

Abstract
This paper provides an overview of the state of the art of subspace identification methods for both open-loop and closed-loop systems. Practical
considerations and future directions are given at the end of the paper.
© 2006 Elsevier Ltd. All rights reserved.

Keywords: Subspace identification; Closed-loop identification; Overview; State space models; Causal model; Practice

1. Introduction such as the prediction error methods (PEMs) (Forssell & Ljung,
1999). It causes additional difficulty for SIMs.
Subspace identification methods (SIM) have enjoyed tremen- Although SIM algorithms are attractive because of the state
dous development in the last 15 years in both theory and practice. space form that is very convenient for estimation, filtering, pre-
SIMs offer an attractive alternative to input-output methods due diction and control, several drawbacks have been recognized. In
to simple and general parametrization for MIMO systems (there general, the estimates from SIMs are not as accurate as those
is no linear input-output parametrization that is general enough from prediction error methods. Further, it is not until recently
for all linear MIMO systems, see (Katayama, 2005)). Most SIMs some SIMs are applicable to closed-loop identification, even
fall into the unifying theorem proposed by van Overschee and de though the data satisfy identifiability conditions for traditional
Moor (1995), among which are canonical variate analysis (CVA) methods such as PEMs.
(Larimore, 1990), N4SID (van Overschee & de Moor, 1994), Unlike PEMs, the traditional SIMs (e.g., CVA, N4SID and
subspace splitting (Jansson & Wahlberg, 1996), and MOESP MOESP) are biased under closed-loop condition, which requires
(Verhaegen & Dewilde, 1992). Based on the unifying theorem, special treatment. Verhaegen (1993) proposed a closed-loop
all these algorithms can be interpreted as a singular value decom- SIM via the identification of an overall open-loop state space
position of a weighted matrix. The statistical properties such model followed by a model reduction step to obtain state space
as consistency and efficiency have been investigated recently representations of plant and controller. Ljung and McKelvey
(Bauer, 2003; Bauer & Ljung, 2002; Gustafsson, 2002; Jansson (1996) investigated the SIM through the classical realization
& Wahlberg, 1998; Knudsen, 2001). path and proposed a recursive approach based on ARX model as
The closed-loop identification is of special interest for a large a feasible closed-loop SIM. Formulated in an errors-in-variables
number of engineering applications. For safety reasons or quality (EIV) framework, Chou and Verhaegen (1997) proposed a new
restrictions, it is desirable that identification experiments are car- SIM that can be applied to closed-loop data. The algorithm has to
ried out under the closed-loop or partial closed-loop condition. treat white input from non-white input differently. Wang and Qin
As pointed out by many researchers (Ljung, 1999; Soderstrom (2002) proposed the use of parity space and principal component
& Stoica, 1989), the fundamental problem with closed-loop data analysis (PCA) for EIV and closed-loop identification which is
is the correlation between the unmeasurable noise and the input. applicable to correlated input excitation. Recent work of Qin
This is true for traditional closed-loop identification approaches and Ljung (2003a), Jansson (2003), and Chiuso and Picci (2005)
analyzed SIMs with feedback, proposed several new closed-loop
SIMs and provided theoretical analysis to these methods.
∗ Tel.: +1 512 471 4417; fax: +1 512 471 7060. The purpose of this paper is to provide an overview of the
E-mail address: qin@che.utexas.edu. state of the art in both open-loop and closed-loop SIMs. The

0098-1354/$ – see front matter © 2006 Elsevier Ltd. All rights reserved.
doi:10.1016/j.compchemeng.2006.05.045
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1503

paper starts with basic stochastic system representations and C, D) matrices. Since AK = A − KC is guaranteed stable even
assumptions, then reviews most existing SIMs in the literature though the original process A matrix is unstable, the predictor
to date. Practical considerations and future directions are given form is numerically advantageous for identifying both stable
to conclude the paper. and unstable processes. The other two model forms may lead to
ill-conditioning for unstable processes (Chiuso & Picci, 2005).
2. Models, notations, and assumptions However, the optimal Kalman gain K is time-varying for finite
number of samples, making AK time varying even though the
2.1. Stochastic state space models original process is time invariant. This is a minor drawback of
the predictor form for limited number of samples.
A stochastic linear system can be written in the following Based on state space description in (4), an extended state
process form, space model can be formulated as
xk+1 = Axk + Buk + wk (1a) yf (k) = Γ̄f xk + Ḡf zf −1 (k) + Df uf (k) + ef (k) (5)

yk = Cxk + Duk + vk (1b) where the subscript f denotes the future horizon. The extended
observability matrix is
where yk ∈ Rny , xk ∈ Rn , uk ∈ Rnu , wk ∈ Rn , and vk ∈ Rny are ⎡ ⎤ ⎡ ⎤
the system output, state, input, state noise, and output measure- C D
⎢ ⎥ ⎢D⎥
ment noise, respectively. A, B, C and D are system matrices with ⎢ CAK ⎥ ⎢ ⎥
⎢ ⎥
appropriate dimensions. Γ̄f = ⎢ . ⎥ ; Df = ⎢ ⎥
⎢ .. ⎥
⎢ .. ⎥ ⎣ . ⎦
It is well known that one can design a Kalman filter for this ⎣ ⎦
f −1
system to estimate the state variables if the system is observable, CAK D
⎡ ⎤
x̂k+1 = Ax̂k + Buk + K(yk − Cx̂k − Duk ) (2) 0 0 ... 0
⎢ ⎥
where K is the steady state Kalman gain that can be obtained ⎢ CBK 0 ... 0 ⎥
⎢ ⎥
from an algebraic Ricatti equation. Denoting Ḡf = ⎢ .. .. .. .. ⎥
⎢ . . . . ⎥
⎣ ⎦
ek = yk − Cx̂k − Duk f −2 f −3
CAK BK CAK BK . . . CBK
as the innovations of the Kalman filter and ignoring the ”∧“on
where the overbar means that the matrix is composed of param-
xk in the rest of this paper, we have the following equivalent
eters of the predictor form.
innovation form,
The input and output are arranged in the following form:
xk+1 = Axk + Buk + Kek (3a) ⎡ ⎤
yk
yk = Cxk + Duk + ek (3b) ⎢ y ⎥
⎢ k+1 ⎥
yf (k) = ⎢⎢ ..

⎥ (6a)
where the innovation ek is white noise and independent of past ⎣ . ⎦
input and output data. The system described by (2) can also be
yk+f −1
represented in the predictor form, ⎡ ⎤
zk
xk+1 = AK xk + BK zk (4a) ⎢ z ⎥
⎢ k+1 ⎥

zf −1 (k) = ⎢ . ⎥ (6b)
yk = Cxk + Duk + ek (4b) ⎥
⎣ .. ⎦
T
where zk = [uTk , ykT ] , AK = A − KC, and BK = [B − KD, K]. zk+f −2
The three model forms, that is, the process form, the inno-
uf (k) and ef (k) are formed similar to yf (k).
vation form, and the predictor form, all can represent the input
By iterating (4) it is straightforward to derive the following
and output data (uk , yk ) exactly. Therefore, one has the option to
relation,
use any of these forms for convenience. For example, the well-
p
known N4SID (Overschee & Moor, 1994) algorithm uses the xk = L̄p zp (k) + AK xk−p (7)
process form. The MOESP (Verhaegen, 1994) algorithm uses
the innovation form. For the convenience of closed-loop identi- where
 
fication, Chiuso and Picci (2005) use the predictor form. p−1
L̄p = BK A K BK ... AK BK (8a)
The subspace identification problem is: given a set of

T
input/output measurements, estimate the system matrices (A, zp (k) = zTk−1 zTk−2 ... zTk−p (8b)
B, C, D), Kalman filter gain K up to within a similarity transfor-
mation, and the innovation covariance matrix Re . One can substitute (7) into (5) to obtain
There are also minor differences among these model forms. p
yf (k) = Γ̄f L̄p zp (k) + Γ̄f AK xk−p + Ḡf zf −1 (k)
The process form and the innovation form use the process (A,
B, C, D) matrices, while the predictor form uses the (AK , BK , + Df uf (k) + ef (k) (9)
1504 S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513

It is clear that the product of the observability and controllability which is the well-known high-order ARX (HOARX) model used
matrices, in the asymptotic methods (Ljung, 1999).
⎡ p−1 ⎤
CBK CAK BK . . . CAK BK 2.3. Linear regression and projections
⎢ p ⎥
⎢ CAK BK CA2K BK . . . CAK BK ⎥
⎢ ⎥
H̄fp  Γ̄f L̄p = ⎢ .. .. .. ⎥ We introduce the notation for linear regression and projec-
⎢ ... ⎥
⎣ . . . ⎦ tions used in this paper. Given the input vector x(k) and output
f −1 f f +p−2 vector y(k), a linear relation
CAK BK CAK BK ... CAK BK
(10) y(k) = Θx(k) + v(k)

is the Hankel matrix which contains the predictor Markov can be built by collecting data for input and output variables and
parameters. Ḡf also contains the predictor Markov parameters. forming the data matrices

y(1) y(2) . . . y(N) = Θ x(1) x(2) . . . x(N) + V


2.2. Assumptions    
Y X

To establish the foundation of the SIM, we introduce follow- where V is the matrix of noise.
ing assumptions: By minimizing
J = ||Y − ΘX||2F ,
A1: The eigenvalues of A − KC are strictly inside the unit circle.
where ||·||F is the F-norm, we have the least squares solution
A2: The system is minimal in the sense that (A, C) is observable
−1
and (A, [B, K]) is controllable. Θ̂ = YXT (XXT )
A3: The innovation sequence ek is a stationary, zero mean, white
noise process with second order moment The model prediction is
−1
Ŷ = Θ̂X = YXT (XXT ) X
E(ei eTj ) = Re δij
Defining
where δij is the Kronecker delta. ΠX = XT (XXT )
−1
X
A4: The input uk and innovation sequence ej are uncorrelated
for open-loop data, but uk is directly related to past inno- as the projection matrix to the row space of X, then
vation ek for closed-loop data. −1
Ŷ = YXT (XXT ) X = YΠX
A5: The input signal is quasi-stationary (Ljung, 1999) and is
persistently exciting of order f + p, where f and p stand for is a projection of Y on X. The least square residual is
future and past horizons, respectively, to be defined later. ⊥
Ỹ = Y − Ŷ = Y (I − ΠX ) = YΠX
From these assumptions we can relate the state space model where
forms to more traditional input-output models. For example, the ⊥ −1
ΠX = I − ΠX = I − XT (XXT ) X
innovation form (3) can be converted to the following input-
output model, is the projection to the orthogonal complement of X. It is easy
to verify that the model Ŷ and the residual Ỹ are orthogonal.
yk = [C(qI − A)−1 B + D]uk + [C(qI − A)−1 K + I]ek (11) Furthermore,
−1
from which the Box-Jenkins model or the ARMAX model can XΠX = XXT (XXT ) X=X
be recovered. Equivalently, the noise term in the Box-Jenkins
model plays the role of innovation in the Kalman filter. The ⊥ −1
XΠX = X(I − XT (XXT ) X) = X − X = 0
predictor form (4) can be converted to
For a model with two sets of input X and U with noise V
yk = C(qI − AK )−1 BK zk + Duk + ek (12)  
X
Y = ΓX + HU + V = [ Γ H ] +V (14)
Since AK is strictly stable based on Assumption A1, U

we can find [ Γ H ] by least squares, assuming V is independent
(qI − AK )−1 = AiK q−i of both regressors X and U.
i=1 If we are only interested in estimating Γ , using the fact that
can be truncated to a large number p and (12) reduces to V is independent of U, that is
p 1 1 T
VU T = [ v(1), . . . , v(N) ] [ u(1), ..., u(N) ]
yk =
˙ CAiK BK zk−i + Duk + ek (13) N N
i=1 → 0 as N → ∞
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1505

we have where the subscript f denotes future horizon, respectively. The


−1 −1
YΠU⊥ = V (I − U T (UU T ) U) = V − VU T (UU T ) U extended observability matrix is
  −1 ⎡ ⎤
1 T 1 T
C
=V− VU UU U → V as N → ∞ ⎢ CA ⎥
N N ⎢ ⎥
Γf = ⎢⎢ .. ⎥
⎥ (17)
Therefore, ⎣ . ⎦
YΠU⊥ = (ΓX + HU + V )ΠU⊥ = ΓXΠU⊥ + V CAf −1

Γ can be found by regressing YΠU⊥ on XΠU⊥ as follows, and Hf and Gf are Toeplitz matrices:
−1 ⎡ ⎤
Γ̂ = YΠU⊥ XT (XΠU⊥ XT ) (15) D 0 ··· 0
⎢ CB D ... 0 ⎥
2 ⎢ ⎥
where the relation (ΠU⊥ ) = ΠU⊥ is used. It is straight-forward Hf = ⎢
⎢ .. .. .. .. ⎥
⎥ (18a)
to show that Γ̂ from (15) is identical to the least squares solution ⎣ . . . . ⎦
of (14). See Appendix of (van Overschee & de Moor, 1995). CAf −2 B CAf −3 B ··· D
⎡ ⎤
2.4. General SIM procedures I 0 ··· 0
⎢ CK I ... 0⎥
⎢ ⎥
Most SIMs involve some or all of the following steps: Gf = ⎢
⎢ .. .. .. .. ⎥
⎥ (18b)
⎣ . . . .⎦
(1) Step 1: Pre-estimation. In this step either the Markov param- CAf −2 K CAf −3 K ··· I
eters as in (13) (Jansson, 2003; Larimore, 2004; Shi &
MacGregor, 2001) or the innovation sequence ek (Qin & The input data are arranged in the following Hankel form:
Ljung, 2003b) is pre-estimated from a high-order ARX ⎡ ⎤
uk uk+1 · · · uk+N−1
(HOARX) model.
⎢ u uk+2 . . . uk+N ⎥
(2) Step 2: Regression or Projection. In this step a least squares ⎢ k+1 ⎥
Uf = ⎢⎢ ⎥ (19a)
regression or projection is performed to estimate one or sev- .. .. .. .. ⎥
⎣ . . . . ⎦
eral (up to f) high-order models.
(3) Step 3: Model Reduction. The high-order model identified uk+f −1 uk+f · · · uk+f +N−2
in Step 2 is reduced to an appropriate low dimensional sub-
space that is observable. This step gives the estimates of Γ f Uf = [ uf (k) uf (k + 1) ... uf (k + N − 1) ] (19b)
or the state sequence xk . Similar formulations are made for Yf and Ef . The state sequences
(4) Step 4: Parameter Estimation. The reduced observability are defined as:
matrix or the realized state sequence from Step 3 is used to
estimate the state space parameters A, B, C, D and K. Xk = [ xk , xk+1 , . . . , xk+N−1 ] (20)
(5) Step 5: Iteration. The above steps can be iterated to improve
accuracy. The Kalman state Xk is unknown, but we know that the Kalman
state is estimated from past input and output data based on (7),
Pre-estimation in Step 1 is usually designed to deal with p
Xk = L̄p Zp + AK Xk−p (21)
closed-loop identification. It is also used to enforce the trian-
gular structure of Hf and thus a causal model. Sometimes Steps where Xk−p = [ xk−p , xk−p+1 , . . . , xk−p+N−1 ]. For a
2 and 3 are done in one combined step, but they can always be p
sufficiently large p, AK  0. Hence, from (21) and (16),
written in two separate steps. Step 4 is where parametrization
takes place, which is unique up to a similarity transform. Yf = Γf L̄p Zp + Hf Uf + Gf Ef
= Hfp Zp + Hf Uf + Gf Ef (22)
3. Open-loop subspace methods
where Hfp = Γf L̄p is the product of the process observability
The early developments of SIMs are applicable to open- matrix and the predictor controllability matrix. It is analogous to
loop identification where the input data are assumed indepen- H̄fp in (10) but it is not exactly a Hankel matrix. However, it does
dent of past noise, which admits no feedback. These methods have a reduced rank n which is less than the matrix dimensions.
include N4SID, MOESP and the CVA method without the pre- Eqs. (16) and (22) can both be used to explain open-loop
estimation step. SIMs, whichever is more convenient. Under open-loop condi-
Based on the innovation form in (3), an extended state space tions, Ef is uncorrelated to Uf , that is,
model can be formulated as
1
Yf = Γf Xk + Hf Uf + Gf Ef (16) Ef UfT → 0 as N → ∞, (23)
N
1506 S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513

or & Longman, 1992) uses this approach. The MOESP algorithm


−1 (Verhaegen, 1994) uses this linear regression and performs SVD
Ef ΠU⊥f = Ef (I − UfT (Uf UfT ) Uf ) = Ef
on Ĥfp Zp ΠU⊥f .
Furthermore, Ef is uncorrelated to Zp from the Kalman filter
theory. Therefore, 3.1.3. CVA approach
Since the coefficient matrix Hfp = Γf L̄p in (25) is not full
1
Ef ZpT → 0 as N→∞ (24) rank, the exact solution to (25) should be the canonical correla-
N tion analysis (CCA) which performs SVD on
The above two relations (23) and (24) are very useful in open-
Wr Yf ΠU⊥f ΠU⊥f ZpT Wc = Wr Yf ΠU⊥f ZpT Wc  Un Sn VnT
loop SIMs.
The open-loop SIMs do not involve a pre-estimation step.
and chooses Γ̂f = Wr−1 Un Sn
1/2
for balanced realization.
Most of them involve only three major steps: projection or −1/2
regression, model reduction, and parameter estimation. We will In the above equation Wr = (Yf ΠU⊥f YfT ) , Wc =
−1/2
summarize each step in the following subsections. (Zp ΠU⊥f ZpT ) . This is exactly canonical correlation
analysis which extracts the n smallest angles between Yf ΠU⊥f
3.1. SIM projections and model reduction
and Zp ΠU⊥f .
3.1.1. N4SID
Open-loop SIMs such as the N4SID (Overschee & Moor, 3.1.4. A unified formulation
1994) first eliminate Uf by post-multiplying ΠU⊥f on (22), van Overschee and de Moor (1995) have unified several SIMs
in the open-loop case which offer insights into the relations
Yf ΠU⊥f = Hfp Zp ΠU⊥f + Hf Uf ΠU⊥f + Gf Ef ΠU⊥f among the SIMs. For the three SIM algorithms presented above,
they are all equivalent to performing SVD on
= Hfp Zp ΠU⊥f + Gf Ef (25)
W1 Ĥfp W2 = Un Sn VnT (27)
Then the noise term is removed by multiplying ZpT from the
where Ĥfp is the least squares estimate in (26) and for
result of (24),
the regression approach, W1 = I, W2 = I, for N4SID, W1 = I,
Yf ΠU⊥f ZpT = Hfp Zp ΠU⊥f ZpT + Gf Ef ZpT = Hfp Zp ΠU⊥f ZpT 1/2 1/2
W2 = (Zp ZpT ) , for MOESP, W1 = I, W2 = (Zp ΠU⊥f ZpT ) ,
−1/2 1/2
and for CVA, W1 = (Yf ΠU⊥f YfT ) , W2 = (Zp ΠU⊥f ZpT ) .
⊥ T ⊥ T −1 It is pointed out in (Gustafsson & Rao, 2002) that the weight-
Ĥfp = Yf ΠUf Zp (Zp ΠUf Zp )
ing W1 has little impact on the results and the solution to Γ f will
N4SID performs SVD on undo this weighting

Ĥfp Zp = Γ T T Γ̂f = W1−1 Un Sn1/2 .
f L̄p Zp = USV  Un Sn Vn

where Sn contains the n largest singular values, and chooses However, for finite data length W1 can make a difference since
1/2 (25) is indeed a reduced rank regression. A requirement for the
Γ̂f = Un Sn as the estimated observability matrix, which is a
weights is that W1 is nonsingular and W2 does not reduce rank
special, balanced realization. Since Γ f and L̄p are observabil-
for Hfp W2 .
ity and controllability matrices for different models, this is not
exactly a balanced realization.
3.2. Enforcing causal models
3.1.2. Regression approach
In the extended state space model (22) Hf is block-triangular,
Perform least square solution to (25) by minimizing
which makes the model causal. However, this information is
J = ||Yf ΠU⊥f − Hfp Zp ΠU⊥f ||2F , not normally taken care of in SIMs, as pointed out in (Shi &
MacGregor, 2001). While there is no problem from a consis-
 ⊥ ⊥ T ⊥ ⊥ T −1
tency point of view given proper excitation of the input, known
Ĥfp = Γ f L̄p = Yf ΠUf (ΠUf Zp )(Zp ΠUf ΠUf Zp ) parameters are estimated from data. Shi (2001) proposes an algo-
−1 rithm known as CVAHf that removes the impact of future input
= Yf ΠU⊥f ZpT (Zp ΠU⊥f ZpT ) (26) from the future output using pre-estimated the Markov parame-
ters and then performs sub-space projections. Shi (2001) further
Note that the rank of Γf L̄p should be n. In the model reduction
shows that this procedure achieves consistency. Larimore (2004)
step we perform SVD,
argues that the CVAHf was implemented in Adaptx and that it

Ĥfp = Γ T T
f L̄p = USV  Un Sn Vn
is efficient, but he does not discuss the impact of imperfect pre-
estimates.
1/2
and choose Γ̂f = Un Sn as the observability matrix. The To avoid these problems the SIM model must not include
observer-Kalman filter method (OKID) (Phan, Horta, Juang, these non-causal terms, Peternell, Scherrer, and Deistler (1996)
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1507

propose a few methods to exclude these extra terms. Specifi- (2) Perform SVD for the following weighted matrix
cally, they recommend a two-step procedure: (i) use a conven-  
tional (unconstrained) SIM to estimate the deterministic Markov W1 Γ f L̄p W2  Un Sn Vn
T
(33)
parameters CAi−1 B; and (ii) form Hf with these Markov param-
eters to ensure that it is lower triangular and then estimate the where W1 is nonsingular and L̄p W2 does not lose rank. Un ,
extended observability matrix. Qin and Ljung (2003a), Qin et al. Sn and Vn are associated to the n largest singular values. For
(2005) propose a causal subspace identification method (PAR- CVA weighting we can choose W1 = (Yf ΠUf ⊥ Y T )−1/2 and
f
SIM) which remove these non-causal terms by performing f least ⊥ ZT )
W2 = (Zp ΠUf
1/2
We choose
p
squares projections in parallel. To accomplish this we partition
the extended state space model row-wise as follows: Γ̂f = W1−1 Un Sn1/2 (34)
⎡ ⎤ ⎡ ⎤
Yf 1 Yf 1 from which the estimate of A and C can be obtained
⎢Y ⎥ ⎢Y ⎥ (Verhaegen, 1994).
⎢ f2 ⎥ ⎢ f2 ⎥
Yf = ⎢ ⎥
⎢ .. ⎥ ; Yi  ⎢ ⎥
⎢ .. ⎥ ; i = 1, 2, . . . , f (28) (3) The estimate of B and D is discussed in the end of this
⎣ . ⎦ ⎣ . ⎦ section.
Yff Yfi
4.1. Estimating A and C from f
where Yfi = [ yk+i−1 yk+i . . . yk+N+i−2 ]. Partition Uf
and Ef in a similar way to define Ufi , Ui , Efi and Ei , respectively, In the subspace identification literature A and C are extracted
for i = 1, 2, . . ., f. Denote further from Γ f by choosing f ≥ n + 1, making Γ f−1 also full column
⎡ ⎤ rank. Denoting
Γf 1
2:f
⎢Γ ⎥ Γf = Γf (ny + 1 : ny f, :)
⎢ f2 ⎥
Γf = ⎢ ⎥
⎢ .. ⎥ (29a) which is the bottom (f − 1) block rows of Γ f , we have
⎣ . ⎦
2:f
Γff Γf = Γf −1 A

Hfi  [ CAi−2 B Therefore,


... CB D] (29b)
† 2:f
 = Γ̂f −1 Γ̂f
Gfi  [ CAi−2 K ... CK I] (29c)
The estimate of C is simply
where Γ fi = CAi−1 . We have the following equations by parti- Ĉ = Γ̂f (1 : ny , :).
tioning (22),

Yfi = Γfi L̄p Zp + Hfi Ui + Gfi Ei (30) 4.2. Estimation of K


for i = 1, 2, . . ., f. Note that each of the above equations is guar- A simple method to estimate the Kalman filter gain K is to
anteed causal. Now we have the following parallel PARSIM extract it from L̄p . From the unified expression (27), we have:
algorithm.
ˆ W = U S VT
W1 Γ̂f L̄p 2 n n n
4. Parallel PARSIM (PARSIM-P) 1/2
and W1 Γ̂f is chosen a Un Sn ,
(1) Perform the following LS estimates, for i = 1, 2, . . ., f, ˆ W = S 1/2 V T
L̄p 2 n n
 †
  Zp Therefore,

Γfi L̄p Ĥfi = Yfi (31)
Ui ˆ = S 1/2 V T W −1
L̄ p n n 2

where [·]† is the Moore-Penrose pseudo-inverse. Stack Note that L̄p is the extended controllability matrix of the predic-
 tor. Similar to the extraction of  and Ĉ from Γ̂f , we can extract
Γfi L̄p , together to obtain Γ̂f L̄p as
ÂK and [ B̂K K̂ ].
⎡ ⎤ Another approach to estimating K is to extract it from Gf (Qin
Γ
f 1 L̄p
⎢ ⎥ et al., 2005).
⎢ ⎥
⎢ Γf 2 L̄p ⎥ From (22) we have
⎢ ⎥ 
⎢ . ⎥ = Γf L̄p (32)
Y f Π
⊥  = Gf E f Π 
⊥  = G f Ef
⎢ .. ⎥ (35)
⎣ ⎦ Zp Zp

Γff L̄p Uf Uf
1508 S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513

since Ef is not correlated with Zp and Uf in open-loop. Perform- The process output can be represented as
ing QR decomposition,
⎡ ⎤ ⎡ ⎤⎡ ⎤ yk = C(qI − AK )−1 x0 + [C(qI − AK )−1 BK + D]uk
Zp R11 Q1
⎢ ⎥ ⎢ ⎥⎢ ⎥ + C(qI − AK )−1 Kyk + ek (45)
⎣ Uf ⎦ = ⎣ R21 R22 ⎦ ⎣ Q2 ⎦ (36)
Yf R31 R32 R33 Q3 or

then yk = [I − C(qI − AK )−1 K]yk

R33 Q3 = Gf Ef (37) = C(qI − AK )−1 x0 + [C(qI − AK )−1 BK + D]uk + ek


(46)
Denoting ek = Fe∗k such that cov(e∗k ) = I, from Assumption A3
we have FFT = Re . Using this notation we have using ek = Fe∗k where e∗k has an identity covariance matrix, and
defining
Gf Ef = G∗f Ef∗ (38)
ỹk = F −1 [I − C(qI − AK )−1 K]yk (47a)
where
⎡ ⎤ G(q) = F −1 C(qI − AK )−1 (47b)
F 0 ... 0
⎢ CKF F ... 0⎥ D∗ = F −1 D (47c)
⎢ ⎥
G∗f =⎢

⎥ ∈ ny f ×ny f

..
.
..
.
..
. ... ⎥
⎦ we obtain,
CAf −2 KF CAf −3 KF ··· F ỹk = G(q)BK uk + D∗ uk + G(q)x0 δk + e∗k
From Eqs. (37) and (38) and using the fact that Q3 is an orthonor- = G(q) ⊗ uTk vec(BK ) + Iny ⊗ uTk vec(D∗ ) + G(q)x0 δk + e∗k
mal matrix, we choose
(48)
Êf∗ = Q3 (39a)
where vec(BK ) and vec(D* ) are vectorized BK and D* matrices
Ĝ∗f = R33 (39b) along the rows. δk is the Kronecker delta function. Now vec(BK ),
vec(D* ) and x0 can be estimated using least squares from the
Denoting the first block column of G∗f by G∗f 1 , above equation. The B, D matrices can be backed out as:
⎡ ⎤ D̂ = F D̂∗ (49a)
F
⎢ ⎥   
⎢ CKF ⎥ Iny 0 F B̂ = B̂K + KD̂ (49b)
G∗f 1 =⎢
⎢ .. ⎥=
⎥ (40)
⎣ . ⎦ 0 Γ̂f −1 KF
4.4. Estimating all parameters from the state
CAf −2 KF
An alternative approach is to estimate all model parameters
KF and F can be estimated as ˆ we have
   † from the state sequence. With the estimate of L̄ p
F̂ Iny 0 ˆ z (k)
= G∗f 1 (41) x̂k  L̄p p

KF 0 Γ̂f −1
From (3) we obtain
Finally, ⎧   2 ⎫
⎨ N 
x̂k  ⎬
 
)F̂ −1
K̂ = (KF (42) [ Ĉ D̂ ] = arg min yk − [ C D]  êk
⎩  uk  ⎭
k=1
and  
x̂k
R̂e = F̂ F̂ T (43) = yk − [ Ĉ D̂ ]
ûk

4.3. Determination of B, D ⎧  ⎡ ⎤2 ⎫


⎪  x̂k 
⎨ N   ⎪⎬
 ⎢ ⎥
Qin et al. (2005) give an optimal approach to estimate B [ Â B̂ K̂ ] = arg min xk − [ A B K ] ⎣ uk ⎦
and D and the initial state using A, C, K and F for the general ⎪
⎩ k=1   ⎪⎭
 êk 
innovation form. Since the initial state is estimated this step does
not introduce a bias for finite p. For more detail see (Ljung & McKelvey, 1996; Overschee &
From the innovation form of the system we have: Moor, 1994), and (Chiuso & Picci, 2005).
As one can see from the above illustration, two major paths
xk+1 = AK xk + BK uk + Kyk (44) for open-loop SIMs are to use the estimates of Γ f or xk to further
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1509

estimate the model parameters. However, no results are available a high-order ARX model to decouple the correlation between Uf
as to which path leads to a better model. and Ef . The well-known CVA algorithm proposed by Larimore
(1990) actually pre-estimates Hf using a high-order ARX and
5. Closed-loop SIMs then move Ĥf Uf to the LHS of (22). Shi and MacGregor (2001)
also use this technique.
In order to identify a state space model with closed-loop data, Inspired from the SSARX approach, Chiuso and Picci (2005)
a couple of closed-loop subspace identification methods (SIMs) give a variation known as the whitening filter approach (WFA)
have been proposed in the last decade (Ljung & McKelvey, 1996; that uses the predictor model form and carry out multi-stage
van Overschee & de Moor, 1997; Verhaegen, 1993). More recent projections row by row. In each block row projection causality is
work is presented in (Jansson, 2003; Qin & Ljung, 2003b), strictly enforced, similar to (Qin et al., 2005). No pre-estimation
which has been regarded as a significant advance in subspace is involved but the projections have to be done block-row wise
identification of feedback systems (Chiuso & Picci, 2005). The to decouple noise from control input. In the rest of this section
consistency of the algorithms has been investigated in (Chiuso we briefly introduce these closed-loop SIMs.
& Picci, 2005; Lin, Qin, & Ljung, 2004).
Due to the feedback control the future input is correlated with 5.1. Innovation estimation method
past output measurement or past noise, making the traditional
SIMs biased. That is, the last two terms of (22) are correlated for Partitioning the last term of (30) into two parts, we obtain
closed-loop systems. Therefore, most of the closed-loop SIMs
try to decouple these two terms. The SIMPCA methods proposed Yfi = Γfi L̄p Zp + Hfi Ui + G−
fi Ei−1 + Efi (50)
in (Wang & Qin, 2002) and a later modification in (Huang, Ding,
& Qin, 2005) move Hf Uf to the LHS and use principal compo- where
nent analysis on the joint input/output data simultaneously. The G−
fi = [ CA
i−2 K ... CK ].
observer/Kalman filter identification (OKID) algorithm (Phan
et al., 1992), which is not traditionally known as SIMs, does not For i = 1, (50) becomes,
use an extended future horizon, therefore is free from the bias
problem. These are some of the closed-loop SIMs which do not Yf 1 = CL̄P ZP + DU1 + Ef 1 (51)
require special manipulations. which is a high-order ARX model. Typically D = 0 in (51).
Most closed-loop SIMs involve four or five of the steps out- Hence, (51) is suitable for closed-loop data since Ef1 is always
lined in Section II-D. Based on the notation in Section II-A, we uncorrelated of past data Zp . In the case that D = 0, there must be
have four different approaches to estimate the model parameters: a delay in the feedback loop, making U1 uncorrelated with Ef1 .
As a consequence, we can obtain unbiased estimates of CL̄P ,
(1) Estimate the Markov parameters from a high-order ARX D, and Ef1 from (51) using closed-loop data.
(HOARX) model, form the Hankel matrix Hfp , then perform The innovation estimation method proposed in (Lin, Qin, &
SVD on Hfp to estimate AK , BK and C. (OKID, Phan et al., Ljung, 2006) (IEM1) uses (51) to pre-estimate Êf 1 , then form
1992); Êi−1 by using the shift structure of Ei−1 , and replace Ei−1 in (50)
(2) Estimate the Markov parameters from a high-order ARX using Êf 1 to estimate Γfi L̄p . Since the only error term in (50) is
model, form Ḡf , then estimate Γ̄f L̄p from (9) and perform Efi which is “future” relative to Ui , it is suitable for closed-loop
SVD to estimate AK , BK and C (SSARX, Jansson, 2003; data.
CVA, Larimore, 2004); and The innovation estimation method proposed in (Lin et al.,
(3) Partition (9) row-wise into f separate sub-problems, enforce 2004; Qin & Ljung, 2003b) (IEM) involves estimating innova-
causal relations similar to (Qin & Ljung, 2003a), estimate tion sequence repeatedly row-wise and estimating Γ f through a
Γ̄f L̄p (or L̄p zp (k) as the state vector), and then estimate A, weighted singular value decomposition. A, B, C, D and K can
B, C and D. (WFA, Chiuso & Picci, 2004; Chiuso & Picci, also be obtained as illustrated in the previous section.
2005).
(4) Pre-estimate the innovation Ef from a HOARX and use (22)
5.2. SIMs with pre-estimation
to estimate the state space model (Qin & Ljung, 2003b).
For convenience we assume D = 0 to simplify the presenta-
Since (22) is actually composed of f block rows in each term and p
tion. Suppose that p is chosen large enough so that AK  0, (9)
the first block row gives an estimate of the innovation, Qin and
can be written as
Ljung (2003b) propose an innovation estimation method (IEM)
that partitions (22) into f block rows and uses the estimated yf (k) = Γ̄f L̄p zp (k) + Ḡf zf −1 (k) + ef (k) (52)
innovation from previous block rows to further estimate model
parameters of the next block row sequentially. An alternative Due to feedback ef (k) is correlated with zf−1 (k). Since Ḡf con-
method known as IEM1 (Lin et al., 2004) estimates the innova- tains the Markov parameters of the predictor form, Jansson
tion from the first block row and then treats êk as known to esti- (2003), Shi and MacGregor (2001) and Larimore (2004) pre-
mate other model parameters. The SSARX approach proposed estimate Ḡf (or part of Ḡf that is related to uf (k)) from a
in (Jansson, 2003) uses the predictor form (4) and pre-estimates high-order ARX model (13). Then, the estimate Ḡf is used to
1510 S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513

define a new vector


ˆ z
ỹf (k) = yf (k) − Ḡ f f −1 (k) = Γ̄f L̄p zp (k) + ef (k)

Now the error term ef (k) is uncorrelated with past data zp (k),
making it suitable for closed-loop data. The model coefficient
Γ̄f L̄p can be estimated using least squares and then SVD or
weighted SVD is performed to obtain Γ̄ˆ f . Alternatively, one
can perform CCA between ỹ(k) and zp (k) to obtain Γ̄ˆ f and L̄
ˆ
p
in one step, leading to the CVA approach in (Larimore, 2004).

5.3. Whitening filter approach

Chiuso and Picci (2005) observe that one does not need to
pre-estimate Ḡf if the triangle structure of Ḡf is exploited.
Partitioning (52) row-wise and denoting
Γ̄fi = CAi−1
K , Fig. 1. Closed-loop SIMs comparison.

Ḡfi = [ CAi−2
K BK CAi−3
K BK ... CBK ],
It is interesting to compare the innovation estimation method
T
and the whitening filter approach. They all partition the extended
zi−1 (k) = [ zTk zTk+1 ... zTk+i−2 ] , state space row-wise and utilize a multi-stage least squares
the ith row of (52) is method to estimate system matrices. The innovation estimation
method starts from a state space model in innovations form,
yk+i−1 = Γ̄fi L̄p zp (k) + Ḡfi zi−1 (k) + ek+i−1 for while the whitening filter approach is based on a state space
model in predictor form.
i = 1, 2, . . . , f. (53)
The IEM, CVA and SIMPCA use the process A matrix to form
Using least squares Γ̄fi L̄p can be estimated for i = 1, 2, . . . f, the observability matrix, while the WFA, OKID, and SSARX

which then form Γ̄f L̄p . Two subsequent options can be used in
use the predictor matrix AK . For open-loop unstable systems the
the model reduction step similar to the open-loop SIM proce- whitening filter approach can be numerically advantageous, as
dure in the previous section. The first one is to perform SVD demonstrated in (Chiuso & Picci, 2005). However, for bounded
  systems such as stable or integrating systems, this advantage
or weighted SVD on Γ̄ f L̄p to obtain Γ̄f , then estimate model disappears. For limited data length where K is time varying, it
 . The other option is to form
parameters from Γ̄ f is better to use process A matrix.
The major difference between closed-loop SIMs and open-
Γ̄f Xk = Γ̄f L̄p Zp  Γ̄f
L̄p Zp loop SIMs is in estimating the observability subspace Γ f or Γ̄f .
and perform SVD to obtain the state sequence Xk , from which The remaining steps to estimating model parameters are essen-
the process A, B, C, D, and K are estimated (Chiuso & Picci, tially the same.
2005).
6. Simulation example
5.4. Summary of closed-loop SIMs
To demonstrate how SIM works in closed-loop case, we use
Subspace identification methods are difficult to apply to the example in (Verhaegen, 1993). The model of the plant is
closed-loop data because of the use of an extended future hori- given in transfer function form:
zon that introduces correlation between inputs and past noise.
To avoid this correlation several methods such as CVA and 10−3 (0.95q4 + 12.99q3 + 18.59q2 + 3.30q − 0.02)
(54)
SSARX use pre-estimation to separate these two terms. The q5 − 4.4q4 + 8.09q3 − 7.83q2 + 4q − 0.86
SIMPCA algorithm avoids the correlation by using the parity
space instead of the observability subspace. Interestingly, the The output disturbance of the plant is a zero-mean white noise
extended future horizon is not a necessary requirement for the with standard deviation 1/3 filtered by the linear filter
projection or regression step of SIMs. It is only necessary to 0.01(2.89q2 + 11.13q + 2.74)
extend the order of the Hankel matrix, from which the observ- F1 (q) =
q3 − 2.7q2 + 2.61q − 0.9
ability matrix is reduced. The OKID (Phan et al., 1992) and the
SMARX (Ljung & McKelvey, 1996) do not require the extended The controller is
future horizon for the regression step. See (Qin & Ljung, 2006)
for more discussions. The closed-loop SIMs can be summarized (0.61q4 − 2.03q3 + 2.76q2 − 1.83q + 0.49)
F (q) =
in Fig. 1. q4 − 2.65q3 + 3.11q2 − 1.75q + 0.39
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1511

Fig. 2. The eigenvalues of estimated A matrix using IEM: (×) estimated pole, Fig. 4. The eigenvalues of estimated A matrix using SSARX: (×) estimated
(+) system pole. pole, (+) system pole.

The feedback mechanism is


uk = −F (q)yk + rk
where rk is a zero-mean white noise sequence with standard devi-
ation 1. We take the number of data points j = 1200, and generate
100 data set, each time with the same reference input rk but with
different noise sequence ek . We choose f = p = 20 for “innova-
tion estimation” approaches, and f = p = 30 for “whitening filter”
approaches. In our simulation, we observe that to obtain unbi-
ased estimation the “whitening filter” approach needs larger f
and p than the “innovation estimation” approach.
The pole estimation results for the closed-loop experiments
are shown in Figs. 2–5. From the results we can see that all

Fig. 5. The eigenvalues of estimated A matrix using WFA: (×) estimated pole,
(+) system pole.

the methods can provide consistent estimates. The “whitening


filter” approach produces the worst results, but there is no general
statement we can make from this specific example. The SSARX
has one pair of outlying poles, but this could happen to other
methods due to the actual noise based on our experience.

7. Further discussions and conclusions

7.1. Statistical properties

The statistical properties such as consistency and efficiency


Fig. 3. The eigenvalues of estimated A matrix using IEM1: (×) estimated pole, of SIMs have been investigated recently (Bauer, 2003, 2005;
(+) system pole. Bauer & Ljung, 2002; Chiuso & Picci, 2004; Gustafsson, 2002;
1512 S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513

Jansson & Wahlberg, 1998; Knudsen, 2001; Larimore, 1996). 5. Confidence intervals. It is desirable to be able to estimate the
Consistency is concerned with the bias of the estimates while confidence intervals for the estimated models. This is also
efficiency is concerned with variance. All these variants are true for SIMs. A possible approach is to derive the model
shown to be generically consistent. For some special cases, it confidence intervals based on the variance estimates of the
has also been shown that CVA gives statistical efficiency and/or model parameters (Chiuso & Picci, 2005).
gives the lowest variance among available weighting choices. 6. Disturbance models: use or do not use? It is generally true that
Simulations also seem to indicate that CVA may have better vari- correlated disturbances happen to the process to be identified
ance properties in overall comparisons, see, e.g. (Ljung, 2003). even during the data collection phase. Therefore, it is usually
While most SIMs are consistent, few if any can achieve the a good idea to identify the process model and the distur-
efficiency of the maximum likelihood estimate (MLE). For open- bance model. However, most industrial practice does not use
loop SIMs Bauer and Ljung (2002) show that SIMs with the CVA the identified disturbance model. One rationale behind this is
type of weighting are optimal for the special case of white noise that the disturbance characteristics change very often. How-
input perturbation and f → ∞. They also show that the variance ever, without using a disturbance model, the power of Kalman
of the estimates improves as f increases. For closed-loop SIMs filtering is ignored. An important issue is to decide whether
only variance expressions of the estimates are available (Chiuso the disturbance model is representative for most of the distur-
& Picci, 2005). bance scenarios, that is, whether the process is “fully” excited
in the disturbance channel.
7.2. Practical considerations 7. Model quality assessment. It is important to assess the model
quality both during the identification phase and during on-
The popularity of SIM in industry has increased tremen- line use. Due to the time-varying nature of industrial pro-
dously in recent years (Larimore, 2004; Zhao & Harmse, 2006). cesses, on-line model assessment is necessary to determining
One of the reasons behind the rapid adoption of SIMs in practice whether model re-identification is needed. The assessment
is the simplicity of SIMs and the inherent characteristics of mul- task includes process model assessment and disturbance
tivariable industrial control problems, such as model predictive model assessment.
control (MPC) problems. While significant progress has been
made in the analysis and understanding of SIMs, the following 7.3. Conclusions
issues are still standing to some extent.
Subspace identification methods have enjoyed rapid develop-
1. Optimal input design for SIMs. Since SIMs are related to ment for both closed-loop systems and open-loop processes. The
high-order ARX, low-order input excitations such as sinu- attractive features include simple parametrization for MIMO
soidal signals are not very suitable for SIMs. Most SIMs systems and robust noniterative numerical solutions. These fea-
achieve favorable results when the input is white or close tures lead to their rapid adoption in industry. There are, however,
to white. For industrial applications closed-loop testing and many unsolved issues in both statistical analysis and practical
simultaneous multivariable testing are preferred (Zhu & Van considerations. Future research should be focused on further
Den Bosch, 2000). understanding of the statistical properties and resolving the prac-
2. Connection to asymptotic methods. Since SIMs are closely tical issues.
related to HOARX with model reduction using state space
models, it is natural to probe the connection to the asymptotic
Acknowledgements
methods, which has enjoyed surprising success in industry
(Zhu, 1998). The two types of methods essentially perform
Financial support from Natural Science Foundation under
the same first step, which is HOARX, see (Qin & Ljung,
CTS-9985074, National Science Foundation of China under
2006). The only difference is in the model reduction step:
an Overseas Young Investigator Award (60228001), and the
SIMs perform model reduction in time domain, while the
Texas-Wisconsin Modeling and Control Consortium are grate-
asymptotic methods do it in frequency domain.
fully acknowledged.
3. Time delay and order estimation. To improve the accuracy
of the final model one must estimate the time delay. This is a
trivial task for SISO models, but can be difficult for MIMO References
processes. The order estimation is also an important problem.
4. Zero model responses. For MIMO processes an input-output Bauer, D., & Ljung, L. (2002). Some facts about the choice of the weighting
matrices in larimore type of subspace algorithms. Automatica, 38, 763–
pair can have zero responses, while the overall system is inter- 773.
acting. For this case it is desirable to identify input-output Bauer, D. (2003). Subspace methods. In Proceedings of the 13th IFAC SYSID
channels that have zero responses and keep them zero in the Symposium (pp. 763–773).
model parametrization. While this is easily done for input- Bauer, D. (2005). Asymptotic properties of subspace estimators. Automatica,
output models, such as FIR models in DMC practice, it is 41, 359–376.
Chiuso, A., & Picci, G. (2004). The asymptotic variance of subspace estimates.
not trivial for state space models. Including parameters in J. Econometrics, 118, 257–291.
the model that are known to be zero usually increases the Chiuso, A., & Picci, G. (2005). Consistency analysis of some closed-loop sub-
variance of the model estimates. space identification methods. Automatica, 41, 377–391.
S.J. Qin / Computers and Chemical Engineering 30 (2006) 1502–1513 1513

Chou, C., & Verhaegen, M. (1997). Subspace algorithms for the identification Peternell, K., Scherrer, W., & Deistler, M. (1996). Statistical analysis of novel
of multivariable dynamic errors-in-variables models. Automatica, 33(10), subspace identification methods. Signal Processing, 52, 161–177.
1857–1869. Phan, M., Horta, L. G., Juang, J. -N., & Longman, R. W. (1992). Improvement
Forssell, U., & Ljung, L. (1999). Closed-loop identification revisited. Automat- of observer/kalman filter identification (OKID) by residual whitening, In
ica, 35, 1215–1241. AIAA Guidance, Navigation and Control Conference, Hilton Head, South
Gustafsson, T. (2002). Subspace-based system identification: Weighting and Carolina.
pre-filtering of instruments. Automatica, 38, 433–443. Qin, S. J., & Ljung, L. (2003a) Parallel QR implementation of subspace identi-
Gustafsson, T., & Rao, B. D. (2002). Statistical analysis of subspace-based fication with parsimonious models, In IFAC Symposium on System Identifi-
estimation of reduced-rank linear regression. IEEE Transactions on Signal cation.
Processing, 50, 151–159. Qin, S. J., & Ljung, L. (2003b). Closed-loop subspace identification with inno-
Huang, B., Ding, S. X., & Qin, S. J. (2005). Closed-loop subspace identifica- vation estimation. In Proceedings of the 13th IFAC SYSID Symposium (pp.
tion: an orthogonal projection approach. Journal of Process Control, 15, 887–892).
53–66. Qin, S. J., Lin, W., & Ljung, L. (2005), A novel subspace identification approach
Jansson, M. (2003). Subspace identification and ARX modelling. In Proceedings with enforced causal models, Automatica, 41, 2043–2053.
of the 13th IFAC SYSID Symposium. Qin, S. J., & Ljung, L. (2006). On the role of future horizon in closed-loop
Jansson, M., & Wahlberg, B. (1996). A linear regression approach to state-space subspace identification. Accepted by the 14th IFAC Symposium on System
subspace system. Signal Processing, 52, 103–129. Identification.
Jansson, M., & Wahlberg, B. (1998). On consistency of subspace methods for Shi, R. (2001). Subspace identification methods for dynamic process modeling,
system identification. Automatica, 34(12), 1507–1519. Ph.D. dissertation, MacMaster University.
Katayama, T. (2005). Subspace methods for system identification. Springer. Shi, R., & MacGregor, J. F. (2001). A framework for subspace identification
Knudsen, T. (2001). Consistency analysis of subspace identification methods methods. In Proceeding of ACC (pp. 3678–3683).
based on a linear regression approach. Automatica, 37, 81–89. Soderstrom, T., & Stoica, P. (1989). System identification. London: Prentice-
Larimore, W. E. (1990). Canonical variate analysis in identification, filtering Hall.
and adaptive control. In Proceedings of the 29th Conference on Decision van Overschee, P., & de Moor, B. (1994). N4SID: Subspace algorithms for the
and Control (pp. 596–604). identification of combined deterministic-stochastic systems. Automatica, 30,
Larimore, W. E. (1996). Statistical optimality and canonical variate analysis 75–93.
system identification. Signal Processing, 52, 131–144. van Overschee, P. & de Moor, B. (1997). Closed-loop subspace system identi-
Larimore, W. E. (2004). Large sample efficiency for adaptx subspace system fication. In Proceedings of the 36th Conference on Decision and Control,
identification with unknown feedback, In Proc. IFAC DYCOPS’04. 1997. pp. 1848–1853.
Lin, W., Qin, S. J., & Ljung, L. (2004). On consistency of closed-loop sub- Verhaegen, M. (1993). Application of a subspace model identification tech-
space identification with innovation estimation. In 43rd IEEE Conference nique to identify LTI systems operating on closed-loop. Automatica, 29,
on Decision and Control (pp. 2195–2200). 1027–1040.
Lin, W., Qin, S. J., & Ljung, L. (2006). A framework for closed-loop subspace Verhaegen, M. (1994). Identification of the deterministic part of MIMO state
identification with innovation estimation. Revised for Automatica. space models given in innovations form from input-output data. Automatica,
Ljung, L. (1999). System identification: Theory for the user. Englewood Cliffs, 30(1), 61–74.
New Jersey: Prentice-Hall, Inc. Verhaegen, M., & Dewilde, P. (1992). Subspace model identification, part i: The
Ljung, L. (2003). Aspects and experiences of user choices in subspace identi- output-error state-space model identification class of algorithms. Interna-
fication methods, In 13th IFAC Symposium on System Identification, 2003. tional Journal of Control, 56, 1187–1210.
pp. 1802–1807. Wang, J., & Qin, S. J. (2002). A new subspace identification approach based on
Ljung, L., & McKelvey, T. (1996). Subspace identification. Signal Processing, principal component analysis. Journal of Process Control, 12, 841–855.
52, 209–215. Zhao, H., & Harmse, M. (2006). Subspace identification in industrial APC
van Overschee, P., & de Moor, B. (1995). A unifying theorem for three application-a review of recent process and industrial experience. Accepted
subspace system identification algorithms. Automatica, 31(12), 1853– by the 14th IFAC Symposium on System Identification, 2006.
1864. Zhu, Y. (1998). Multivariable process identification for mpc: the asymptotic
Overschee, P. V., & Moor, B. D. (1994). N4SID: Subspace algorithms for the method and its applications. Journal of Process Control, 8, 101–115.
identification of combined deterministic-stochastic systems. Automatica, Zhu, Y., & Van Den Bosch, P. P. J. (2000). Optimal closed-loop identification
30(1), 75. test design for internal model control. Automatica, 36, 1237–1241.

You might also like