Professional Documents
Culture Documents
DISSERTATION
By
*****
2007
Sliding mode controls with time-varying sliding surfaces are proposed to solve the
optimal control problem for both linear and nonlinear systems. The time-varying
sliding surfaces are designed such that the system state is on the sliding surface from
the beginning of the motion without reaching phase. Therefore, the behavior of the
optimal control problem is also transformed into finding the optimal sliding surfaces.
In some cases, the new problem is easy to solve than the original one. The main
advantage of this kind of controls is that they can provide a more robust optimal
control to the original problem. The optimal sliding mode control system should work
in such a manner. In the region dominated by the system nominal part, the system
becomes dominant, sliding mode control will take over the main control task. Several
approaches are applied to find the optimal solution or an approximation of the optimal
solution for linear continuous-time, linear discrete-time and nonlinear optimal control
problems.
est in both theoretical research and real applications. For underactuated systems,
ii
which don’t satisfy Brockett’s necessary conditions, non-smooth sliding mode con-
trol could be used to robustly stabilize those underactuated systems. Two kinds of
underactuated systems, those in the cascaded form and those in the chained form,
are studied in the second part of the dissertation. Two sliding mode controllers are
iii
This is dedicated to my family . . .
iv
ACKNOWLEDGMENTS
for his guidance and continuous support during my Ph.D. study at The Ohio State
University. Especially, I want to thank him for providing me the precious opportuni-
ties to participate in the first and second DARPA Grand Challenge. I would also like
to thank Dr. Vadmin Utkin and Dr. Andrea Serrani for serving in my dissertation
committee.
I want to thank Dr. Keith Redmill, Dr. Qi Chen, Dr. Lu Xu, Dr. Zhijun Tang,
Dr. Yiting Liu, Yongjie Zhu, Mingfeng (Alex) Hsieh and everyone else who I was
working with during the past five years. It was my pleasure to work with them.
I would like to acknowledge the financial support from the Intelligent Transporta-
tion System Fellowship program, the center for automotive research, the department
Finally, I would like to thank all my friends and my family for their help and
support.
v
VITA
PUBLICATIONS
Research Publications
R. Xu and Ü. Özgüner, “Extremum seeking control via sliding mode with two sliding
surfaces,” Proceedings of the 16th IFAC World Congress, pp. 2212-2217, Prague,
Czech, July 2005.
R. Xu and Ü. Özgüner, “Optimal sliding mode control for linear systems,” Proceedings
of the 9th International Workshop on Variable Structure Systems, pp. 143-148, June,
2006.
vi
R. Xu and Ü. Özgüner, “Sliding mode control of a class of underactuated systems,”
Automatica, vol. 44, no. 2, February, 2008.
FIELDS OF STUDY
vii
TABLE OF CONTENTS
Page
Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ii
Dedication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv
Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
Vita . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Chapters:
1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.2 An overview of classical sliding mode control for linear systems . . 17
2.3 Chattering phenomenon and boundary layer control . . . . . . . . . 19
2.4 Sliding mode control with time-varying sliding mode manifold . . . 22
2.5 Sliding mode control for infinite-time optimal quadratic regulators . 24
2.5.1 Sliding mode manifold design . . . . . . . . . . . . . . . . . 24
2.5.2 Seeking the optimal controller parameters . . . . . . . . . . 26
2.5.3 Illustrative example . . . . . . . . . . . . . . . . . . . . . . 33
viii
2.6 Sliding mode control for finite-time optimal quadratic regulators . . 38
2.7 Sliding mode control for minimum-energy control . . . . . . . . . . 40
2.7.1 Minimum-energy control . . . . . . . . . . . . . . . . . . . . 40
2.7.2 Illustrative example . . . . . . . . . . . . . . . . . . . . . . 44
ix
5.3 Sliding mode control of underactuated systems in the chained form 132
5.3.1 System model . . . . . . . . . . . . . . . . . . . . . . . . . . 132
5.3.2 Sliding mode control design . . . . . . . . . . . . . . . . . . 135
5.3.3 Stabilization of a Dubin’s car . . . . . . . . . . . . . . . . . 138
5.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
6. Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
x
LIST OF TABLES
Table Page
2.2 I1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
2.3 I2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
xi
LIST OF FIGURES
Figure Page
2.9 Time responses of cross range yE (t) and cross range velocity η(t) . . . 49
xii
3.5 Response of s(t) using HJB equation approach . . . . . . . . . . . . . 74
xiii
4.11 |sk | and δk (xk ) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
5.12 Time evolution of the forward velocity control v1 (t) and of the steering
velocity control v2 (t). . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
5.14 Time evolution of the two switching functions s1 (t) and s2 (t). . . . . 140
xiv
CHAPTER 1
INTRODUCTION
The concept of sliding mode control (SMC) has received much attention in the
past few decades. Sliding mode control is a technique in which an appropriate input is
provided so that the system states are confined to a desired sub-manifold of the state
space. The concept of sliding mode control was proposed by Utkin [17], who showed
that sliding mode could be achieved by changing the controller structure. Therefore,
sliding mode control is based on variable structure systems, in which the control input
is switched between two control signals. The system state trajectory is forced to move
along a chosen manifold in the state space, called the sliding manifold, by using an
appropriate variable structure control signal. The sliding manifold itself guarantees
the system stability, once the control restrictions are achieved. These restrictions
define the existence or not of the sliding mode. The behavior of the closed-loop
system is thus governed by the dynamics of the system on the sliding manifold [19],
[21]. The main advantage of this technique is that, once the system state reaches the
sliding manifold and sliding mode happens, the system dynamics remain insensitive to
sliding mode control is simpler to implement, since only two input control values are
1
required. Unlike similarly implemented, standard on-off control, system stability is
guaranteed by the sliding manifold. Sliding mode control is considered as a high gain
control, which provides a fast and robust response to the system behavior. The basic
ideas of sliding mode control can be illustrated by the following simple example. For
a second-order continuous system and output error e, we have the switching function
s to be defined as
d
s = ce + e = ce + ė (1.1)
dt
in a first-order system. Its solution is e(t) = e(t0 ) exp[−c(t − t0 )], which implies that,
for c > 0, the system’s output error e(t) tends exponentially to zero. The condition
sṡ ≤ 0 (1.2)
u = −M sgn(s) (1.3)
The analysis of (1.2) results in the limit values of M > 0. With large enough M ,
the switching control (1.3) can ensure the existence of the sliding mode even in the
presence of matched uncertainties and model errors. After the sliding mode happens,
the system behavior is determined only by the structure of the sliding manifold s =
2
1.1 Literature in sliding mode control
In the works of early years, the plant considered was a linear second-order system
modelled in phase variable form. Since then, intensive studies have been carried out
on the theory and applications of sliding mode control for various systems, including
with input and/or state delays [11], and stochastic systems [1]. In addition, the ob-
jectives of sliding mode control have been greatly extended from stabilization to other
control functions. Sliding mode control approaches can also be used to construct state
observers [19]. To analyze the sliding motion for systems affine in the control input,
technique [17]. It has been shown that the traditional discontinuous control switches
in sliding mode so as to imitate the equivalent control in the average sense, which
can be broadly defined as the continuous control that would lead to the invariance
conditions for the sliding motion. According to the equivalent control methodology,
the slow component of the discontinuous control is actually the equivalent control and
this component can be extracted by passing the discontinuous control through a low
pass filter, whose time constant is sufficiently large to filter out the high frequency
switching terms, yet sufficiently small not to eliminate any slow component [19]. The
equivalent control idea has also been used in the design of disturbance estimators
Classical sliding mode control is based on the possibility of making and keeping the
first time derivative of s, and switching between high amplitude opposite values with
3
theoretically infinite frequency. Nevertheless, the implementation of sliding mode
effort usually needed to assure robustness properties, and the possibility that the so-
called chattering phenomenon can rise due to the limit frequency of the control input
devices [19]. In particular the latter may lead to large undesired oscillations that can
damage the controlled system. In order to avoid chattering effect, several continuous
which the switching function s is enforced into a ∆ boundary layer of the sliding
manifold s = 0. The side effect is the control accuracy is degraded, because the
robustness of sliding mode control is lost inside the boundary layer. Moreover, the
influence of the uncertainties and disturbance grows with the size ∆ of the boundary
layer. Another way to eliminate the chattering effect is by means of high order
acting on the higher order time derivative of the switching function s instead of acting
on its first time derivative, as it happens in classical sliding mode. Preserving the
main advantages of the original approach with respect to robustness and easiness of
implementation, they totally remove the chattering effect, because the discontinuity
happens only on the first or higher order time derivative of the control input u.
Due to the use of computers for control purpose, the concept of digital sliding
mode control (DSMC) has also been a topic of study in the past two decades. In the
case of DSMC design, the control input is applied only at certain sampling instants
and the control effort is constant over the entire sampling period. Using an Euler
approximation of the time derivative over the sampling period T , Milosavljevic [12]
4
proposed the concept of quasi-sliding mode by a discrete-time extension of continuous-
time sliding mode, i.e., the condition for sliding mode is given as
to obtain ueq
k , while the other employs [4]
It should be noted that (1.5) implies (1.6), however, the reverse is not true. Since
(1.6) does not guarantee sk = 0, the DSMC based on (1.6) should be more carefully
Researchers have worked on the idea of robust optimal control in a system using
the concept of variable structure control. As mentioned before, the main advantage
of this technique is that, once the system state reaches a sliding surface, the system
However, robustness of sliding mode control is guaranteed only after the system state
reaches the sliding surface, and therefore robustness is not guaranteed during the
the advantage of sliding mode control, namely the desired dynamic behavior of the
system, is not obtained for some time after the beginning of motion. As tracking error
convergence may merely be asymptotic, the advantage might not even be achieved
at all. Furthermore, for given initial conditions, there is trade-off between a short
5
reaching phase and fast system response in the sliding phase [13]. The more robust
the system is owing to a short reaching phase, the slower the response in the sliding
phase is. Usually, a conventional SMC is conservatively designed for the worst cases in
which the system uncertainties and external disturbances are dominant. Stability or
convergence is the main concern of SMC design in such circumstance. However, most
real physical systems are not totally unknown to us, and they can be represented by
nominal systems with a small portion of uncertainties. When the system nominal part
and other performance requirements like minimizing input energy should be taken
into consideration. On the other hand, it is well known that optimal control provides
a systematic design methodology which makes the designed controller “optimal” ac-
sliding mode design, Al-abbass and Özgüner [23] used a time-varying sliding surface
ensured in the resulting sliding mode dynamics. Young and Özgüner [45], [46] further
developed this kind of time-varying sliding surfaces and constructed a sliding surface
with the co-states of the plant for both continuous-time and discrete-time systems.
was introduced by Utkin to derive an optimal sliding mode. The resulting sliding
6
mode control is introducing extra dynamics into a sliding surface. A technique for
designing the sliding surface using the linear quadratic (LQ) approach has been stud-
ied by Young and Özgüner [20]. The basic idea is that some states of the system are
considered as the control inputs to the subsystem of the other states and LQ methods
can be used to find the optimal control, or more precisely the optimal sliding mode.
A sliding mode manifold using a linear operator as s = L(x) was proposed by Young
and Özgüner [42]. Using a state space realization of the linear operator L(·), i.e.
as the intersection of linear hyper-planes in the extended state space [x, z]. Later,
Young and Özgüner [44] proposed frequency shaping sliding mode control, which is
linked with LQ optimal control and sliding mode control in the frequency domain.
They designed compensators using the optimal control and realization methods for
linear systems in the sliding mode. The sliding system with a compensator (extra
the original system. However, the designed compensators may not only improve the
stability of the sliding system but also yield desired performance and characteristics.
Concepts such as model following control [1] and integral sliding mode [2] have been
used to design control algorithms that give robust performance. Although most of
the recently proposed variable structure control algorithms employ sliding surfaces
ing optimal properties with respect to various performance indices were proposed in
[18]. Unfortunately, research reported there concerned the sliding phase only, and
no attention was paid to the system behavior in the reaching phase. Moreover, in
the optimal sliding mode design the associated quadratic cost functional only has the
7
states in its arguments without consideration of the control effort, or uses the equiva-
lent control in the sliding mode as the real control input. In order to overcome those
difficulties, Xu and Özgüner [39] presented an optimal sliding mode control approach
for linear systems and showed that integrating optimal design methodologies into the
sliding surface design can give a robust solution to the optimal control problem with
sliding phase but also in the reaching phase, and the real control input is used in the
performance index instead of the equivalent control input used in the previous results.
Later, they [41] extended the optimal sliding mode control for a class of nonlinear
systems.
1950s [61]. The time-optimal control laws (in terms of switch curves and surfaces)
were obtained for a variety of second and third-order systems in the early fifties.
Then the finite- and infinite-time linear quadratic optimal control problem has been
intensively studied for both the unconstrained [58], [49] and constrained [65], [52], [64]
cases. The nonlinear optimal control problem has also been the subject of intense
research efforts for a long time; however, it is far less mature than the research of the
linear optimal control problem. One of the main approaches to solve the nonlinear
nately, in most cases this partial differential equation cannot be solved analytically
tried to extend the existing results, which were developed in the linear optimal control
8
problem, to solve the nonlinear optimal control problem. Al’Brecht [48] first assumed
that the cost functional and system equations are both analytic, and showed that
optimal control could be obtained in the form of a power series, whose terms could
lem for the linearized system and subsequent solution of a series of linear differential
equations. He also established the convergence form of this power series for single-
input systems of the form ẋ = f (x) + Bu. Lee and Markus [60] again employed the
control law u = u∗ , which stabilizes the system ẋ = f (x, u), gives rise to a finite val-
ued cost functional, and satisfies the HJB equation near the origin. Lukes [62] later
introduced a practical algorithm for computing an optimal feedback control law for
nonlinear systems from a sequence of suboptimal control laws that converge to the
optimal control law. They solved the HJB equation recursively by using a technique
asymptotically stabilizing control law is obtained as the initial solution u(0) , the al-
gorithm will compute a sequence of u(i) with improved performance. Other recursive
algorithms to obtain the optimal control law can also be found in the literature [63].
Freeman and Kokotovic [55] employed the concept of inverse optimality, and showed
that control Lyapunov functions are solutions of the HJB equations associated with
sensible cost functionals. They used the Sontag’s formula [85] to compute the inverse
optimal control law, and pointed out that, if the level curves of the control Lyapunov
function fully agreed in shape with the value function, the Sontag’s formula gave
the real optimal control. For a special class of nonlinear systems which affine in the
9
control input (i.e., ẋ = f (x) + g(x)u) Banks and Mhana [50] introduced the so-called
state dependent Riccati equation (SDRE) approach to solve the infinite-time nonlin-
ear optimal control problem. The basic idea is to factorize f (x) to a linear-like system
as f (x) = A(x)x and solve the Riccati equation at each control instant. Huang and
Lu [57] showed that for a given infinite-time nonlinear optimal control problem there
exists at least one expansion of f (x) such that the resulting control law by means of
In the past decade there has been increasing interest in underactuated systems.
These systems are characterized by the fact that they have fewer actuators than the
applications such as free-flying space robots, underwater robots, surface vessels, ma-
nipulators with structural flexibility, etc. [87]. They are used for reducing weight,
without reducing the reachable configuration space. Some other advantages of un-
deractuated systems include no or less damage while hitting an object, and tolerance
bilize the systems around equilibrium states even locally. The key feature of many
of these problems is the nonlinear coupling between the directly actuated degrees of
freedom and the underactuated degrees of freedom. Brockett [67] has provided the
following necessary conditions for the existence of stabilizing smooth feedback laws:
10
Theorem 1.1. [67] A necessary condition for the existence of a continuously differ-
entiable asymptotically stabilizing feedback law for the system ẋ = f (x, u) is that
For some underactuated systems, these conditions may not be satisfied. If the
linearized system has uncontrollable modes associated with eigenvalues whose real
part is positive, then the original system cannot be stabilized, even locally, by any
smooth state feedback control law. However, Kawski [76] has proved that there exists
locally controllable affine systems, even if the linearized system has uncontrollable
modes associated with eigenvalues whose real part is positive. Qian et al. [82] pro-
power integrators perturbed by a C 1 lower-triangular vector field, which may not sat-
isfy Brockett’s necessary conditions. Those results suggest that continuous feedback
tion [85].
strategies have been proposed, such as back-stepping control [86], energy based or
passivity based control [70], [78], adaptive non-smooth control [75], fuzzy control [77]
11
or intelligent control [68], hybrid control [71] , etc. From those underactuated me-
great interest. One is in the cascaded normal form introduced by Olfati-saber [79] and
the other is in the chained form introduced by Murray and Sastry [91]. Olfati-saber
tems (e.g. the Acrobot [68], the TORA system [73], the VTOL aircraft [81], and the
pendubot [70]) to the cascaded normal form. Although he did not present any general
control law to stabilize underactuated systems in the cascaded normal form, the nor-
mal form itself is more convenient for control design. On the other hand, the chained
Sastry [91] have also given sufficient conditions to convert (via state feedback and
puts to the chained form. Many nonholonomic mechanical systems can be described
by kinematical models in the chained form or are feedback equivalent to chained form,
among which the most interesting examples are a Dubin’s car and a Dubin’s car with
Many researchers have put a great deal of effort on the control of a car-like vehicle
in recent years. Most of these works have been concentrated on tracking and posture
(including both the position and the orientation of a vehicle with respect to a fixed
difficult than the tracking problem due to the fact that nonholonomic systems with
more degrees of freedom than control inputs cannot be stabilized by any static state
feedback control law [67]. For nonholonomic systems, control methods based on lin-
earization techniques are not adequate and a complete nonlinear analysis is required.
12
Moreover, unlike the conventional differential type mobile robot, a car-like vehicle
has a low limit on the turning radius and this constraint makes the problem more
difficult. Therefore, many researchers tired to solve the stabilization problem via
with path-planning [90], [92] and feedback stabilizing [88], [89] laws can be found in
the literature.
Chapter 2, optimal sliding mode controls (OSMC) with time-varying sliding mode
are proposed to solve optimal control problems for linear continuous-time systems.
By applying time-varying sliding mode, the reaching phase of SMC disappears or the
system state is on the sliding surface from the beginning of the motion. Infinite-time
LQR, finite-time LQR and minimum energy control methods are used to obtain the
optimal sliding manifold, which makes the closed-loop system’s behavior “optimal”
with respect to some performance indices. The concept of Itô Integral is employed
to seek a set of optimal parameters for the optimal sliding mode controller. In the
remainder of Chapter 2, simulation results will show that the OSMC provides robust
solutions to the optimal control problems. In Chapter 3, optimal sliding mode control
is extended for a class of nonlinear systems. Three different approaches (i.e. the HJB
equation method, the CLF method and the SDRE method) are used to design the
optimal sliding manifold. The performance of each method will be compared with
rate bounded disturbance and state dependent disturbance are considered. Optimal
13
discrete-time sliding mode controls are proposed to robustly stabilize such systems
with optimal performance. In Chapter 5, sliding mode controls are designed to stabi-
lize two classes of underactuated systems in two general forms, the cascaded normal
form and the chained form. Several real application examples validate the effec-
tiveness of sliding mode control of underactuated systems. In the final chapter, the
dissertation is concluded with some concluding remarks and some possible directions
14
CHAPTER 2
2.1 Introduction
Sliding mode control (SMC) has been widely recognized as a powerful control
strategy for its ability of making a control system very robust, which yields complete
studies as well as application researches are reported ([19], [44], [21], etc.). In recent
decade, SMC has widely been extended to incorporate new techniques, such as higher-
order sliding mode control, dynamic sliding mode control and optimal sliding mode
control. These techniques retain the main advantages of SMC and also yield more
accuracy and desired performances. A technique for designing the sliding surface
using the linear quadratic (LQ) approach has been studied by Young et al. [20]. The
basic idea is that some states of the system are considered as the control inputs to the
subsystem of the other states and LQ methods can be used to find the optimal control,
or more precisely the optimal sliding mode. Young and Özgüner [44] have proposed
frequency shaping sliding mode control, which is linked with LQ optimal control and
sliding mode control in the frequency domain. They designed compensators using the
15
optimal control and realization methods for linear systems in the sliding mode. The
is a higher-order system compared with the original system. However, the designed
compensators may not only improve the stability of the sliding system but also yield
desired performance and characteristics. But these methods can only be applied to a
linear time-invariant system, and the quadratic cost function only has the states in
A typical SMC is conservatively designed for the worst cases in which the system
ity or convergence is the main concern of SMC design. However, most real physical
systems are more or less transparent to us with only a small portion of uncertain-
ties. When the system nominal part is dominant, robustness is no longer the only
concern of control design and other performance requirements should be taken into
etc. It is well known that optimal control provides a systematic design methodology
which makes the designed controller “optimal” according to the performance index.
Numerous researches have shown that the LQ optimal control problem can be solved
by using its associated Riccati equation [58]. Other optimal control problems such
systems may be solved by means of Pontryagin Principle [61]. The main limitation
this problem is to introduce SMC into optimal control whenever uncertainties are
present. This can make optimal or suboptimal control more robust and applicable
16
to systems with uncertainties. Though SMC and optimal control are two extremal
control strategies, it is still possible integrate them together. The integrated control
system should work in such a manner. In the region dominated by the system nom-
inal part, the system behavior is mainly governed by optimal control. In the region
where perturbation becomes dominant, SMC will take over the main control task.
The integrated control method is especially effective when the exogenous disturbance
is persistent (non-vanishing) nearby the equilibrium, and the system nominal part is
In this chapter, a sliding mode control with dynamic sliding surface is proposed
to solve the minimum-energy control, the infinite- and finite-time LQ optimal control
problem for linear continuous-time systems. With carefully design of the sliding
surface, the reaching phase is eliminated and the robustness of the sliding mode
control is ensured from the beginning of the motion. Several simulations results will
show that the proposed optimal sliding mode control has better performance than
With the aim of outlining the basic ideas of sliding mode control, we consider the
ẋ = Ax + Bu (2.1)
S = {x ∈ Rn : s(x) = Cx = 0} (2.2)
17
where C is chosen to make CB 6= 0 and Ae = [I − B(CB)−1 C]A a Hurwitz matrix.
And the switching control usw is designed to force the system converge to the sliding
where M > 0 is a constant number and sgn(s) is the signum function, which is defined
as
½
1 s≥0
sgn(s) = (2.5)
−1 s < 0
The existence of sliding mode can be ensured by applying sliding mode control
on the linear system (2.1). This can be showed by using a Lyapunov function candi-
V̇ = sṡ
= s[CAx + CBu]
= −sM sgn(s)
= −M |s|
≤ 0
Since V̇ < 0 when s 6= 0, the system will converge to the sliding manifold s = 0.
Moreover, the convergence time is finite, because ṡ = −M sgn(s). Once the system
18
reaches s = 0, it will stay in the sliding manifold forever. In sliding mode, s ≡ 0 and
Since Ae is a Hurwitz matrix, the system state x will converge to zero asymptotically.
Classical sliding mode control is based on the possibility of making and keeping
the first time derivative of s, and switching between high amplitude opposite values
as switching time delays and small time constants in the actuators, the discontinuity
in the feedback control produces particular dynamics in the vicinity of the sliding
surface, which is usually called as chattering [19]. This phenomenon may lead to
large undesired oscillations that degrade the performance of the system and damage
the controlled system. In order to avoid chattering effect, several continuous approx-
imations of the discontinuous sliding mode control have been presented, in which the
We also call it boundary layer control, because the control objective now is to make
the system approach to and stay in a boundary layer. In the simplest case, the control
with a high gain in the boundary layer. For example, the signum function can be
replaced by saturation functions shown in Fig. 2.1. The side effect is the control
accuracy is degraded, because the robustness of sliding mode control is lost inside the
boundary layer. Moreover, the influence of the uncertainties and disturbance grows
19
σε
1
ε
ε s
−ε
with the size of the boundary layer. Another way to eliminate the chattering effect
is by means of high order sliding mode control approaches. They are characterized
by discontinuous controls acting on the higher order time derivative of the switching
sliding mode. The actual dynamics of the controlled system in the boundary layer
second-order example
ẋ1 = x2
ẋ2 = u (2.8)
The switching function is defined as s = cx1 +x2 where c > 0. Following the procedure
20
If the signum function sgn() is approximated by a saturation function whose slope is
1/ε, the system dynamics in the boundary layer |s| < ε can be described by
ẋ1 = x2
It is worthwhile to note that the slow motion is identical to the sliding motion in the
sliding mode s = cx1 + x2 = 0. In the time scale t/ε, the fast motion is defined by
The real dynamics is approximated by superposition of the slow motion and the fast
motion
x2 = x2s + x2f
εc+M Mc h εc+M
i
= −cx10 e−c(t−t0 ) + x20 e− ε
(t−t0 )
− x1s 1 − e− ε (t−t0 ) (2.14)
εc + M
where t0 is the time instant when the system first reaches the switching surface s = 0,
and x10 and x20 are the values of x1 and x2 at t = t0 , respectively. Clearly, the
continuous approximation of the sliding mode control using saturation functions also
21
2.4 Sliding mode control with time-varying sliding mode man-
ifold
ẋ = Ax + Bu + Bξ (2.15)
where x = [x1 , · · · , xn ]T is the state vector, u the single control input, ξ the bounded
extern disturbance, |ξ| < ρ, ρ a positive constant number. A ∈ Rn×n , B ∈ Rn×1 and
(A, B) is a controllable pair. Without loss of generality A and B are assumed to have
T which makes (T AT −1 , T B) have the above forms. To design a sliding mode control
obvious that s(0) = 0 if φ(0) = −Cx(0). Therefore, the reaching time to the sliding
manifold is zero, or the system is on the sliding manifold at the beginning. A sliding
mode control is designed to keep the system on the sliding manifold s = 0, or inside a
boundary layer Ωc = {s||s| < ε} where ε > 0 is a small constant number. The latter
one is chosen in order to avoid the chattering effect in the control input. Thus, the
22
where M > ρ|CB|. σε (·) is a saturation function defined as
1 if s > ε
σε (s) = s/ε if |s| ≤ ε (2.19)
−1 if s < −ε
Theorem 2.1. The state of (2.15) stays in the boundary layer Ωc = {s||s| < ε}
from the beginning of the motion under the sliding mode control defined in (2.18) if
s(0) = 0.
Proof. A Lyapunov function candidate can be chosen as V = sT s/2. Its time deriva-
tive is
³ ´
V̇ = sṡ = s CAx + CBu + CBξ + φ̇ = s[−M σε (s) + CBξ]
because M > ρ|CB| > |CBξ|. Therefore, Ωc is an invariant set of system (2.15)
under the sliding mode control (2.18). If s(0) = 0, which locates inside Ωc , the state
of (2.15) will stay inside Ωc forever from the beginning of the motion even in the
Denote xn+1 = φ, x = [x1 , x2 , · · · , xn , xn+1 ]T and v = φ̇. After substituting the new
variables and the sliding mode control (2.20) into the original system (2.15), we obtain
ẋ = Ax + Bv + B ξ ξ (2.21)
23
where
· ¸
A − B(CB)−1 (CA + Mε C) −B(CB)−1 Mε
A=
0 0
· −1
¸ · ¸
−B(CB) B
B= Bξ =
1 0
Since the sliding mode control takes care of the disturbance, we now can focus on
designing v for the regulation or tracking problem of the above closed-loop system
control problems will be discussed in the coordinate of (x, v) after applying the sliding
A special case of the sliding mode manifold design is directly related to linear
where
Q = [qij ] ≥ 0, R>0
Similarly, the cost functional I(x, u) can be represented by using the new coordinate
(x, v) as
Z ∞
1
I(x, v) = (xT Qx + 2xT N v + Rv 2 )dt (2.23)
2 0
24
where
· ¸
Q + R(CB)−2 (AT C T + Mε C T )(CA + M
ε
C) R(CB)−2 (AT C T + Mε C T ) Mε
Q = 2 ,
R(CB)−2 (CA + M C) Mε R(CB)−2 Mε2
· T T M T ¸ε
A C + εC
N = R(CB)−2 M , R = R(CB)−2
ε
If the disturbance is not considered, then the original optimal control problem is
s.t. ẋ = Ax + Bv
This optimal control problem can be solved by solving the Riccati equation if the pair
−1 T T
φ̇ = v ∗ = R (B P − N )x (2.24)
−1 T T T
−(P B − N )R (B P − N ) − P A − A P + Q = 0 (2.25)
−1 T T
Denote K = [k1 , · · · , kn+1 ] = R (B P − N ) ∈ R1×(n+1) . The optimal choice of φ̇
can be represented by
n
X
φ̇ = ki xi + kn+1 φ (2.26)
i=1
with initial condition φ(0) = −Cx(0). The sliding mode control defined in (2.18) is
then implementable.
25
2.5.2 Seeking the optimal controller parameters
Now, we take the disturbance ξ into consideration. For simplicity, we can pick up
cn = 1 such that CB = 1. By using the optimal sliding mode control derived above,
where
0 1 0 ··· 0 0 0
0 0 1 0 0 0
.. ... .. ..
. . .
 = , B̂ =
0 0 ··· 1 0 0
a1 a2 ··· an an+1 1
k1 k2 ··· kn kn+1 0
and ai are some constant parameters. Without the disturbance ξ, this system will
ξ, however, the system states will move randomly but inside a small neighborhood
around the origin. Our goal is to find optimal values for M and/or ε which minimize
the size of the neighborhood given sufficient information of the disturbance ξ. For
the sake of simplicity, we consider the disturbance ξ(t) as a stochastic process and
number.
3. {ξ(t)} is stationary, i.e. the (joint) distribution of {ξ(t1 + t), ..., ξ(tk + t)} does
not depend on t.
26
4. The expectation function of ξ(t), E[ξ(t)] = 0 for all t.
With the above assumptions, the equations of system (2.27) can be represented by
dx1 = x2 dt
..
.
dxn−1 = xn dt (2.28)
" n+1 #
X M M
dxn = ci xi − xn + xn+1 dt + dWt (ω)
i=1
ε ε
n+1
X
dxn+1 = ci xi dt
i=1
where Wt (ω) is 1-dimensional Brownian motion starting at the origin. The solution
to (2.28) gives us (n + 1) stochastic processes: x1 (t, ω), x2 (t, ω), · · · , xn+1 (t, ω). In
order to obtain the solution to (2.28), the Itô integral is needed. In the following, the
Definition (Itô Integral) For functions f (t, ω) ∈ [0, ∞) × Ω → R, the Itô integral
is
Z tf
I[f ](ω) = f (t, ω)dWt (ω) (2.29)
t0
R tf R t1 R tf
1. t0
f dWt = t0
f dWt + t1
f dWt
R tf R tf R tf
2. t0
(cf + g)dWt = c t0
f dWt + t0
gdWt with c constant
Rt Rt
3. E[ t0f f dWt ] = 0. Especially, E[ t0f Wt dWt ] = 0
Rt Rt
4. 0
f (τ )dWτ = f (t)Wt − 0
Wτ dfτ
27
Theorem 2.2. [98] (Existence and uniqueness theorem for stochastic differ-
ential equations) Let tf > 0 and b(·, ·) : [0, tf ] × <n → <n , λ(·, ·) : [0, tf ] × <n →
|b(t, x) − b(t, y)| + |λ(t, x) − λ(t, y)| ≤ D|x − y|; x, y ∈ Rn , t ∈ [0, tf ] (2.31)
for some constant D. Let Z be a random variable such that E[|Z|2 ] < ∞. Then the
Theorem 2.4. [98] (The 1-dimensional Itô formula) Let X(t) be an Itô process
given by dXt = udt + vdWt . Let g(t, x) ∈ C 2 ([0, ∞) × R). Then Yt = g(t, Xt ) is again
∂g ∂g 1 ∂ 2g
dYt = (t, Xt )dt + (t, Xt )dXt + (t, Xt ) · (dXt )2 (2.36)
∂t ∂x 2 ∂x2
28
where (dXt )2 = (dXt ) · (dXt ) is computed according to the rules
Definition (n-dimensional Itô process) Let W (t, ω) = [w1 (t, ω), ..., wm (t, ω)]T
denote m-dimensional Brownian motion. We can form the following n Itô processes
dx1 = u1 dt + v11 dw1 + · · · + v1m dwm
.. .. (2.38)
. .
dx = u dt + v dw + · · · + v dw
n n n1 1 nm m
where
x1 u1 v11 · · · v1m dw1
.. , v = .. , dW = ..
X = ... , u = .
..
. . . (2.40)
xn un vn1 · · · vnm dwm
Theorem 2.5. [98] (The general Itô formula) Let dX(t) = udt + vdW (t) be
an n-dimensional Itô process. Let g(t, x) = [g1 (t, x), ..., gp (t, x)]T be a C 2 map from
[0, ∞) × <n into <p . Then the process Y (t, ω) = g(t, X(t)) is again an Itô process,
∂gk X ∂gk 1 X ∂ 2 gk
dYk = (t, X)dt + (t, X)dxi + (t, X)dxi dxj (2.41)
∂t i
∂xi 2 i,j ∂xi ∂xj
29
where dX = [dx1 , ..., dxn ]T , A ∈ Rn×n is time invariant, B ∈ Rn×1 is time invariant
given by
· Z t ¸
X(t) = exp(At) X(0) + exp(−Aτ )BdWτ (2.43)
0
According to the property of integration by parts, the left hand side of (2.44) is
Now, let us go back to the optimization problem. The cost functional is modified
to meet the stochastic property of the disturbance ξ by using the expectation of the
cost. Suppose that the system states have reached the origin, we want to find optimal
30
values for M and/or ε such that the average size of the neighborhood of the variation
of the states under the influence of the disturbance ξ is minimized during a long
x̄(0) = 0
where
q1 · · · 0 0
.. ..
. .
Q̂ = , qi > 0 for i = 1, · · · , n
0 qn 0
0 ··· 0 0
Applying (2.43) to the original closed-loop system with initial condition x̄(0) = 0, we
obtain
Z t
x̄(t) = exp(Â(t − τ ))B̂dWτ (2.48)
0
Denote exp(Ât) = [âij (t)], where âij (t) is the ith row and jth column entry of exp(Ât).
It should be noted that âij (t) is also a function of M and ε. Recall that
0 0 ··· 0
.. .. ..
. £ ¤ .
.
B̂ B̂ T = 0 0 ··· 0 1 0 = 0 ··· 0 0 0
1 0 ··· 0 1 0
0 0 ··· 0 0 0
31
We have
Z t Z t
E[x2i (t)] = â2in (t − τ )dτ = â2in (τ )dτ for i = 1, · · · , n (2.50)
0 0
and ε, some numerical methods have to be applied to find the optimal values for M
and/or ε.
Because the two optimization problems are coupled, we cannot obtain one solution
• Step 1: Pick up two suitable M (0) and/or ε(0) for M and/or ε respectively;
(k)
• Step 2: Compute the optimal ci for i = 1, · · · , n + 1 from (2.24) by using
• Step 3: Find the numerical minimizers M (k) and/or ε(k) from (2.51) by using
(k)
ci ;
• Step 4: If (|M (k) − M (k−1) | < ²1 and |ε(k) − ε(k−1) | < ²2 ) or (| min E[J](k) −
(k)
min E[J](k−1) | < ²3 ), then stop and give out the optimal values ci , M (k−1) ,
ε(k−1) . ²1 > 0, ²2 > 0 and ²3 > 0 are error tolerances. Otherwise, go back to
32
Name Value Name Value
c1 0 c2 1
M 18.6 k1 -10
k2 175.0455 k3 -186
ẋ1 = x2
ẋ2 = u + ξ (2.52)
where
· ¸
100 0
Q= , R=1
0 100
We pick up ε = 0.1. The optimal parameters for the optimal sliding mode control
(OSMC) uopt smc are computed from the aforementioned iterative optimization proce-
dure as tabulated in Table (2.1). For the sake of comparison, optimal state feedback
control (optimal control) uopt based on Q and R and a conventional sliding mode
control (conventional SMC) usmc are also designed. The optimal control uopt can be
33
15
10
−5
−10
−15
0 2 4 6 8 10 12 14 16 18 20
time (sec)
For the disturbance ξ, a band-limited white noise is applied, whose maximum mag-
nitude is less then 15. One realization of the white noise is shown in Fig. 2.2. Ten
realizations of white noises are generated. Each realization is used in three simu-
lations by using uopt , usmc and uopt smc , respectively. The initial conditions for the
x1 (0) = 3, x2 (0) = 0
optimal-SMC under one realization of white noise are shown in Fig. 2.3, 2.4 and
2.5, respectively. For comparison, the responses of x1 of three different controls are
put together in Fig. 2.6. From Fig. 2.6, the following observations are made:
34
x (t)
1
4
x1
0
−2
0 2 4 6 8 10 12 14 16 18 20
x2(t)
5
x2
0
−5
0 2 4 6 8 10 12 14 16 18 20
uopt(t)
20
0
uopt
−20
−40
0 2 4 6 8 10 12 14 16 18 20
time (sec)
1. Before 2 seconds, when x1 is far from the origin, the response of the optimal
SMC is very close to that of the optimal control. This means that the optimal
2. After 6 seconds, when x1 is around the origin, the variation of the response of the
optimal control is much larger than those of the other two sliding mode controls.
This is because that the optimal control is too weak to offset the influence of
the disturbance when the states are close to the origin. While the sliding mode
controls can perform more robustly because of their variable structures. This
means that the optimal SMC is more robust than the optimal control.
of cost functionals are needed. Since we can not integrate till infinity time, the cost
35
x (t)
1
4
x1
0
−2
0 2 4 6 8 10 12 14 16 18 20
x2(t)
x2 5
−5
0 2 4 6 8 10 12 14 16 18 20
usmc(t)
20
usmc
−20
0 2 4 6 8 10 12 14 16 18 20
time (sec)
Figure 2.4: The simulation results of the conventional sliding mode control
x1(t)
4
2
x1
−2
0 2 4 6 8 10 12 14 16 18 20
x2(t)
5
x2
−5
0 2 4 6 8 10 12 14 16 18 20
uopt_smc(t)
20
uopt_smc
−20
−40
0 2 4 6 8 10 12 14 16 18 20
time (sec)
Figure 2.5: The simulation results of the optimal sliding mode control
36
x1(t)
3
optimal control
sliding mode control
2.5 optimal sliding mode control
x1 1.5
0.5
−0.5
0 2 4 6 8 10 12 14 16 18 20
time (sec)
The values of cost functionals I1 and I2 of each simulation are tabulated in Table 2.2
and 2.3, respectively. From Table 2.2 and 2.3, the following observations are made:
1. For I1 , the optimal control is the best, and the conventional SMC is the worst.
However, the performance the optimal SMC is quite close to that of the opti-
mal control, while the conventional SMC is much worse. This means that the
37
Noise realization Optimal control Sliding mode control Optimal SMC
1 482.4 562.5 485.9
2 479.7 560.4 485.1
3 478.6 558.0 485.5
4 478.4 557.0 485.2
5 480.1 560.1 485.7
6 481.1 561.5 486.2
7 481.3 561.0 485.3
8 481.1 560.7 485.8
9 479.0 558.4 484.4
10 478.8 556.2 485.3
Average 480.1 559.6 485.4
Table 2.2: I1
2. For I2 , the optimal SMC is much better than the optimal control. The average
value of I2 of the optimal SMC is only 5.97% of that of the optimal control. This
means that the optimal SMC is much more robust than the optimal control.
For a finite-time linear quadratic optimal control problem where the terminal state
is free and the terminal time is fixed, the optimal solution is well known, and can be
summarized as follows. Over the time interval [t0 , tf ], the optimal control v(t) that
where
−1 T
N t ≥ 0, Q − N R N ≥ 0, R≥0
38
Noise realization Optimal control Sliding mode control Optimal SMC
1 2.074 0.1650 0.1584
2 2.312 0.1538 0.1470
3 2.909 0.1665 0.1582
4 2.272 0.1652 0.1588
5 2.189 0.1621 0.1558
6 2.683 0.1592 0.1514
7 2.924 0.1589 0.1509
8 2.356 0.1627 0.1561
9 3.000 0.1470 0.1381
10 3.046 0.1717 0.1640
Average 2.577 0.1612 0.1539
Table 2.3: I2
T T
v(t) = −R−1 [B P (t) + N ]x(t) (2.59)
−1 T T −1 T −1 T −1 T
−Ṗ = P (A − BR N ) + (A − N R B )P − P BR B P + Q − N R N (2.60)
with
P (tf ) = N t
After we obtain P (t) by backward integrating (2.60) from P (tf ), the optimal solution
the optimal sliding mode control design is similar to the optimal sliding mode control
design for the infinite-time LQR problem. By applying the sliding mode control, we
39
can transform the original finite-time LQR problem to minimize (2.58) with respect
to ẋ = Ax + Bv. Then, the optimal sliding mode manifold can be computed from
(2.59), φ̇ = v and s = Cx + φ.
In this section, we shall consider the problem of transferring the initial state
x(0) of a linear system to the origin in a fixed time Tf = tf − t0 and keeping the
energy consumption of the control input minimum. This optimal control problem is
with respect to the linear time-invariant system (2.15) with initial condition x(0) = x0
Applying the sliding mode control (2.20) to the original system (2.15), we obtain
the extended system (2.21). The minimum-energy cost functional (2.62) can also be
where
· ¸
e = (CB)−2 (AT C T + Mε C T )(CA + M
ε
C) (CB)−2 (AT C T + Mε C T ) Mε
Q 2 ,
(CB)−2 (CA + M C) M (CB)−2 Mε2
· T T M T ε¸ ε
e = (CB)−2 A C M
N
+ εC
, e = (CB)−2
R
ε
Now let us derive the necessary conditions provided by the Pontryagin Minimum
Principle (PMP) for this problem. We first form the Hamiltonian function H for this
40
problem:
1 e e v + 1 v T Rv
e + xT AT p + v T B T p
H = H[x(t), p(t), v(t), t] = xT Qx + xT N (2.64)
2 2
trol and if x∗ (t) is the resultant optimal trajectory, then there is a corresponding
1. The state x∗ (t) and the costate p∗ (t) satisfy the differential equations
∗
ẋ = Ax∗ + Bv ∗ (2.65)
ṗ∗ = −Qx e v ∗ − AT p∗
e ∗−N (2.66)
1 ∗T e ∗ e v ∗ + 1 v ∗T Rv
e ∗ + x∗T AT p∗ + v ∗T B T p∗
x Qx + x∗T N
2 2
1 ∗T e ∗ e v + 1 v T Rv
e + x∗T AT p∗ + v T B T p∗ (2.68)
≤ x Qx + x∗T N
2 2
or, equivalently,
e v ∗ + 1 v ∗T Rv
x∗T N e v + 1 v T Rv
e ∗ + v ∗T B T p∗ ≤ x∗T N e + v T B T p∗ (2.69)
2 2
e v + 1 v T Rv
ϕ(v) = x∗T N e + v T B T p∗ (2.70)
2
41
has an absolute minimum at v(t) = v ∗ (t). Since v(t) is not constrained and since
ϕ(v) is a smooth function of v(t), we can find the minimum by setting ∂ϕ(v)/∂v = 0.
But
∂ϕ e x∗ + B T p∗
e +N
= Rv (2.71)
∂v
We thus immediately obtain the optimal solution
v ∗ = −R e ∗ + B T p∗ ]
e−1 [Hx (2.72)
The next step is to substitute the optimal solution (2.72) into the canonical equations
After substituting the optimal solution (2.72), we can thus rewrite the system con-
sisting of x∗ and p∗ as
ż = W z (2.73)
Let Ψ(t, t0 ) be the (2n + 2) × (2n + 2) fundamental matrix for the system (2.73), i.e.,
42
It is necessary that x∗ (t0 ) = x0 and x∗ (tf ) = 0. Therefore, from (2.76) we find that
If the matrix Ψ12 (tf , t0 ) is nonsingular or invertible, then we can solve for p∗ (t0 ) in
p∗ (t0 ) = −Ψ−1
12 (tf , t0 )Ψ11 (tf , t0 )x0 (2.78)
n
v ∗ (t) = −Re−1 N e [Ψ11 (t, t0 ) − Ψ12 (t, t0 )Ψ−1 (tf , t0 )Ψ11 (tf , t0 )]
12
T
o
−1
+ B [Ψ21 (t, t0 ) − Ψ22 (t, t0 )Ψ12 (tf , t0 )Ψ11 (tf , t0 )] x0 (2.81)
Remark The matrix Ψ12 (tf , t0 ) is independent of the initial state. It depends on the
e N
matrices A, B, Q, e , R,
e on the initial time t0 , and on the terminal time tf .
Theorem 2.8. If an optimal control exists and if the matrix Ψ12 (tf , t0 ) is nonsingular,
Proof. If Ψ12 (tf , t0 ) is nonsingular, then (2.78) implies that there is a unique initial
costate p∗ (t0 ) corresponding to x0 , t0 and tf . Therefore, x∗ (t) and p∗ (t) are unique
and so is the control v ∗ (t). This establishes the uniqueness of the extremal control.
43
Remark If the matrix Ψ12 (tf , t0 ) is singular, then there may exist many initial
is satisfied. In this case, the extremal controls are not unique, and so there may be
within a given time interval when it is travelling at a high speed, as depicted in Fig.
θ̇r = r
where xE and yE are the longitudinal and lateral displacements of the vehicle with
respect to a fixed inertial reference frame, θr the angular displacement of the vehicle
with respect to an inertial axis normal to the plane of motion, U the longitudinal
velocity, V the lateral velocity, and r the yaw rate. The vehicle dynamics are governed
44
Lane 1 Direction of travel
Lane 2
In this model, Tf and Tr are two inputs to the system, which are the front and
rear drive/brake torques respectively. τf is the total torque applied through the
steering linkage via the power steering system. τd is the disturbance torque. Aρ is the
aerodynamic drag coefficient in the longitudinal direction. Ff and Fr are the front
and rear tire forces respectively. αf and αr are the slip angles of the front and rear
tire respectively. a and b are the distances from the front and rear ends of the vehicle
to the center of gravity. Rf and Rr are the effective rolling radii of the front and
rear tires respectively. bf is the viscous damping coefficient. m is the total vehicle
mass. Iz and Iδ are the yaw moment of inertia and the front steering system inertial
shown in Fig. 2.8. The critical system nonlinearities and uncertainties come from
the tire forces, which are determined by the slip angles. These nonlinearities can be
45
V
b a Ff
Fr
U
δ
Direction of travel
approximated as follows:
µ ¶
V − ar
αf = δ − arctan (2.85)
U
µ ¶
V + br
αr = − arctan (2.86)
U
λ 0≤λ≤4
f (λ) = 0.125λ + 3.5 4 < λ < 8 (2.87)
4.5 8≤λ
½
0.5w 0 ≤ w ≤ 4000
fw (w) = (2.88)
0.05w + 600 4000 < w
Ff (αf ) = −Cr fw (Fnf )f (|αf |)sgn(αf ) (2.89)
b
Fnf = mg (2.90)
a+b
Fr (αr ) = −Cr fw (Fnr )f (|αr |)sgn(αr ) (2.91)
a
Fnr = mg (2.92)
a+b
46
In order to simplify the problem, we assume that the yaw angle θr is small during
the maneuver, i.e., sin θr ≈ θr and cos θr ≈ 1. Let η be the lateral velocity, and the
ẏE = η (2.93)
η̇ = V̇ + U̇ θr + U r − V θr r (2.94)
After substituting the dynamics of U̇ and V̇ into the above equations, we can express
as follows:
ẏE = η (2.95)
η̇ = ay yE + aη η + bv δ + ξ(U, V, r, θr , δ) (2.96)
where ξ includes all the nonlinearities and uncertainties and can be considered as a
from a linearization process with respect to a nominal operation point such that
The numerical values of the automotive vehicle model are tabulated in Table 2.4. If
47
Name Value Name Value
m 1300.2 kg Aρ 0.36
Iz 1969 kg m2 Iδ 5 kg m2
Rf 0.33 m Rr 0.33 m
a 1.258 m b 1.612 m
h 0.05 m bf 100 Nm rad s−1
the nominal and initial longitudinal velocity is U = U (t0 ) = 30 m/s, the nominal
ay = 0, aη = −3.7471, bv = −63.389
Using this nominal linear model and following the aforementioned minimum-energy
sliding mode control design procedure in the previous subsection, the optimal sliding
mode controller for the lane change maneuver can be developed. The simulation
48
5
cross range y (m)
E
−1
−2
−3
0 0.5 1 1.5 2 2.5 3
time (sec)
Figure 2.9: Time responses of cross range yE (t) and cross range velocity η(t)
30.1
30
29.9
longitudinal velocity U (m/s)
29.8
29.7
29.6
29.5
29.4
29.3
29.2
0 0.5 1 1.5 2 2.5 3
time (sec)
49
5
4.5
3.5
cross range yE (m)
2.5
1.5
0.5
0
0 10 20 30 40 50 60 70 80 90
down range xE (m)
50
CHAPTER 3
In this chapter, the optimal sliding mode control is extended for a class of nonlinear
ẋi = xi+1 , i = 1, · · · , n − 1
where x = [x1 , · · · , xn ]T is the state vector, f a smooth function with f (0) = 0 and ξ
Because of the existence of the exogenous disturbance ξ, the sliding mode control
design is applied to obtain robustness against the disturbance around the equilibrium.
Instead of the conventional sliding mode control, an integral sliding mode control is
51
applied to gain one more degree of freedom in control design. This degree of freedom
as
where ci ’s are constant parameters, and the continuous approximated sliding mode
control is designed as
where M > sup |ξ|, φ is a smooth function and will be defined later, and σε (·) is a
Theorem 3.1. The sliding mode control (3.4) will drive the system (3.1) inside a
boundary layer Ω = {s||s| < ε} from the beginning t = 0 even in the presence of the
Pn−1
disturbance ξ, if we enforce s(0) = 0 by choosing φ(0) = − i=1 ci xi (0) − xn (0).
V̇ = sṡ
µ ¶
M
= s ξ− s
ε
M
= sξ − s2
ε
On the boundary of Ω, s = |ε|
M 2
V̇ = sξ − ε
ε
= ε(|ξ| − M )
< 0
52
Therefore, the boundary layer Ω is an invariant set of the closed-loop system of (3.1)
using the sliding mode control. Since we enforce s(0) = 0 ∈ Ω, the closed-loop system
of (3.1) will stay inside the boundary layer Ω from the beginning t = 0 even in the
Since s always stays insider Ω or |s| ≤ ε, the actually applied control input is
M
u = −c1 x2 − · · · − cn−1 xn − f − φ̇ − s (3.6)
ε
M
Denote xn+1 = φ, v = φ̇ and η = ε
, the dynamics equations of the system (3.1)
becomes
ẋ1 = x2
..
.
ẋn−1 = xn
n−1
X
ẋn = −ηc1 x1 − (ci + ηci+1 )xi+1 − ηxn+1 − v (3.7)
i=1
ẋn+1 = v
or in matrix form
ẋ = Ax + Bv (3.8)
53
Pn−1
where f = f + ηc1 x1 + i=1 (ci + ηci+1 )xi+1 + ηxn+1 . It is easy to show that (A, B)
becomes a optimal control problem for a linear system with a non-quadratic cost
functional. The solution to this optimal control problem then can be used to construct
an optimal sliding mode control for the original optimal control problem. To ensure
the existence and uniqueness of the optimal control, we should choose ci ’s such that
2
L + f > 0. In the following sections, three different approaches will be proposed to
We assume that the solution to this optimal control problem exists and its value
function is J(x) = minv I. The HJB equation for this optimal control problem is
"µ ¶ #
T
∂J ∂J 2
+ min (Ax + Bv) + L + f + 2f v + v 2 = 0 (3.10)
∂t v ∂x
∂J
By assuming that ∂t
= 0, we can obtain the optimal control as
1 ∂J
v ∗ = −f − B T (3.11)
2 ∂x
and the value function J(x) satisfies the following HJB equation
· ¸T · ¸T · ¸T · ¸
∂J ∂J 1 ∂J T ∂J
Ax − Bf + L − BB =0 (3.12)
∂x ∂x 4 ∂x ∂x
We further assume that J(·) is analytic, infinite differentiable in <n+1 and can be
P
expanded to a power (Taylor) series around the origin J(x) = ∞ i=0 Ji (x) with J0 (x) =
J1 (x) = 0. These are justified by the following requirements: The cost-to-go starting
from the origin is zero (i.e., J(0) = 0) and the control action v = −f − 12 B T ∂J
∂x
is zero
∂J
at the origin (i.e., f (0) = 0 and ∂x
(0) = 0). Ji (x) denotes i th order polynomials.
54
For example, if n = 2, J2 (x) = J200 x21 + J020 x22 + J002 x23 + J110 x1 x2 + J101 x1 x3 +
P P∞
J011 x2 x3 . Similarly, we expand f and L as f (x) = ∞
i=0 f i (x) and L(x) = i=0 Li (x),
f 0 (x) = 0.
To make (3.12) hold for any x it is necessary and sufficient that terms of (3.12) of
any order be zero. Since L0 (x) = L1 (x) = 0 and J0 (x) = J1 (x) = 0, the 0th-order and
1st-order terms of (3.12) are automatically satisfied. The 2nd order term of (3.12) is
· ¸T · ¸T · ¸T · ¸
∂J2 ∂J2 1 ∂J2 T ∂J2
Ax − Bf 1 + L2 − BB =0 (3.13)
∂x ∂x 4 ∂x ∂x
Substituting L2 and f 1 into (3.13), we can obtain J2 (x). Generally, the l th order
term of (3.12) is
· ¸T l ·
X ¸T l · ¸T · ¸
∂Jl ∂Ji 1 X ∂Ji T ∂Jl−i+2
Ax − Bf l−i+1 + Ll − BB = 0 (3.14)
∂x i=2
∂x 4 i=2 ∂x ∂x
£ ∂Jl ¤T £ ∂J2 ¤ £ ∂J2 ¤T £ ∂Jl ¤ £ ∂Jl ¤
Note that ∂x
BB T ∂x
= ∂x
BB T ∂x
. Isolating the highest order ∂x
in
Examining (3.15), we conclude that Jl (x) can be computed from Ji (x), i < l. After
we get J2 (x) by solving (3.13), we can obtain J3 (x), J4 (x), · · · step by step by solving
(3.15).
x
ė = Ae
x + Bv (3.16)
55
with a quadratic cost functional
Z ∞
2
Ie = [L2 + f 1 + 2f 1 v + v 2 ]dt (3.17)
0
2 2
This is an LQR problem. Since L + f > 0, L2 + f 1 > 0 too. Since (A, B) is
stabilizable, we conclude that the above LQR problem has a unique optimal solution
and its value function Je2 has only the quadratic terms. The optimal solution is given
as
1 ∂ Je2
ve∗ = −f 1 − B T (3.18)
2 ∂e x
where Je2 is the solution to the following HJB equation
" #T " #T " #T " #
∂ Je2 ∂ Je2 1 ∂ Je2 ∂ e2
J
Ae
x− Bf 1 + L2 − BB T =0 (3.19)
∂e x ∂e x 4 ∂e x ∂ex
Due to the uniqueness of the optimal solution, Je2 is identical to the solution to (3.13),
x
ė = Ae v∗
x + Be
à !
1 ∂ Je2
= A − Bf 1 − BB T x
e
2 ∂e x
= Âe
x (3.20)
e
where  = A − Bf 1 − 21 BB T ∂∂e
J2
x
. It should be noted that the linearized closed-loop
system x
ė = Âe
x is globally asymptotically stable.
Theorem 3.2. [62] The ith (i ≤ l) order term of this lth order approximation, Ji (x)
is the same as the ith order term of the exact solution of the HJB equation.
56
A suboptimal control then can be obtained by using the l th order approximation
function. If J l (x) > 0 and J˙l (x) < 0, then (3.22) gives a stabilizing control. Other-
wise, increase the approximation order until J l (x) is a Lyapunov function. And the
First, let us review the concept of control Lyapunov function and Sontag’s formula.
called a Control Lyapunov Function (CLF) for system (3.27) if for all x 6= 0,
∂V ∂V
G=0⇒ F <0 (3.28)
∂x ∂x
57
Freeman and Kokotovic [72] have shown that every CLF solves the HJB equation
associated with a sensible cost functional. If we have a CLF V (x) for system (3.27),
then a suboptimal control for system (3.27) with respect to the cost functional (3.26)
value function in the HJB approach to the optimal control problem. If the level curves
of a CLF V fully agree in shape with the value function, the Sontag’s formula gives
the real optimal control. For an arbitrary CLF V , though in most cases those level
curves may not fully agree, this approach based on the Sontag’s formula will result
in a suboptimal controller.
To use Sontag’s formula, we need find a CLF for the original system. Again, using
the optimal sliding mode control the nonlinear system (3.1) can be transformed to
the first n rows and first n columns of A is in a controllable canonical-like form with
state feedback and the last row of A is a zero vector. Therefore, an appropriate set
58
Note that η > 0. Obviously, if we pick up c1 > 0 and c1 6= η, then the above
Lemma 3.3. A has all distinct eigenvalues in the left half of the plane, with at most
one at the origin, there exists a constant vector H ∈ R1×(n+1) such that the Lyapunov
P.
Proof. Since all the eigenvalues of A are non-positive, in which only one eigenvalue
equal to zero and all eigenvalues of A are different, there exists an invertible matrix
Therefore,
Z ∞
T
P = etA H T HetA dt (3.35)
0
59
T
exists, and P ≥ 0 because etA H T HetA ≥ 0 for each t.
Z ∞ µ ¶ ¯∞
T d tAT T tA tAT tA ¯
PA + A P = e H He dt = e H He ¯ = −H T H
T
(3.36)
0 dt 0
P A + AT P = −H T H.
−H T H.
Proof.
T
V = xT P x + s2 /2 = xT (P + C C/2)x (3.37)
T
If P + C C/2 > 0, then V is positive definite. Noting that
0
..
∂s .
B = [ c1 · · · cn−1 1 1 ] 0 = 0 (3.38)
∂x
−1
1
and
∂s
Ax = [ −ηc1 · · · −ηcn−1 −η −η ]x = −ηs (3.39)
∂x
Thus,
∂V ∂s
B = 2xT P B + s B = 2xT P B (3.40)
∂x ∂x
∂V
Letting ∂x
B = 0, we have xT P B = B T P x = 0. Also,
∂V ∂s
Ax = xT (P A + AT P )x + s Ax = −xT H T Hx − ηs2 ≤ 0 (3.41)
∂x ∂x
60
If we can prove that −xT H T Hx − ηs2 6= 0 for any x 6= 0 under the condition B T P x =
0, then V is a CLF for system (3.8). We assume that there exists xa 6= 0, xa ∈ R(n+1)
such that −xTa H T Hxa − ηs2 (xa ) = 0, then −xTa H T Hxa = 0 and s(xa ) = 0 because
Hxa = 0 (3.42)
Cxa = 0 (3.43)
T
Remark If P > 0, then P + C C/2 > 0 is satisfied for any C.
After substituting the CLF V into the Sontag’s formula, we can obtain a subopti-
mal sliding mode control. In order to apply the Sontag’s formula, we modify the cost
functional by neglecting the cross term between the state and the control in (3.9) as
Z ∞ Z ∞
2 2
I= [L + f + v ]dt = [L + v 2 ]dt (3.44)
0 0
2
where L = L + f . Then the optimal sliding mode control is given as
( √
Vx Ax+ (Vx Ax)2 +L(Vx B)2
φ̇ = − Vx B
if Vx B 6= 0 (3.45)
0 if Vx B = 0
n−1
X
φ(0) = − ci xi (0) − xn (0) (3.46)
i=1
n−1
X
s = cxi + xn + φ (3.47)
i=1
n−1
X
u = − ci xi+1 − f − φ̇ − ηs (3.48)
i=1
∂V
where Vx = ∂x
.
61
3.5 State dependent Riccati equation approach
The state dependent Riccati equation (SDRE) approach can be used to solve the
where x(t) ∈ Rn , u(t) ∈ R, Q(x) is positive semi-definite, R(x) positive definite, Q(x),
R(x), F (x) and G(x) are all sufficiently smooth functions of the state vector x(t),
and x0 is the initial conditions of the process. Furthermore, F (0) = 0, i.e. the origin
B(x)u) be the Hamiltonian function for the problem. Then, the necessary conditions
fore, the HJB equation for the above problem can be written as
· ¸T
T 2 ∂J
x Q(x)x + R(x)u + [F (x) + G(x)u] = 0 (3.52)
∂x
Substituting the optimal control input u∗ yields The HJB equation for the above
62
where J(x) is interpreted as the value function of the infinite time nonlinear optimal
control problem (3.49). Solving the HJB equation (3.53) will lead to the optimal
most cases. In order to avoid directly solving (3.12), Banks and Mhana [50] have
introduced the so-called state dependent Riccati equation (SDRE) approach to solving
this optimal control problem. The state dependent Riccati equation (SDRE) approach
is derived from the following two expansions. First, the state vector, F (x) is expanded
as the product of a state dependent system matrix, A(x), and the state vector x
It should be noted that this expansion is generally not unique. Clearly, if A0 (x) is
e
such that f (x) = A0 (x)x, then any matrix A(x) e
such that A(x)x = 0 gives F (x) =
e
[A0 (x) + A(x)]x. The second expansion is rewriting the infinite-time value function
J(x) as
where P (x) is a symmetric, matrix valued function of the state x. The quadratic-like
form of J(x) is justified by the fact that the first order and constant terms of J(x)
63
∂pij (x)
where xT ∂P∂x(x) is a matrix with row i, column j elements defined by xT ∂x
(pij (x)
For this equation to hold for any x, it is necessary and sufficient that Π(x) satisfies
If (A(x), G(x)) is stabilizable and (A(x)x, Q(x)) is zero-detectable, for a given x, the
SDRE (3.60) will possess a unique positive definite solution, Π(x). After obtaining
the solution to (3.60), we can derive the optimal control policy from
R∞
If the cost functional has a cross term, i.e., I = 0
[xT Q(x)x+2xT N (x)u+R(x)u2 ]dt,
64
e
Lemma 3.5. opt{A(x), G(x), Q(x), e (x)} and opt{A(x), G(x), Q(x), R(x), N (x)}
R(x), N
e
have the same optimal solution, if Q(x) e (x) = N (x) + N0 (x)
= Q(x) + Q0 (x) and N
e
Proof. After substituting Q(x) e (x) into the HJB equation, we find that both
and N
optimal control problems have the same generalized SDRE as in (3.63). Therefore,
e T (x)x, their optimal solutions from
they have the same Π(x). Since N T (x)x = N
Although the computational complexity of the SDRE method is far less than
solving directly the HJB equation, Huang and Lu [57] pointed out that the success of
this method completely depends on one good choice of the state dependent coefficients
A(x). Not only can the SDRE controller provide a performance which is far off
the optimal, it may even fail to yield global stability due to the local nature of
this design technique. The right A(x) is hard to guess from the dynamics of the
plant unless the optimal cost or value function J(x) is known. In light of the above
discussion, it would appear that the SDRE approach has only succeeded in converting
the nearly impossible problem of solving the HJB equation into an equally difficult
task of determining a good A(x) function. However, for a class of nonlinear systems
as defined in (3.1), this factorization can be easily achieved by using the sliding mode
control. We assume that the nonlinear function f (x) in (3.1) can be factorized as
where K(x) = [k1 (x), · · · , kn (x)]. We further assume that L(x) in the cost functional
65
where qij (x) is the ith row, jth column element of Q(x). The actually applied control
Pn−1 ¡Pn−1 ¢
where Ĉ ∈ R1×(n+1) is a constant vector with Ĉx = i=1 ci xi+1 +η i=1 ci xi + xn+1 ,
and W (x) = [K(x), 0]. By applying this control, the original optimal control problem
is changed to
Z ∞
min [xT Q(x)x + 2xT N (x)v + v 2 ]dt (3.67)
v 0
s.t. ẋ = Ax + Bv
where
0 1 ··· 0
0 0 ··· 0
.. .. ... ..
A = . . .
−ηc1 −(c1 + ηc2 ) · · · −η
0 0 ··· 0
£ ¤T
B = 0 · · · 0 −1 1
· ¸
Q(x) 0
Q(x) = + Ĉ T Ĉ + W T (x)W (x) + Ĉ T W (x) + W T (x)Ĉ
0 0
N (x) = [Ĉ + W (x)]T
f (x) = K(x)x.
e
Proof. Suppose that there exists another factorization f (x) = K(x)x, e
where K(x) =
we conclude that A and B are both independent of the factorization f (x) = K(x)x
e
or f (x) = K(x)x. e
Suppose that we use the factorization f (x) = K(x)x instead of
66
e and N
f (x) = K(x)x. The associated Q e are
· ¸
e Q(x) 0 f T (x)W
f (x) + Ĉ T W
f (x) + W
f T (x)Ĉ
Q(x) = + Ĉ T Ĉ + W
0 0
= Q(x) + W0T (x)W0 (x) + W T (x)W0 (x) + W0T (x)W (x) + Ĉ T W0 (x) + W0T (x)Ĉ
= Q(x) + Q0 (x)
e (x) = [Ĉ + W
N f (x)]T = N (x) + W T (x)
0
W T (x)W0 (x) + W0T (x)W (x) + Ĉ T W0 (x) + W0T (x)Ĉ. Since K0 (x)x = 0 and W0 (x)x =
0, xT Q0 (x)x = 0 and xT W0T (x) = 0. By Lemma 4, the optimal solution using the
e
factorization f (x) = K(x)x is the same as the optimal solution using the factorization
Theorem 3.6 tells us that any factorization of f (x) will lead to the same optimal
control problem as the ‘right’ factorization of f (x) does. This fact gives us the freedom
For second order systems the computation of solving SDRE (3.63) can be done
symbolically to obtain a closed form expression for the optimal sliding surface. This
is generally not possible for systems of order larger than two. In these cases numerical
solution to the Riccati equation is sufficient for our approach. For single input systems
After we obtain the optimal v ∗ from the generalized SDRE (3.63), the optimal
φ̇ = v ∗ (3.68)
n−1
X
φ(0) = − ci xi (0) − xn (0) (3.69)
i=1
67
3.6 Simulation Study
θ̇ = ω
where θ is the angular displacement, ω the angular velocity, m the mass of the inverted
pendulum, J the inertia of the inverted pendulum, β the coefficient of viscous friction,
τ the control torque, d the exogenous disturbance including the wind effect. The
use a bandlimited white noise as the exogenous disturbance. One realization of the
ẋ1 = x2
mgl β
ẋ2 = sin x1 − x2 + u + ξ (3.71)
J J
Throughout this section the parameters of the inverted pendulum used to verify
the design approaches are tabulated in Table 3.1. Substituting the parameter values
into (3.71), we have the inverted pendulum model in the form of (3.1)
ẋ1 = x2
with f (x) = 10 sin x1 − 10x2 , n = 2 and |ξ| < 5. A cost functional is chosen by the
designer as
Z ∞ ¡ 2 ¢
I= x1 + 10x22 + u2 dt (3.73)
0
68
mg
l
θ
m 0.1 kg
g 10 m s−2
l 0.1 m
β 0.1 N m s rad−1
J 0.01 kg m2
69
d
0.05
0.04
0.03
0.02
0.01
−0.01
−0.02
−0.03
−0.04
−0.05
0 2 4 6 8 10 12 14 16 18 20
time (sec)
π
The initial conditions are x1 (0) = 2
and x2 (0) = 0. The switching function is designed
as
s = x1 + x2 + φ (3.74)
L = x1 + 10x22 (3.79)
70
3.6.1 The HJB equation approach
1 T
xT P Ax + xT Qx − x P BB T P x = 0 (3.84)
à ! 4
T
P A + A P 1 T
xT x + xT Qx − x P BB T P x = 0 (3.85)
2 4
Denote P = P/2
³ T
´
T T
x P A + A P + Q − P BB P x = 0 (3.86)
T
P A + A P + Q − P BB T P = 0 (3.87)
The solution to the above Lyapunov equation can be obtained by using the command
care in MATLAB.
223.1263 20.0499 0 446.2526 40.0998 0
P = 20.0499 2.2515 0 , P = 40.0998 4.5030 0 (3.88)
0 0 0 0 0 0
71
J2 = 223.1263x21 + 40.0988x1 x2 + 2.2515x22 (3.89)
J3 = 0 (3.90)
J5 = 0 (3.92)
Remark The terms involving x3 do not appear in Ji ’s. The reason is that x3 =
fact can help decreasing the computation complexity by letting the coefficients before
J 5 = J2 + J3 + J4 + J5
1 ∂J 5
v ∗ = −f − B T
2 ∂x
= −10 sin x1 − 50x1 − 41x2 − 50x3 + 20.0499x1 + 2.2515x2
The simulation results using HJB equation approach are shown in Fig. 3.3-3.6.
We choose
£ ¤
H= 1 3 1 (3.95)
72
x1(t)
1.6
1.4
1.2
0.8
x1
0.6
0.4
0.2
−0.2
0 2 4 6 8 10 12 14 16 18 20
time (sec)
x2(t)
0.1
−0.1
−0.2
−0.3
x2
−0.4
−0.5
−0.6
−0.7
−0.8
−0.9
0 2 4 6 8 10 12 14 16 18 20
time (sec)
73
s(t)
0.05
0.04
0.03
0.02
0.01
0
s
−0.01
−0.02
−0.03
−0.04
−0.05
0 2 4 6 8 10 12 14 16 18 20
time (sec)
u(t)
5
−5
u
−10
−15
−20
0 2 4 6 8 10 12 14 16 18 20
time (sec)
74
Denote P = P T = [pij ]. Through P A + AT P = −H T H, we have the following
equations:
−100p12 = −1
−50p12 − 50p23 = −1
2(p12 − 51p22 ) = −9
−100p23 = −1
p11 = 1.93, p12 = 0.01, p13 = 1.93, p22 = 0.0884, p23 = 0.01. We cannot obtain p33
from the above equations. Except for the only requirement that P should be positive
1
V = xT P x + s 2
µ 2 ¶
T 1 T
= x P+ C C x
2
= xT Px
2.43 0.51 2.43
= xT 0.51 0.5884 0.51 x (3.98)
2.43 0.51 10.50
75
x1(t)
1.6
1.4
1.2
x1 0.8
0.6
0.4
0.2
−0.2
0 2 4 6 8 10 12 14 16 18 20
time (sec)
where P = P + 12 C T C.
2
L = xT Qx + f = x21 + 10x22 + (10 sin x1 − 10x2 + 50x1 + 51x2 + 50x3 )2 (3.99)
( √
xT P Ax+ (xT P Ax)2 +L(xT P B)2
v∗ = − xT P B
if xT P B 6= 0 (3.100)
0 if xT P B = 0
The simulation results using CLF approach are shown in Fig. 3.7-3.10.
ẋ1 = x2
³x ´
1
ẋ2 = 10sinc x1 − 10x2 + u + ξ (3.101)
π
76
x2(t)
0.2
−0.2
−0.4
x2
−0.6
−0.8
−1
−1.2
0 2 4 6 8 10 12 14 16 18 20
time (sec)
s(t)
0.05
0.04
0.03
0.02
0.01
0
s
−0.01
−0.02
−0.03
−0.04
−0.05
0 2 4 6 8 10 12 14 16 18 20
time (sec)
77
u(t)
5
−5
−10
−15
u
−20
−25
−30
−35
−40
0 2 4 6 8 10 12 14 16 18 20
time (sec)
T T
Π[A − BN ] + [AT − N B T ]Π + Q − ΠBB T Π − N N = 0 (3.104)
0 1 0 0
A = −50 −51 −50 , B = −1 ,
0 0 0 1
2
1 + µ 41µ 50µ µ
Q(x) = 41µ 1691 2050 , N (x) = 41 ,
50µ 2050 2500 50
³x ´
1
µ = 50 + 10sinc
π
The simulation results using SDRE approach are shown in Fig. 3.11-3.14.
78
x1(t)
1.6
1.4
1.2
0.8
x1
0.6
0.4
0.2
−0.2
0 2 4 6 8 10 12 14 16 18 20
time (sec)
x2(t)
0.1
−0.1
−0.2
−0.3
x2
−0.4
−0.5
−0.6
−0.7
−0.8
−0.9
0 2 4 6 8 10 12 14 16 18 20
time (sec)
79
s(t)
0.05
0.04
0.03
0.02
0.01
0
s
−0.01
−0.02
−0.03
−0.04
−0.05
0 2 4 6 8 10 12 14 16 18 20
time (sec)
u(t)
5
−5
−10
u
−15
−20
−25
0 2 4 6 8 10 12 14 16 18 20
time (sec)
80
1.6
HJB
SDRE
1.4
CLF
1.2
x1 0.8
0.6
0.4
0.2
−0.2
0 2 4 6 8 10 12 14 16 18 20
time (sec)
The responses of x1 (t) of three approaches are shown in Fig. 3.15. All of them
are quite close. Especially, the response using HJB approach and the response using
SDRE approach are almost identical. This partly shows that the trajectories are
the non-vanished disturbance. The variation in the response using CLF approach is
3.7 Conclusion
All three approaches can give a robust optimal control for the inverted pendulum
system. The HJB equation approach should be able to get a more accurate solution if
we keep increasing the approximation order. Moreover, the optimal sliding switching
81
function can be designed off-line. But the computation burden of this approach is very
heavy if the number of states of the system is larger than 5. For the CLF approach,
this method is far less than the HJB equation approach. Theoretically, the SDRE
approach can obtain the exact optimal solution. However, solving a SDRE off-line
for a given system is almost impossible. In that case, numerically solving the SDRE
82
CHAPTER 4
Due to the use of computers for control purpose, discrete-time sliding mode con-
trol has been received substantial interest. As pointed out in [5], discrete-time sliding
part by means of simple equivalence. When the sliding mode control is used in a
discrete-time system (e.g., a sampled-data system), the sliding mode motion does not
exist because the switching frequency is limited by the sampling rate [12]. Thus, the
mode control will be presented for linear discrete-time systems with rate bounded dis-
This kind of optimal discrete-time sliding mode control is featured with the ability
to robustly stabilize linear discrete-time systems and the optimality with respect to
83
The system model of disturbed linear discrete-time systems is generalized as fol-
lows. Using a sampling and hold process on the continuous time systems, the nominal
xk = x(kT )
uk = u(kT )
where xk = [xk1 , xk2 , · · · , xkn ]T ∈ Rn and uk ∈ R1 are the state vector and input,
respectively. T is the sampling period, F and G are obtained from integrating the
F = exp(AT ) (4.4)
Z T
Γ = exp(Aτ )dτ (4.5)
0
G = ΓB (4.6)
It is assumed that (F, G) is controllable. Note that F is nonsingular for any nonneg-
ative value of T . The effects of the parametric uncertainties and exogenous distur-
bances can be modelled by adding an unknown disturbance term into the nominal
discrete-time systems as
The disturbance term is expressed as Gdk to guarantee the matching condition, where
dk has the same dimension as uk . It is assumed that there are two kinds of distur-
bances. The first kind of disturbance is rate bounded, i.e., |dk+1 − dk | < δd T for any
84
k ≥ 0, where δd > 0 is a known constant. The second kind of disturbance is state-
where q > 0 is a known constant. Two different optimal discrete-time sliding mode
control (ODSMC) approaches will be proposed to overcome those two kinds of dis-
turbances respectively.
sk = Cxk + φk = 0 (4.8)
where C ∈ R1×n is a constant row vector and CG 6= 0 and φk will be defined later.
We can choose φ0 = −Cx0 such that s0 = 0, i.e., the system state is on the sliding
mode manifold sk = 0 at the beginning. The ideal sliding mode should satisfy
sk+1 = sk = 0, k = 0, 1, 2, · · · (4.9)
However, because of the disturbance and the limited sampling period, the ideal sliding
mode is not obtainable for discrete-time systems, [12]. In this case, the concept of
quasi-sliding mode was proposed in [5]. Instead of requiring sk moving exactly along
the sliding mode manifold sk = 0, the quasi-sliding mode only requires that sk stays
as
85
Solving for uk , we obtain the equivalent control
and the sliding mode control uk consisting of the equivalent control uek and a switching
control usk
where
However, dk is generally unknown, so the sliding mode control using the informa-
from
Theorem 4.1. The closed-loop disturbed system (4.7) with the sliding mode control
86
in finite steps and stays inside it thereafter, if |∆k | < δd T , 0 < r < 1 and ε > |CG|δd .
Proof. First, we consider the case when sk is outside the boundary layer Ω, i.e.
|sk | > (ε + |CG|δd )T . We assume that sk > (ε + |CG|δd )T . From (4.16), we have
Therefore,
(r − 1)(ε + |CG|δd )T < rsk − (ε + |CG|δd )T < sk+1 < rsk (4.20)
Since 0 < r < 1, sk+1 will enter or approach to the boundary layer Ω. Thus, sk
will converge to the boundary layer Ω when sk > (ε + |CG|δd )T . And because r is
constant, the maximum number of the steps Nmax , which the system requires to enter
Similarly, using the property of symmetry we can prove that sk will enter the boundary
Second, we consider the case when sk is inside the boundary layer Ω, i.e. |sk | ≤
87
which also infers |sk+1 | ≤ (ε + |CG|δd )T or sk+1 is inside the boundary layer Ω. If
sk < 0, then
sk+1 is inside the boundary layer Ω too, because 0 < r < 1 and |sk | ≤ (ε + |CG|δd )T .
Therefore, the system state will stay inside the boundary layer Ω once it enters Ω.
In all, the sliding mode control law (4.15) makes the state of the closed-loop
disturbed system (4.7) enter the boundary layer Ω in finite steps and stay inside it
thereafter.
Remark The size of this boundary layer, i.e., (ε+|CG|δd )T , is O(T ) and will decrease
Remark For a given sampling period T and a fixed δd , the size of this boundary
layer is determined by |CG| because the choice of ε is also limited by ε > |CG|. G is
system parameter and not changeable, but C is free to pick up in controller design.
A good choice of C, which makes |CG| small, will lead to a small size boundary layer
and in turn good control performance. However, the side effect is that the control
effort will increase, because its inverse (CG)−1 appears in the control input uk .
Remark Using the proposed sliding mode control, we ensure that the quasi-sliding
mode can happen in finite steps. Since we can enforce s0 = 0, the quasi-sliding mode
Now we will use the optimal control method to design an optimal sliding mode
with respect to a quadratic performance index. Considering the nominal system (4.1),
88
we choose a quadratic performance index as
∞
1X T
J= (x Qxk + u2k ) (4.25)
2 k=0 k
sliding mode, we assume that the disturbance dk = 0. Since the sliding mode can
happen from the beginning by enforcing s0 = 0, sk = 0 (k ≥ 0), and the control input
becomes
where
· ¸ · ¸
F − G(CG)−1 (CF − rC) G(CG)−1 (r − 1) −G(CG)−1
F = , G=
0 1 1
89
where
· ¸
Q + (CG)−2 (CF − rC)T (CF − rC) (CG)−2 (CF − rC)T (1 − r)
Q = ,
(CG)−2 (1 − r)(CF − rC) (CG)−2 (1 − r)2
· ¸
(CG)−2 (CF − rC)T
H = , R = (CG)−2
(CG)−2 (1 − r)
If we choose C and r such that the pair (F , G) is stabilizable, then the optimal control
vk∗ is given by
where
T T T
K = (G P G + R)−1 (G P F + H ) (4.31)
−1 T T T −1 T
P = (F − GR H )T [P − P G(G P G + R)−1 G P ](F − GR H )
−1 T
+(Q − HR H ) (4.32)
sk = Cxk + φk (4.33)
ẋ1 = x2
ẋ2 = u (4.35)
90
When the sampling period T = 0.05 second, the discrete-time model is
· ¸ · ¸
1 0.05 0.00125
xk+1 = F xk + Guk = xk + uk (4.36)
0 1 0.05
So, if we choose ε > 10|CG| in the sliding mode control (4.15), the quasi-sliding mode
Controllers are designed for both T = 0.05 and T = 0.01 , respectively. In order
to explore the effect of the choice of C, two controllers using different C are designed
for T = 0.05. The controller parameters are designed by using the aforementioned
approach and are tabulated in Table 4.1. The simulation results are shown in Fig.
4.1-4.9. Comparing the results of Controller 1 and Controller 2, we observe that the
is the deviation of the state caused by the disturbance. However, the control input
the choice of C does affect the controller’s performance. Comparing the results of
Controller 1 and Controller 3, we can see that the size of the boundary layer, in
91
Controller 1 Controller 2 Controller 3
T 0.05 0.05 0.01
r 0.5 0.5 0.5
C [1, 1] [0.1, 0.1] [1 1]
ε 0.6 0.06 0.2
K [0.3535, 0.3590, 0.5] [0.0353, 0.0359, 0.5] [0.4689, 0.4700, 0.5]
which the switching function sk is confined, decreases when the sampling period T
decreases. In turn, the control accuracy of the closed-loop system is better when T
is smaller.
1
x1
0.8 x2
0.6
0.4
0.2
x
−0.2
−0.4
−0.6
−0.8
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.1: Response of the system state x using Controller 1 when T = 0.05
92
0.06
0.05
0.04
0.03
0.02
sk
0.01
−0.01
−0.02
−0.03
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.2: Response of the switching function s using Controller 1 when T = 0.05
0.5
−0.5
−1
−1.5
uk
−2
−2.5
−3
−3.5
−4
−4.5
0 1 2 3 4 5 6 7 8 9 10
time (sec)
93
1
x
1
0.8 x2
0.6
0.4
0.2
0
x
−0.2
−0.4
−0.6
−0.8
−1
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.4: Response of the system state x using Controller 2 when T = 0.05
0.01
−0.01
−0.02
−0.03
sk
−0.04
−0.05
−0.06
−0.07
−0.08
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.5: Response of the switching function s using Controller 2 when T = 0.05
94
0
−2
−4
−6
−8
uk
−10
−12
−14
−16
−18
0 1 2 3 4 5 6 7 8 9 10
time (sec)
1
x1
0.8 x2
0.6
0.4
0.2
x
−0.2
−0.4
−0.6
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.7: Response of the system state x using Controller 3 when T = 0.01
95
−3
x 10
12
10
6
sk
−2
0 1 2 3 4 5 6 7 8 9 10
time (sec)
Figure 4.8: Response of the switching function s using Controller 3 when T = 0.01
0.5
−0.5
−1
−1.5
uk
−2
−2.5
−3
−3.5
−4
−4.5
0 1 2 3 4 5 6 7 8 9 10
time (sec)
96
4.3 ODSMC of single input systems with state-dependent
disturbance
sk+1 = Cxk+1
= CF xk + CGuk
= CF xk + CG(uek + usk )
= sk + CGusk (4.44)
97
where M > 0. δk (xk ) is defined as
n
|CG| X
δk (xk ) = M |xki | (4.47)
2r i=1
Theorem 4.2. For the nominal discrete-time system (4.1), the control law (4.45)
makes the switching function sk move from any initial condition into the boundary
Proof. Considering the switching function outside the boundary layer Υ, i.e., |sk | >
sk+1 = sk + CGusk
Multiplying both sides of this equation by rsk and after a simple manipulation, we
have
98
Therefore,
or
1
Vk = s2k (4.54)
2
then
which implies that sk will move from the outside into the boundary layer Υ.
Lemma 4.3. If sk ∆sk+1 < − 12 ∆s2k+1 , then the sliding mode can happen.
1 2
Vk+1 = s
2 k+1
1
= (sk + ∆sk+1 )2
2
1 2 1
= sk + sk ∆sk+1 + s2k+1
2 2
1 2
= Vk + sk ∆sk+1 + sk+1 (4.56)
2
and sk will converge to sk = 0. Therefore, if sk ∆sk+1 < − 12 ∆s2k+1 , the sliding mode
can happen.
99
Theorem 4.4. For the nominal discrete-time system (4.1), the control law (4.45)
makes the switching function sk converge to the sliding mode manifold sk = 0 when
sk ∈ Υ.
Proof. When the switching function sk is inside the boundary layer Υ, the switching
From (4.44)
From Lemma 4.3, we conclude that the sliding mode exists and sk will converge to
Theorem 4.5. For the nominal discrete-time system (4.1), the control law (4.45)
makes the switching function sk converge to the sliding mode manifold sk = 0 from
zk = N xk (4.61)
100
to transform the system (4.1) into the following controllable canonical form
where Ĉ = CN −1 = [ĉ1 , ĉ2 , · · · , ĉn ]T . In the new coordinate, the performance index
can be represented as
∞
1 X −1 T
J = (N zk ) Q(N −1 zk )
2 k=m
∞
1 X T −1 T
= zk (N ) QN −1 zk
2 k=m
∞
1X T
= z Q̂zk (4.64)
2 k=m k
101
where z k = [zk1 , zk2 , · · · , zk(n−1) ]T ∈ Rn−1 and
0 1 0 ··· 0 0
0 0 1 ··· 0 0
... . . . ...
W = ... ∈ R(n−1)×(n−1) , V = ..
. ∈ Rn−1
0 0 0 ··· 1 0
0 0 0 ··· 0 1
zkn can be considered as a virtual control input to the system (4.66). Correspondingly,
where
· ¸
Q̂11 Q̂12
Q̂ = , Q̂11 ∈ R(n−1)×(n−1) , Q̂12 ∈ Rn−1 , Q̂22 ∈ R
Q̂T12 Q̂22
Note that Q̂11 is positive definite and Q̂22 6= 0 because Q̂ is positive definite. Since
(W, V ) is controllable, the optimal solution to minimize this performance index for
and P is a symmetric positive definite matrix which satisfies the following discrete-
P = (W − V Q̂−1 T T −1 T −1 T
22 Q̂12 )[P − P V (V P V + Q̂22 ) V P ](W − V Q̂22 Q̂12 )
102
After comparing (4.65) and (4.68), we can easily obtain the optimal switching function
as
where C ∗ = [K, 1]N . The closed-loop system on this sliding mode manifold is
xk+1 = F xk + Guk
= F xk − G(C ∗ G)−1 C ∗ F xk
Corollary 4.6. [F − G(C ∗ G)−1 C ∗ F ] is a stable matrix, whose eigenvalues are all
Now we consider the disturbed system (4.7) with state dependent disturbance
To analyze the robustness of the sliding mode controller, we discuss the case outside
Theorem 4.7. For the disturbed system (4.7), the following sliding mode control law
uk ensures the system state converge to the boundary layer Υ∗ = |s∗k | ≤ δk∗ (xk ) from
103
where M > q and δk∗ (xk ) is defined as
X n
|C ∗ G|
δk∗ (xk ) = (M + q) |xki | (4.76)
2r i=1
Proof. Considering the switching function outside the boundary layer Υ∗ , i.e., |s∗k | >
Multiplying both sides of this equation by rs∗k and after a simple manipulation, we
have
Therefore,
−r2 s∗2 ∗ ∗ 2 ∗2
k < rsk sk+1 < r sk (4.81)
104
or
which implies that s∗k will move from the outside into the boundary layer Υ∗ .
When the system state is inside the boundary layer Υ∗ , i.e., |s∗k | ≤ δk (xk ), the
105
Proof. Let Vk = xTk P xk be a candidate Lyapunov function.
< Vk (4.86)
Therefore, the closed-loop system is asymptotically stable inside the boundary layer
Υ∗ .
q
1
Remark For a given system, if q < |GT P G|
is not satisfied, we need redesign the
Theorem 4.9. For the disturbed discrete-time system (4.7), the control law (4.74)
q
1
with q < |GT P G|
makes the switching function s∗k converge to the sliding mode
106
with a state dependent disturbance dk = [q1 , q2 ]xk , where |q1 | < 0.1 and |q2 | < 0.1.
the nominal system (4.87) can be transformed into a controllable canonical form as
Then the optimal sliding mode can be designed by using the aforementioned approach
as follow
· ¸
−1 T −1 14.41 −12
Q̂ = (N ) QN =
−12 10
K = −0.8315
The dynamics of the closed-loop nominal system on the sliding mode manifold is
107
1.5
x1
x2
1
0.5
−0.5
−1
−1.5
−2
−2.5
−3
−3.5
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
time (sec)
conditions are set as x1 (0) = 1 and x2 (0) = 0. The controller parameters are picked
up as r = 0.5 and M = 0.1. The simulation results are shown in Fig. 4.10-4.12. In
Fig. 4.11, we can clearly observe that sk first converge to the boundary layer Υ, and
4.4 Conclusion
The design of optimal sliding mode control has been proposed to provide robust-
ness to linear discrete-time systems. Two kinds of disturbances, i.e., rate bounded
disturbance and state dependent disturbance, are considered respectively. The slid-
ing mode manifold is designed to obtain optimal performance with respect to some
108
4
δk(xk)
|sk|
3.5
2.5
1.5
0.5
0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
time (sec)
uk
0
−0.5
−1
−1.5
−2
−2.5
−3
−3.5
−4
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
time (sec)
109
quadratic performance indices. For systems disturbed by a rate bounded disturbance,
a time-varying sliding mode manifold is used. This time-varying sliding mode provides
additional freedom in achieving robustness and optimality. The control accuracy can
the sampling period. On the other hand, for systems disturbed by a state depen-
dent disturbance, the sliding mode manifold is designed by using Lyapunov theory to
that the optimal sliding mode control provides robust solutions to the optimal control
problems.
110
CHAPTER 5
5.1 Introduction
[79] and the other is in the chained form introduced by Murray and Sastry [91]. Olfati-
systems (e.g. the Acrobot [68], the TORA system [73], the VTOL aircraft [81], and
the pendubot [70]) to the cascaded normal form. Although he did not present any
general control law to stabilize underactuated systems in the cascaded normal form,
the normal form itself is more convenient for control design. On the other hand, the
ray and Sastry [91] have also given sufficient conditions to convert (via state feedback
inputs to the chained form. Many nonholonomic mechanical systems can be described
by kinematical models in the chained form or are feedback equivalent to chained form,
among which the most interesting examples are a Dubin’s car and a Dubin’s car with
111
multiple trailers [94]. Those two classes of underactuated systems may not satisfy
the well known Brockett’s necessary conditions (Theorem 1.1). Therefore, it is im-
possible to apply a smooth feedback law to, even locally, stabilize them. Thus, many
control laws or a mixture of both. In this chapter, sliding mode control approaches
will be developed, which can globally stabilize all degrees of freedom, including those
which are indirectly actuated through the nonlinear coupling, for those two classes of
underactuated systems. Since the sliding mode control is not smooth, the proposed
sliding mode controllers can stabilize systems, which do not satisfy Brockett’s nec-
essary conditions. The advantage of sliding mode control is its insensitivity to the
model errors, parametric uncertainties and other disturbances [19]. Once the system
is on the sliding manifold, the system behavior is determined by the structure of the
sliding surface. This advantage gives us a little more freedom in controller design, as
we can modify the system model by introducing virtual disturbances to satisfy some
conditions or requirements.
where q = [q1 , q2 ]T , q and q̇ are the states, τ the control input, and hi ’s contain Cori-
olis, centrifugal and gravity terms. In [79] and [80] a systematic method is proposed
112
to transform an underactuated system (5.1) into the following cascaded normal form
ẋ1 = x2 + d1
ẋ2 = f1 (x1 , x2 , x3 , x4 ) + d2
ẋ3 = x4 (5.2)
1, 2, 3) disturbances. The block diagram of (5.2) is depicted in Fig. 5.1. Many under-
actuated systems can be transformed into the form of (5.2). The inverted pendulum
system [78], the translational oscillations with a rotational actuator (TORA) example
[73], the vertical takeoff and landing (VTOL) aircraft [81], and the quadrotor heli-
copter [66] are all examples, to name a few. Our control objective is to stabilize all
states of (5.2) to zero. For system (5.2), we have the following assumptions:
Assumption 1. f1 (0, 0, 0, 0) = 0.
∂f1 ∂f1
Assumption 2. ∂x3
is invertible or ∂x4
is invertible.
are requirements to develop an overall sliding mode controller, and are not necessarily
satisfied by some practical systems. However, the existence of d2 gives us the freedom
of choice of f1 , which meets these assumptions. We will show this in TORA example
in Section 4.
113
d3 d2 d1
x4 x3 x2 x1
f2 + f1 + +
bu
In this subsection we will present an overall sliding mode control, which considers
and controls all states at the same time. First we denote error variables as
e1 = x1 e3 = f1 (x1 , x2 , x3 , x4 )
∂f1 ∂f1 ∂f1
e2 = x2 e4 = ∂x 1
x2 + ∂x 2
f1 + ∂x 3
x4
E1 = [e1 , e2 , e3 ]T E2 = [e1 , e2 ]T
where E1 and E2 are defined only for the system stability analysis. It is more conve-
nient to exclude e3 and/or e4 from E1 and E2 in later definitions and stability proof.
To satisfy the matched disturbance condition of the sliding mode control design [19],
Assumption 4. If d1 = 6 0, then the maximum absolute row sum norm of ∂f1 /∂x1 is
° °
° ∂f1 °
bounded, i.e., ° ∂x 1
° ≤ β1 where β1 is a nonnegative constant.
∞
Assumption 5. If d2 = 6 0, then the maximum absolute row sum norm of ∂f1 /∂x2 is
° °
° ∂f1 °
bounded, i.e., ° ∂x 2
° ≤ β2 where β2 is a nonnegative constant.
∞
∂f1
Assumption 6. If d3 6= 0 and if ∂x3
is invertible, then the maximum absolute row
° °
∂f1 ° ∂f1 °
sum norm of ∂x3
is bounded, i.e., ° ∂x 3
° ≤ β3 , where β3 is a positive constant.
∞
∂f1
Assumption 7. If d3 6= 0 and if ∂x4
is invertible, then the maximum absolute row
° °
∂f1 ° ∂f1 °
sum norm of ∂x4
is bounded, i.e., ° ∂x 4
° ≤ β4 , where β4 is a positive constant.
∞
114
∂f1
Assumption 8. The disturbances di ’s are bounded. If ∂x4
is invertible, then
∂f1
kd1 k < d1 kE2 k2 , kd2 k < d2 kE2 k2 and kd3 k < d3 + d4 kξ(x)k2 . If ∂x3
is invertible, then
kd1 k < d1 kE1 k2 , kd2 k < d2 kE1 k2 and kd3 k < d3 + d4 kξ(x)k2 . di ≥ 0 (i = 1, 2, 3, 4)
are constant numbers, and ξ(x) is a known vector field of states x = [x1 , x2 , x3 , x4 ]T .
∂f1 ∂f1
If ∂x4
= 0 and ∂x3
is invertible, then a switching surface is defined as s = c1 e1 +
∂f1
c2 e2 + c3 e3 + e4 . If ∂x4
is invertible, then a switching surface is defined as s =
The requirement for the choice of ci will be discussed later. The reason why we apply
two different surfaces is that s needs to be relative degree 1 with respect to the control
input u.
It is worthwhile to note that the systems, which are described by (5.2) and satisfy
ẋ1 = x2
ẋ3 = x4 (5.3)
ẋ4 = u
ical system studied by Rui et al. [83], where xi are the states, u the input, g the
acceleration of gravity, and l the pendulum length. It is obvious that the linearized
system of (5.3) around the origin has an unstable uncontrollable mode associated with
115
an eigenvalue whose real part is positive. Therefore, this underactuated system can-
not be stabilized by any smooth state feedback, because it violates the first Brockett’s
necessary condition. However, there exists a continuous state feedback law, which
If we add x3 to and subtract x3 from the second equation of (5.3), then system (5.3)
⇒ |x3 | ≤ |(1 + x23 )x3 | = |e3 − g sin x1 /l| ≤ |e3 | + g|e1 |/l
where In is n × n identity matrix, and denote λleft (−An1 ) as the real part of the
max{d1 , d2 , β1 d1 + β2 d2 } < λleft (−An1 ). The sliding mode control consists of two
116
parts: the equivalent control part ueq and the switching control part usw . The equiv-
Theorem 5.1. If ∂f1 /∂x4 = 0 and ∂f1 /∂x3 is invertible, with the control (5.7) all
Proof. First we prove that the sliding mode exists. A Lyapunov function candidate
yield
Therefore, under the control (5.7), the system can reach and thereafter stay on the
117
original system is reduced to
ė1 0 In 0 e1 d1
ė2 = 0 0 In e2 + d2 (5.9)
∂f1 ∂f1
ė3 −c1 In −c2 In −c3 In e3 d + ∂x2 d2
∂x1 1
or
where λmax (P ) and λmin (P ) are the largest and smallest eigenvalues of a positive
definite matrix P , respectively. If γ < λmin (Q)/2λmax (P ), then V̇1 < 0 and the
this ratio can be maximized with a choice Q = I3n (pp.342, [97]). When Q = I3n ,
λmin (Q) = 1 and λmax (P ) = 1/(2λleft (−An1 )). If γ < λleft (−An1 ), the reduced
system is asymptotically stable at the origin, and e1 , e2 and e3 will all converge to
to zero too. Recalling e3 = f1 and Assumption 3, we can easily prove that x3 and x4
Remark Placing the eigenvalues of An1 can be done by picking up appropriate ci for
A11 when n = 1 or In = 1, because the eigenvalues of An1 are just multiple (n) copies
118
Case 2: ∂f1 /∂x4 is invertible
ci should be picked up to make An2 Hurwitz and max{d1 , d2 } < λleft (−An2 ). The
equivalent control and the switching control are designed respectively as follows
· ¸−1 · ¸
∂f1 ∂f1 ∂f1 ∂f1 ∂f1
ueq = − b c1 x2 + c2 f1 + x2 + f1 + x4 + f2 (5.12)
∂x4 ∂x1 ∂x2 ∂x3 ∂x4
·¸−1
∂f1
usw =− b [M sign(s) + λs] (5.13)
∂x4
Theorem 5.2. If ∂f1 /∂x4 is invertible, with the control (5.14) all states of (5.2) will
asymptotically converge to 0.
Proof. Following the lines of the proof of Theorem 5.1, we can prove that the sliding
or
Ė2 = A2 E2 + D2 (5.16)
119
where D2 = [d1 , d2 ]T , and kD2 k2 < max{d1 , d2 }kE2 k2 . Since A2 is Hurwitz and
max{d1 , d2 } < λleft (−An2 ), the above perturbed system of E2 is asymptotically stable
Fig. 5.2. It has been constructed as a benchmark system to evaluate the performance
of various nonlinear controllers ([69], [73], [74]). Let z1 be the normalized displacement
of the platform from the equilibrium position, z2 = ż1 , θ1 be the angle of the rotor,
ż1 = z2
−z1 + εθ22 sin θ1 ε cos θ1
ż2 = 2 2
− v
1 − ε cos θ1 1 − ε2 cos2 θ1
θ̇1 = θ2
ε cos θ1 (z1 − εθ22 sin θ1 ) 1
θ̇2 = 2 2
+ v (5.17)
1 − ε cos θ1 1 − ε cos2 θ1
2
where v is the control input and ε a constant parameter which depends on the rotor
platform masses and eccentricity. All quantities are normalized in dimensionless units.
120
M
k
x1 = z1 + ε sin θ1
x2 = z2 + εθ2 cos θ1
x3 = θ1
x4 = θ2
ε cos x3 [x1 − (1 + x24 )ε sin x3 ] + u
v =
1 − ε2 cos2 x3
ẋ1 = x2
ẋ3 = x4
ẋ4 = u (5.18)
121
Parameter Value Initial condition Value
ε 0.1 z1 (0) 1m
z2 (0) 0m
θ1 (0) π
θ2 (0) 0
d3 = 0. Note that
∂f1
= ε cos x3 − 11ε 6= 0 (5.19)
∂x3
Assumption 1-3 are all satisfied. Since this system has only one disturbance d2 =
∂f1 ∂f1
11εx3 and ∂x1
== 0, the matched disturbance condition can be examined only
∂x2
¯ ¯
¯ ∂f1 ¯
for Assumption 6 and 8. It is obvious that ¯ ∂x 3
¯ < 12ε. Recalling that e1 = x1 and
⇒ 10ε|x3 | ≤ 2kek2
Thus, Assumption 6 and 8 are satisfied as well, and the aforementioned sliding mode
control approach can stabilize the TORA system. A simulation is implemented on the
original TORA system equations, and the following parameters and initial conditions
are tabulated in Table 5.1. The parameters of the sliding mode controller are tabu-
lated in Table 5.2. The simulation results are shown in Fig. 5.3 and 5.4. We can see
that the sliding mode controller successfully stabilizes the TORA system. However,
122
Name Value Name Value
M 5 c1 21.75
ρ 1 c2 22.25
λ 1 c3 8
z1(t)
1
0.5
z1 (m)
−0.5
0 1 2 3 4 5 6 7 8 9 10
θ1(t)
4
2
θ1 (rad)
−2
−4
0 1 2 3 4 5 6 7 8 9 10
time (sec)
conservative to ensure stability and in turn the control input required is larger than
A UAV quadrotor is a four rotor helicopter and depicted in Fig. 5.5. More
details of its configuration can be found in [66]. The dynamic model of the quadrotor
123
v(t)
40
30
20
10
0
v
−10
−20
−30
−40
−50
0 1 2 3 4 5 6 7 8 9 10
time (sec)
helicopter can be obtained via a Lagrange approach and a simplified model is given
as follow
θ̈ = u2 − lK4 θ̇/I1
ψ̈ = u3 − lK5 ψ̇/I2
φ̈ = u4 − K6 φ̇/I3
where (x, y, z) are three positions; (θ, ψ, φ) three Euler angles, representing pitch, roll
and yaw respectively; g the acceleration of gravity; l the half length of the helicopter;
m the total mass of the helicopter; Ii ’s the moments of inertia with respect to the
124
F4 F1
F3 F2
z
y
x
mg
axes; Ki ’s the drag coefficients; ui ’s the virtual control inputs defined as follow
u1 = (F1 + F2 + F3 + F4 )/m
u3 = l(−F1 + F2 + F3 − F4 )/I2
u4 = C(F1 − F2 + F3 − F4 )/I3
where Fi ’s are thrusts generated by four rotors and can be considered as the real
control inputs to the system, and C the force to moment scaling factor. The quadrotor
125
Since drag is very small at low speeds, the drag terms in the above equations can be
(5.22) we can easily construct a rate bounded PID controller and a sliding mode
controller to move states z and φ to their desired values zd and φd , respectively. The
where ² is a small positive constant, kz1 , kz2 , kz3 , k, k0 , k1 are all positive, and sat(·)
Then, we have
|u̇1 | ≤ k (5.27)
For φ, a sliding mode control is designed to make φ converge to its desired value φd
quickly:
and sφ = cφ (φ − φd ) + φ̇ the designed stable sliding surface for φ. The reason why
we use a rate bounded control for u1 will be discussed later. Now we will use the
tem (5.23). Since the controllers for the fully-actuated subsystem are independent of
126
the controller for the underactuated subsystem, we can design a controller with fast
response for φ, which makes φ converge to its desired value very quickly. Thus, it is
reasonable to assume that after some time φ becomes almost time invariant. Also, u1
is close to its “steady” state value g/(cos θ cos ψ) after some time
¯ ¯
¯ g ¯ η
¯ u1 − ¯≤η≤ (5.29)
¯ cos θ cos ψ ¯ | cos θ cos ψ|
where η > 0 is a small constant. Then the lower bound of |u1 | is given by
¯ ¯
¯ g−η ¯
|u1 | ≥ ¯¯ ¯ (5.30)
cos θ cos ψ ¯
Since we assume that φ is time invariant, T̂ (φ) is also time invariant and simply
u1 represents the total thrust on the body in the z-axis and is generally nonzero to
127
Check the derivatives of the new states
· ¸ µ ¶ · ¸
−1 ẋ d 1 −1 x
ẋ1 = T + T̂
ẏ dt u1 y
u̇1
= x2 − x1
u
· 1 ¸ µ ¶ · ¸
−1 ẍ d 1 −1 ẋ
ẋ2 = T + T̂
ÿ dt u1 ẏ
· ¸ · ¸
sin θ cos ψ −1 K1 /m 0 u̇1
= − T̂ T̂ x2 − x2
sin ψ 0 K2 /m u1
ẋ3 = x4
· ¸ · ¸
u2 −lK4 /I1 0
ẋ4 = + x4
u3 0 −lK5 /I2
128
Parameter Value Initial condition Value
I1 , I2 1.25 Ns2 /rad x(0) 2m
I3 2.5 Ns2 /rad y(0) 1m
K1 , K 2 , K 3 0.010 Ns/m z(0) 0m
K4 , K 5 , K 6 0.012 Ns/rad θ(0) 0
m 2 kg ψ(0) 0
l 0.2 m φ(0) 0
g 9.8 m/s2
∂f1
It should be noted that cos θ cos ψ > 0 and the determinant of ∂x3
is cos θ cos2 ψ > 0
∂f1
when (θ, ψ) ∈ (−π/2, π/2) × (−π/2, π/2), which is satisfied generally. Therefore, ∂x3
is positive definite and invertible. Moreover, the maximum absolute row sum norm
° °
∂f1 ° ∂f1 °
of ∂x3 is bounded: ° ∂x 3
° < 2. Therefore, Assumption 6 is satisfied as well. The
∞
sliding mode control proposed in Section 3 can be used to stabilize the underactuated
subsystem (5.23) after the coordinate transformation. The parameters and the initial
conditions of the quadrotor for simulation are tabulated in Table 5.3. The parameters
of the controllers for the quadrotor are tabulated in Table 5.4. Simulation results for
positions, angles and control inputs are shown in Fig. 5.6-5.9. The simulation results
129
x(t)
4
x (m)
0
−2
0 5 10 15 20 25 30 35 40 45 50
y(t)
2
y (m) 1
−1
0 5 10 15 20 25 30 35 40 45 50
z(t)
3
2
z (m)
0
0 5 10 15 20 25 30 35 40 45 50
t (sec)
show that the sliding mode controllers with the rate bounded PID controller (5.25)
are able to stabilize the quadrotor helicopter. From Fig. 5.7, we find that φ converges
to its desired value faster than other states, so it is safe to consider φ as time invariant
after 5 seconds. This verifies the assumption that T̂ (φ) is time invariant after some
time. Fig. 5.9 shows that the rate bounded control u1 successfully tracks the desired
130
θ(t)
2
θ (degree)
0
−2
−4
0 5 10 15 20 25 30 35 40 45 50
ψ(t)
ψ (degree) 1
−1
−2
0 5 10 15 20 25 30 35 40 45 50
φ(t)
100
φ (degree)
50
0
0 5 10 15 20 25 30 35 40 45 50
t (sec)
F1(t)
20
F1 (N)
10
0
0 5 10 15 20 25 30 35 40 45 50
F2(t)
20
F2 (N)
10
0
0 5 10 15 20 25 30 35 40 45 50
F3(t)
10
F3 (N)
0
0 5 10 15 20 25 30 35 40 45 50
F4(t)
10
F4 (N)
−10
0 5 10 15 20 25 30 35 40 45 50
t (sec)
131
13
u1
u1d
12
11
10
7
0 5 10 15 20 25 30 35 40 45 50
t (sec)
ẋ1 = u1
ẋ2 = u2
ẋ3 = x 2 u1
ẋ4 = x 3 u1 (5.35)
···
ẋn = xn−1 u1
132
This class of systems, which is so-called chained form, has been introduced by Murray
paper the authors have also given sufficient conditions to convert (via state feedback
inputs to the chained form. Many nonholonomic mechanical systems can be described
ẋ = v1 cos θ
ẏ = v1 sin θ
1
θ̇ = v1 tan φ (5.36)
l
φ̇ = v2
where x and y denote the location of the center of the axle between the two rear
wheels, θ the angle of the car body with respect to the x axis, φ the steering angle
with angle with respect to the car body, l the distance from the front axle to the real
axle, and v1 and v2 the forward velocity of the rear wheels and the velocity of the
steering wheels, respectively. The diagram of a Dubin’s car is shown in Figure 5.10.
As discussed in detail in [91], the Dubin’s car model (5.36) can be put into the
x1 = x
1
x2 = sec3 θ tan φ
l
x3 = tan θ
x4 = y
133
Y
L
l
φ θ
X
x
134
and input change
1
v1 = u1
cos θ
3
v2 = − sin2 φ tan θ sec θu1 + l cos2 φ cos3 θu2
l
Next, we will present a variable structure controller to stabilize the chained form
(5.35). In compatible with the Dubin’s car model, we assume the order of the chained
constant.
Based on those observations, the controller design can be divided into two stages. In
the first stage, we shall keep u1 constant or piecewise constant and converge the other
states x2 , · · · , xn to zero in finite time. In the second stage, we shall keep u2 = 0 and
where x1 (0) is the initial condition of x1 , ∆ the switching time interval, T the time
135
For u2 , a multi-layer sliding mode controller is applied to ensure the finite-time
convergence of the states. The first layer sliding surface function is defined as
½
s+
1 , if u1 = +1
s1 = (5.38)
s−
1 , if u1 = −1
where
p /q1
s+
1 = x3 + c 1 x4 1 (5.39)
p /q1
s−
1 = x3 − c1 x41 (5.40)
c1 > 0 is a positive number, q1 > p1 > 0, p1 and q1 odd numbers, and p1 and q1 are
relatively prime. The finite-time convergence can be checked as follows. Assume that
p /q1
⇒ x3 + c 1 x4 1 =0
p /q
⇒ x3 = ẋ4 = −c1 x41 1
· ½ ¾¸ q q−p
1
q1 − p1 q1 q1 −p1 1 1
⇒ x4 (t) = −c1 t + [x4 (t0 )] q1
for t > t0
q1 q1 − p 1
Since q1 > p1 > 0 and q1 and p1 are odd numbers, x4 converges to zero at t =
q1 −p1
q1
t0 + c1 (q1 −p1 )
[x4 (t0 )] q1
, so does x3 . Similarly, if u1 = −1, the same result happens.
where
s+
2 = ṡ+ + p2 /q2
1 + c2 (s1 ) (5.42)
s−
2 = ṡ− − p2 /q2
1 + c2 (s1 ) (5.43)
136
c2 > 0 is a positive number, q2 > p2 > 0, p2 and q2 odd numbers, and p2 and q2 are
finite time.
The construction of sliding surface functions continues till the highest layer sliding
surface function is relative degree 1 with respect to u2 . For n = 4, the second layer
s2 is enough. Then the sliding mode control is designed to make s2 converge to zero
in finite time as
½
u+
2 , if u1 = +1
u2 = (5.44)
u−
2 , if u1 = −1
where
" µ ¶ pq2 #
p1 −q1 p1
d p1 2
u+
2 = − c1 x4 q1 x3 + c2 x3 + c1 x4q1 − M sgn(s+
2) (5.45)
dt q1
" µ ¶ pq2 #
p1 −q1 p1
d p1 2
u−
2 = c1 x4 q1 x3 + c2 x3 − c1 x4q1 + M sgn(s−
2) (5.46)
dt q1
u̇2 , then control inputs in (5.37) and (5.44) can make the states x2 , · · · , xn converge
case, u1 ’s sign doesn’t change during the process of convergence, and u2 doesn’t switch
between u+ −
2 and u2 . However, in real applications, we don’t want x1 to keep increasing
switch when u1 changes its sign. Every time when u2 switches, the converging process
137
Name Value Name Value
∆ 2 T 4
p1 3 q1 5
p2 3 q2 5
c1 1 c2 1
M 3 k 2
stabilize x1 .
u1 = −kx1 (5.47)
+
u2 , if u1 > ε
u2 = 0, if |u1 | ≤ ε (5.48)
−
u2 , if u1 < −ε
We now use the variable structure controller proposed in the previous subsection
£ ¤ £ ¤
x(0) y(0) θ(0) φ(0) = 3 1 0 0
The controller parameters are tabulated in Table 5.5. Figure 5.11 shows the state
evolution, Figure 5.12 shows the two velocity controls, Figure 5.13 shows the x-y
trajectory of the Dubin’s car, and Figure 5.14 shows the evolution of the two switching
functions.
138
3
x(t)
y(t)
2.5 θ(t)
φ(t)
2
1.5
0.5
−0.5
−1
0 1 2 3 4 5 6 7 8
time (sec)
v1(t)
2
−2
−4
−6
0 1 2 3 4 5 6 7 8
v2(t)
10
−5
−10
0 1 2 3 4 5 6 7 8
time (sec)
Figure 5.12: Time evolution of the forward velocity control v1 (t) and of the steering
velocity control v2 (t).
139
1.2
0.8
0.6
y (m)
0.4
0.2
−0.2
0 0.5 1 1.5 2 2.5 3
x (m)
1.2
s1(t)
s2(t)
1
0.8
0.6
0.4
0.2
−0.2
0 1 2 3 4 5 6 7 8
time (sec)
Figure 5.14: Time evolution of the two switching functions s1 (t) and s2 (t).
140
5.4 Conclusion
Stabilizing two classes of underactuated systems by using sliding mode control has
been studied in this chapter. The non-smooth property of the sliding mode control has
been utilized to make it possible to stabilize a nonlinear system, which doesn’t satisfy
Brockett’s necessary conditions. The simulation results of there examples show that
the proposed sliding mode controllers work effectively to stabilize the underactuated
systems.
141
CHAPTER 6
CONCLUSIONS
systems by using sliding mode control are two topics of this dissertation.
In the first part of the dissertation, the concept of optimal sliding mode control
control, much less research has been carried out on the design of the sliding surface.
Most of the existing literature suggests that the sliding surface is designed so that
the closed-loop system is stable and has some desirable properties when confined to
it. However, not much literature is available for identifying and correlating these
presenting methods of designing sliding surfaces such that the systems satisfy some
optimality. Optimal sliding mode control for linear continuous-time systems is studied
in Chapter 2, which integrates the optimal control methodologies into the design of the
sliding surface in sliding mode control. Through this way, the property of optimality
plus robustness may be obtained for controlling disturbed systems. The proposed
optimal sliding mode controls provide robust solutions to the infinite- and finite-
142
utilizes the concept of Itô integral is presented to find a set of optimal parameters of
the designed sliding mode controller for systems disturbed by special signals. Several
simulation results qualitative and quantitatively show the advantages of the proposed
optimal sliding mode control over conventional sliding mode control and conventional
optimal control.
Next, optimal sliding mode control design is extended for a class of nonlinear
(HJB equation approach, CLF approach and SDRE approach) are presented to obtain
an optimal or suboptimal sliding surface for nonlinear systems. The pros and cons
of each approach are discussed via an illustrative example. In Chapter 4, two kinds
of disturbances, rate bounded and state dependent, are considered in the discrete-
time systems. Accordingly, two kinds of optimal discrete sliding mode controllers are
proposed, which are able to robustly and optimally control those disturbed discrete-
time systems.
In the second part, the dissertation is focusing on the stabilization problem of un-
by any smooth feedback law. Because of the discontinuity property of sliding mode
control, it has the potential to stabilize those underactuated systems. Two general
forms of underacuated systems, the cascaded form and the chained form, are of great
interest. Stabilizing sliding mode control laws for underactuated systems in those two
forms are proposed in Chapter 5. Several benchmark real applications including the
TORA system, the quadrotor helicopter and the Dubin’s car are used to test those
sliding mode control laws. The advantages of using sliding mode control to stabilize
143
an underactuated system are flexibility in designing, easiness of implementation and
satisfactory performance.
Although optimal sliding mode control problem has been solved thoroughly for
linear systems, optimal sliding mode control problem for nonlinear systems needs
further research effort. The class of nonlinear systems considered in this dissertation
only has limited examples. For more general forms of nonlinear systems, it requires
tems and analysis of the existence of the optimal control. The SDRE approach is
more promising than the other two approaches if an appropriate method is found to
control has wide-open possibilities for the further research directions. There are
many underactuated systems in real applications. Using different state feedback and
shall explore the possibilities of stabilizing systems using sliding mode control, which
are not covered by the two general forms considered in this dissertation.
144
BIBLIOGRAPHY
[1] M. Basin, A. Ferreira and L. Fridman, “LQG-robust sliding mode control for
linear stochastic systems with uncertainties,” Proceedings of the 9th International
Workshop on Variable Structure Systems, pp. 74–79, June 2006.
[4] K. Furuta, “Sliding mode control of a discrete system,” System & Control
Letters, vol. 14, pp. 145–152, 1990.
[6] Ï. Haskara, Ü. Özgüner and V. I. Utkin, “On sliding mode observers via equiv-
alent control approach,” International Journal of Control, vol. 71, no. 6, pp.
1051–1067, 1998.
[7] D. Krupp and Y. Shtessel, “Chattering-free sliding mode control with unmodeled
dynamics,” Proceedings of American Control Conference, pp. 530–534, 1999.
[9] H. Lee, E. Kim, H. J. Kang and M. Park, “Design of a sliding mode controller
with fuzzy sliding surfaces,” IEE Proceedings - Control Theory and Applications,
vol. 145, no. 5, pp. 411–418, 1998.
145
[10] J. J. Lee. and Y. Xu, “A new method of switching surface design for multivariable
variable structure system,” IEEE Transactions on Automatic Control, vol. 39,
no. 2, pp. 414–419, 1994.
[11] X. Li and S. Yurkovich, “Sliding mode control of delayed systems with appli-
cations to engine idle speed control,” IEEE Transactions on Control Systems
Technology, vol. 9, no. 6, pp. 802–810, 2001.
[13] R. Morgan and Ü. Özgüner, “A decentralized variable structure control algorithm
for robotic manipulators,” IEEE Journal of Robotics and Automations, vol. 1,
no. 1, pp. 57–65, 1985.
[15] E. V. Persson, “Synthesis of control systems operating linearly for small signals
and approximately ’Bang-Bang’ for large signals,” Proceedings of the 2nd IFAC
Congress, 1964.
[16] E. P. Ryan, Optimal Relay And Saturating Control System Synthesis, Peter
Pergrinus Ltd., 1984.
[17] V. I. Utkin, “Variable structure systems with sliding modes,” IEEE Transactions
on Automatic Control, vol. 22, pp. 212–222, April, 1977.
[21] K. D. Young, V. I. Utkin and Ü. Özgüner, “A control engineer’s guide to sliding
mode control,” IEEE Transactions on Control Systems Technology, vol. 7, no.
3, pp. 328–342, 1999.
146
[22] T. Acarman and Ü. Özgüner, “Relation of dynamic sliding surface design and
high order sliding mode controllers,” Proceedings of the 40th IEEE Conference
on Decision and Control, vol. 1, pp. 934–939, 2001.
[23] F. Al-abbass and Ü. Özgüner, “Decentralized model reference adaptive systems
using a variable structure control,” Proceedings of the 24th IEEE Conference on
Decision and Control, vol. 1, pp. 1473–1478, 1985.
[28] C. C. Chiang and Y. S. Tzeng, “Dynamic sliding mode adaptive controller for
nonlinear systems with higher-order and unmatched uncertainties,” Proceedings
of 1999 IEEE International Conference on Systems, Man, and Cybernetics, vol.
1, pp. 44–49, 1999.
[29] S. B. Choi, “Moving switching surfaces for robust control of second order variable
structure systems,” International Journal of Control, vol. 50, pp. 229–245, 1993.
147
[34] A. J. Koshkouei, K. J. Burnham and A. S. I. Zinober, “Dynamic sliding mode
control design,” IEE Proceedings - Control Theory and Applications, vol. 152,
no. 4, pp. 392–396, 2005.
[36] Y. Shtessel and Y. Orlov, “Nonminimum phase output tracking via dynamic
sliding manifolds,” Proceedings of the 1996 IEEE International Conference on
Control Applications, pp. 1066–1071, 1996.
[38] R. Xu and Ü. Özgüner, “Extremum seeking control via sliding mode with two
sliding surfaces,” Proceedings of the 16th IFAC World Congress, Prague, Czech,
July 2005.
[39] R. Xu and Ü. Özgüner, “Optimal sliding mode control for linear systems,”
Proceedings of the 9th International Workshop on Variable Structure Systems,
pp. 143-148, June, 2006.
[40] R. Xu and Ü. Özgüner, “Optimal sliding mode control of linear systems,” sub-
mitted to International Journal of Control.
[41] R. Xu and Ü. Özgüner, “Optimal sliding mode control for a class of nonlinear
systems,” submitted to IEEE Transactions on Automatic Control.
[42] K. D. Young and Ü. Özgüner, “Frequency shaped sliding mode synthesis,” Pro-
ceedings of the 1st International Workshop on Variable Structure Systems, Yu-
goslavia, 1990.
[43] K. D. Young and Ü. Özgüner, “Frequency shaped variable structure control,”
Proceedings of 1990 American Control Conference, vol. 1, pp. 181–185, 1990.
[44] K. D. Young and Ü. Özgüner, “Frequency shaping compensator design for sliding
mode,” International Journal of Control, vol. 57, pp. 1005–1019, 1993.
[45] K. D. Young and Ü. Özgüner, “Co-states for sliding mode design in linear
systems,” Systems & Control Letters, vol. 27, pp. 233–242, 1996.
[46] K. D. Young and Ü. Özgüner, “Sliding-mode design for robust linear optimal
control,” Automatica, vol. 33, no. 7, pp. 1313–1323, 1997.
148
[47] J. Zhang, Y. Zhang, Z. Chen and Z. Zhao, “A control scheme based on discrete
time-varying sliding surface for position control systems,” Proceedings of the 5th
World Congress on Intelligent Control and Automation, pp. 1175–1178, 2004.
[50] S. P. Banks and K. J. Mhana, “Optimal control and stabilization for nonlinear
systems,” IMA Journal of Mathematical Control and Information, vol. 9, no. 2,
pp. 179–196, 1992.
[54] F.A. Freeman and P.V. Kokotovic, Robust nonlinear control design: state-space
and Lyapunov techniques Boston: Birkhauser, 1996.
[55] F.A. Freeman and P.V. Kokotovic, “Inverse optimality in robust stabilization,”
SIAM J. Contr. Optimiz., vol. 34, pp. 1365–1391, 1996.
[57] Y. Huang and W.M. Lu, “Nonlinear optimal control: Alternatives to Hamilton-
Jacobi equation,” Proceedings of the 35th IEEE Conference on Decision and
Control, pp. 3942–3947, 1996.
[58] R.E. Kalman, “Contribution to the theory of optimal control,” Bol. Soc. Mat.
Mexicana, vol. 5, pp. 102–119, 1960.
[60] E. B. Lee and L. Markus, Foundations of Optimal Control Theory, John Wiley,
1967.
149
[61] F. L. Lewis and V.L. Syrmos. Optimal Control. 2nd Edition. John Wiley & Sons,
Inc. 1995.
[63] G. N. Saridis and C. G. Lee, “An approximation theory of optimal control for
trainable manipulations,” IEEE Transcations on Automatic Control, vol. 13, pp.
621–629, 1968.
[65] M. Sznaier and M. J. Damborg, “Suboptimal control of linear systems with state
and control inequality constraints,” Proceedings of the 26th IEEE Conference on
Decision and Control, pp. 761–762, 1986.
[70] I. Fantoni, R. Lozano and M. W. Spong, “Energy based control of the pendubot,”
IEEE Transactions on Automatic Control, vol. 45, no. 4, pp. 725–729, April 2000.
[71] R. Fierro, F. L. Lewis and A. Lowe, “Hybrid control for a class of underactuated
mechanical systems,” IEEE Transactions on systems, man, and cybernetics –
part A: systems and humans, vol. 29, no. 6, pp. 649–654, November 1999.
150
[74] Z. P. Jiang, D. J. Hill and Y. Guo, “Stabilization and tracking via ouptut
feedback for the nonlinear benchmark system,” Automatica, vol. 34, pp.907–915,
1998.
[76] M. Kawski, “Stabilization of nonlinear systems in the plane,” Systems & Control
Letters, vol. 12, pp. 169–175, 1989.
[81] R. Olfati-Saber, “Global configuration stabilization for the VTOL aircraft with
strong input coupling,” IEEE Transactions on Automatic Control, vol. 47, no.
11, pp. 1949–1952, 2002.
[82] C. Qian and W. Lin, “Non-Lipschitz continuous stabilizers for nonlinear systems
with uncontrollable unstable linearization,” Systems & Control Letters, vol. 42,
pp. 185–200, 2001.
151
[86] D. Seto and J. Baillieul, “Control problems in super-articulated mechanical
systems,” IEEE Transactions on Automatic Control, vol. 39, no. 12, pp. 2442–
2453, December 1994.
[89] S. Lee, M. Kim, Y. Youm, and W. Chung, “Control of a car-like mobile robot
for parking problem,” Proceedings of IEEE International Conference on Robotics
and Automation, pp. 1–6, 1999.
[94] C. Samson, “Control of chained systems application to path following and time-
varying point-stabilization of mobile robots,” IEEE Transcations on Automatic
Control, vol. 40, no. 1, pp. 64–77, 1995.
[95] R. Xu and Ü. Özgüner, “Sliding mode control of a quadrotor helicopter,” Pro-
ceedings of the 45th IEEE Conference on Control and Decision, pp. 4957-4962,
December, 2006.
[96] R. Xu and Ü. Özgüner, “Sliding mode control of a class of underactuated sys-
tems,” Accepted by Automatica.
152
[99] Ü. Özgüner, K. A. Ünyelioglu, C. Hatipoglu and F. Kantz, “desgin of a lateral
controller for a cooperative vehicle system,” SAE Paper 950474.
153