Robust and Optimal Control
Robust and Optimal Control
Robust and Optimal Control
CONTROL
KEMIN ZHOU
with
JOHN C. DOYLE and KEITH GLOVER
TO OUR PARENTS
Contents
Preface
xiii
xvi
List of Acronyms
xx
1 Introduction
1.1
1.2
1.3
Historical Perspective : : : : : : : : : : : : : : : : : : : : : : : : : :
How to Use This Book : : : : : : : : : : : : : : : : : : : : : : : : :
Highlights of The Book : : : : : : : : : : : : : : : : : : : : : : : : :
2 Linear Algebra
2.1
2.2
2.3
2.4
2.5
2.6
2.7
2.8
2.9
2.10
2.11
2.12
Linear Subspaces : : : : : : : : : : : : :
Eigenvalues and Eigenvectors : : : : : :
Matrix Inversion Formulas : : : : : : : :
Matrix Calculus : : : : : : : : : : : : : :
Kronecker Product and Kronecker Sum :
Invariant Subspaces : : : : : : : : : : : :
Vector Norms and Matrix Norms : : : :
Singular Value Decomposition : : : : : :
Generalized Inverses : : : : : : : : : : :
Semidenite Matrices : : : : : : : : : : :
Matrix Dilation Problems* : : : : : : : :
Notes and References : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
1
4
6
17
17
20
22
24
25
26
28
32
36
36
38
44
45
45
CONTENTS
viii
3.2
3.3
3.4
3.5
3.6
3.7
3.8
3.9
3.10
3.11
3.12
Normed Spaces : : : : : : : : :
Hilbert Spaces : : : : : : : : : :
Hardy Spaces H2 and H1 : : :
Power and Spectral Signals : :
Induced System Gains : : : : :
Computing L2 and H2 Norms :
Computing L1 and H1 Norms
Notes and References : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
Feedback Structure : : : : : : : : :
Well-Posedness of Feedback Loop :
Internal Stability : : : : : : : : : :
Coprime Factorization over RH1 :
Feedback Properties : : : : : : : :
The Concept of Loop Shaping : : :
Weighted H2 and H1 Performance
Notes and References : : : : : : : :
6 Performance Limitations
6.1
6.2
6.3
6.4
6.5
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
Introduction : : : : : : : : : : : : : : : : : :
Integral Relations : : : : : : : : : : : : : : :
Design Limitations and Sensitivity Bounds :
Bode's Gain and Phase Relation : : : : : :
Notes and References : : : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
47
53
58
63
65
68
71
72
78
80
89
91
91
93
97
102
104
112
114
116
117
117
119
121
126
130
134
137
141
143
143
145
149
151
152
CONTENTS
ix
Hankel Operator : : : : : : : : : : : : : : : :
All-pass Dilations : : : : : : : : : : : : : : : :
Optimal Hankel Norm Approximation : : : :
L1 Bounds for Hankel Norm Approximation
Bounds for Balanced Truncation : : : : : : :
Toeplitz Operators : : : : : : : : : : : : : : :
Hankel and Toeplitz Operators on the Disk* :
Nehari's Theorem* : : : : : : : : : : : : : : :
Notes and References : : : : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
Model Uncertainty : : : : : : : : : : : : : : : : : :
Small Gain Theorem : : : : : : : : : : : : : : : : :
Stability under Stable Unstructured Uncertainties :
Unstructured Robust Performance : : : : : : : : :
Gain Margin and Phase Margin : : : : : : : : : : :
Deciency of Classical Control for MIMO Systems
Notes and References : : : : : : : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
153
154
160
161
167
169
170
175
184
188
192
194
195
200
206
207
207
211
214
222
229
232
237
239
239
246
256
257
260
261
262
266
276
287
290
CONTENTS
14 H2 Optimal Control
14.1
14.2
14.3
14.4
14.5
14.6
14.7
14.8
14.9
14.10
14.11
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
Hankel Operators : : : : : : : : : : : : : : : : : : : : :
Toeplitz Operators : : : : : : : : : : : : : : : : : : : :
Mixed Hankel-Toeplitz Operators : : : : : : : : : : : :
Mixed Hankel-Toeplitz Operators: The General Case*
Linear Quadratic Max-Min Problem : : : : : : : : : :
Notes and References : : : : : : : : : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
291
292
295
301
312
314
315
318
319
320
325
333
342
353
357
358
362
364
365
365
367
372
373
375
378
379
381
388
390
392
393
393
395
397
399
401
404
CONTENTS
Problem Formulation : : : : : : : :
Output Feedback H1 Control : : :
Motivation for Special Problems : :
Full Information Control : : : : : :
Full Control : : : : : : : : : : : : :
Disturbance Feedforward : : : : :
Output Estimation : : : : : : : : :
Separation Theory : : : : : : : : :
Optimality and Limiting Behavior
Controller Interpretations : : : : :
An Optimal Controller : : : : : : :
Notes and References : : : : : : : :
xi
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
General H1 Solutions : : : : : : : : : : : : : : : : :
Loop Shifting : : : : : : : : : : : : : : : : : : : : : :
Relaxing Assumptions : : : : : : : : : : : : : : : : :
H2 and H1 Integral Control : : : : : : : : : : : : : :
H1 Filtering : : : : : : : : : : : : : : : : : : : : : :
Youla Parameterization Approach* : : : : : : : : : :
Connections : : : : : : : : : : : : : : : : : : : : : : :
State Feedback and Dierential Game : : : : : : : :
Parameterization of State Feedback H1 Controllers :
Notes and References : : : : : : : : : : : : : : : : : :
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
18 H1 Loop Shaping
18.1
18.2
18.3
18.4
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
405
405
406
412
416
423
423
425
426
431
436
438
439
441
442
445
449
450
454
456
460
462
465
468
469
469
478
481
487
489
490
497
504
509
513
CONTENTS
xii
515
527
20.1
20.2
20.3
21.1
21.2
21.3
21.4
21.5
21.6
21.7
21.8
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
527
528
537
542
548
553
558
561
Bibliography
563
Index
581
Preface
This book is inspired by the recent development in the robust and H1 control theory,
particularly the state-space H1 control theory developed in the paper by Doyle, Glover,
Khargonekar, and Francis [1989] (known as DGKF paper). We give a fairly comprehensive and step-by-step treatment of the state-space H1 control theory in the style
of DGKF. We also treat the robust control problems with unstructured and structured
uncertainties. The linear fractional transformation (LFT) and the structured singular
value (known as ) are introduced as the unied tools for robust stability and performance analysis and synthesis. Chapter 1 contains a more detailed chapter-by-chapter
review of the topics and results presented in this book.
We would like to thank Professor Bruce A. Francis at University of Toronto for his
helpful comments and suggestions on early versions of the manuscript. As a matter
of fact, this manuscript was inspired by his lectures given at Caltech in 1987 and his
masterpiece { A Course in H1 Control Theory. We are grateful to Professor Andre Tits
at University of Maryland who has made numerous helpful comments and suggestions
that have greatly improved the quality of the manuscript. Professor Jakob Stoustrup,
Professor Hans Henrik Niemann, and their students at The Technical University of
Denmark have read various versions of this manuscript and have made many helpful
comments and suggestions. We are grateful to their help. Special thanks go to Professor
Andrew Packard at University of California-Berkeley for his help during the preparation
of the early versions of this manuscript. We are also grateful to Professor Jie Chen at
University of California-Riverside for providing material used in Chapter 6. We would
also like to thank Professor Kang-Zhi Liu at Chiba University (Japan) and Professor
Tongwen Chen at University of Calgary for their valuable comments and suggestions.
In addition, we would like to thank G. Balas, C. Beck, D. S. Bernstein, G. Gu, W.
Lu, J. Morris, M. Newlin, L. Qiu, H. P. Rotstein and many other people for their
comments and suggestions. The rst author is especially grateful to Professor Pramod
P. Khargonekar at The University of Michigan for introducing him to robust and H1
control and to Professor Tryphon Georgiou at University of Minnesota for encouraging
him to complete this work.
Kemin Zhou
xiii
xiv
PREFACE
John C. Doyle
Keith Glover
xv
xvi
2
[
\
belong to
subset
union
intersection
2
3
~
end of proof
end of example
end of remark
:=
'
/
dened as
asymptotically greater than
asymptotically less than
much greater than
much less than
jj
Re ()
complex conjugate of 2 C
absolute value of 2 C
real part of 2 C
(t)
ij
1+(t)
unit impulse
Kronecker delta function, ii = 1 and ij = 0 if i 6= j
unit step function
xvii
n n identity matrix
a matrix with aij as its i-th row and j -th column element
diag(a1 ; : : : ; an ) an n n diagonal matrix with ai as its i-th diagonal element
AT
transpose
A
adjoint operator of A or complex conjugate transpose of A
A;1
inverse of A
A+
pseudo inverse of A
A;
shorthand for (A;1 )
det(A)
determinant of A
Trace(A)
trace of A
(A)
eigenvalue of A
(A)
spectral radius of A
(A)
the set of spectrum of A
(A)
largest singular value of A
(A)
smallest singular value of A
i (A)
i-th singular value of A
(A)
condition number of A
kAk
spectral norm of A: kAk = (A)
Im(A), R(A)
image (or range) space of A
Ker(A), N(A)
kernel (or null) space of A
X; (A)
stable invariant subspace of A
X+ (A)
antistable invariant subspace of A
In
[aij ]
Ric(H )
gf
\
h; i
x?y
D?
S?
L2 (;1; 1)
L2 [0; 1)
L2 (;1; 0]
L2+
L2;
l2+
l2;
L2 (j R)
L2 (@ D )
H2 (j R)
xviii
H2 (@ D )
H2? (j R)
H2? (@ D )
L 1 ( j R)
L1 (@ D )
H1 (j R)
H1 (@ D )
H1; (j R)
H1; (@ D )
Rp (s)
polynomial ring
rational proper transfer matrices
G (s)
(z )
G
A B
C D
F` (M; Q)
Fu (M; Q)
S (M; N )
lower LFT
upper LFT
star product
xix
LIST OF ACRONYMS
xx
List of Acronyms
ARE
BR
CIF
DF
FC
FDLTI
FI
HF
i
lcf
LF
LFT
lhp or LHP
LQG
LQR
LTI
LTR
MIMO
nlcf
NP
nrcf
NS
OE
OF
OI
rcf
rhp or RHP
RP
RS
SF
SISO
SSV
SVD
Introduction
1.1 Historical Perspective
This book gives a comprehensive treatment of optimal H2 and H1 control theory and
an introduction to the more general subject of robust control. Since the central subject
of this book is state-space H1 optimal control, in contrast to the approach adopted
in the famous book by Francis [1987]: A Course in H1 Control Theory, it may be
helpful to provide some historical perspective of the state-space H1 control theory to
be presented in this book. This section is not intended as a review of the literature
in H1 theory or robust control, but rather only an attempt to outline some of the
work that most closely touches on our approach to state-space H1 . Hopefully our lack
of written historical material will be somewhat made up for by the pictorial history of
control shown in Figure 1.1. Here we see how the practical but classical methods yielded
to the more sophisticated modern theory. Robust control sought to blend the best of
both worlds. The strange creature that resulted is the main topic of this book.
The H1 optimal control theory was originally formulated by Zames [1981] in an
input-output setting. Most solution techniques available at that time involved analytic
functions (Nevanlinna-Pick interpolation) or operator-theoretic methods [Sarason, 1967;
Adamjan et al., 1978; Ball and Helton, 1983]. Indeed, H1 theory seemed to many to
signal the beginning of the end for the state-space methods which had dominated control
for the previous 20 years. Unfortunately, the standard frequency-domain approaches to
H1 started running into signicant obstacles in dealing with multi-input multi-output
(MIMO) systems, both mathematically and computationally, much as the H2 (or LQG)
theory of the 1950's had.
1
INTRODUCTION
from Limebeer and Hung [1987] and Limebeer and Halikias [1988] who showed, for
problems transformable to 2 1-block problems, that a subsequent minimal realization
of the controller has state dimension no greater than that of the generalized plant G.
This suggested the likely existence of similarly low dimension optimal controllers in the
general 2 2 case.
Additional progress on the 2 2-block problems came from Ball and Cohen [1987],
who gave a state-space solution involving 3 Riccati equations. Jonckheere and Juang
[1987] showed a connection between the 2 1-block problem and previous work by
Jonckheere and Silverman [1978] on linear-quadratic control. Foias and Tannenbaum
[1988] developed an interesting class of operators called skew Toeplitz to study the
2 2-block problem. Other approaches have been derived by Hung [1989] using an
interpolation theory approach, Kwakernaak [1986] using a polynomial approach, and
Kimura [1988] using a method based on conjugation.
The simple state space H1 controller formulae to be presented in this book were rst
derived in Glover and Doyle [1988] with the 1984 approach, but using a new 2 2-block
solution, together with a cumbersome back substitution. The very simplicity of the new
formulae and their similarity with the H2 ones suggested a more direct approach.
Independent encouragement for a simpler approach to the H1 problem came from
papers by Petersen [1987], Khargonekar, Petersen, and Zhou [1990], Zhou and Khargonekar [1988], and Khargonekar, Petersen, and Rotea [1988]. They showed that for
the state-feedback H1 problem one can choose a constant gain as a (sub)optimal controller. In addition, a formula for the state-feedback gain matrix was given in terms
of an algebraic Riccati equation. Also, these papers established connections between
H1 -optimal control, quadratic stabilization, and linear-quadratic dierential games.
The landmark breakthrough came in the DGKF paper (Doyle, Glover, Khargonekar,
and Francis [1989]). In addition to providing controller formulae that are simple and
expressed in terms of plant data as in Glover and Doyle [1988], the methods in that
paper are a fundamental departure from the 1984 approach. In particular, the Youla
parameterization and the resulting 2 2-block model-matching problem of the 1984
solution are avoided entirely; replaced by a more purely state-space approach involving
observer-based compensators, a pair of 2 1 block problems, and a separation argument.
The operator theory still plays a central role (as does Redheer's work [Redheer, 1960]
on linear fractional transformations), but its use is more straightforward. The key
to this was a return to simple and familiar state-space tools, in the style of Willems
[1971], such as completing the square, and the connection between frequency domain
inequalities (e.g kGk1 < 1), Riccati equations, and spectral factorizations. This book
in some sense can be regarded as an expansion of the DGKF paper.
The state-space theory of H1 can be carried much further, by generalizing timeinvariant to time-varying, innite horizon to nite horizon, and nite dimensional to
innite dimensional. A
ourish of activity has begun on these problems since the publication of the DGKF paper and numerous results have been published in the literature,
not surprising, many results in DGKF paper generalize mutatis mutandis, to these cases,
which are beyond the scope of this book.
INTRODUCTION
3:9
11
A
U
A
10
12
13
@
@
R
@
15
16
@
I
@
@
17
;
?;
;
19
14
13:2
13:6
13:4
18
20
21
INTRODUCTION
stage with the background of Section 3.9, they may serve as an introduction to sources
of model uncertainties and hence to robustness problems.
Robust Control
4
5
6*
7*
8*
9
10
11
H1 Control
Advanced
H1 Control
Table 1.1: Possible choices for an one-semester course (* chapters may be omitted)
Table 1.1 lists several possible choices of topics for an one-semester course. A course
on model and controller reductions may only include the concept of H1 control and
the H1 controller formulas with the detailed proofs omitted as suggested in the above
table.
X B
min
X
C A
= max
C A
B
A
A Q + QA + C C = 0:
Chapter 4 denes several norms for signals and introduces the H2 spaces and the H1
spaces. The input/output gains of a stable linear system under various input signals
are characterized. We show that H2 and H1 norms come out naturally as measures of
the worst possible performance for many classes of input signals. For example, let
kg uk
n
X
w1
e1
+6
P
K^
++ w2
e2 ?
e
We show that the above closed-loop system is internally stable if and only if
I ;K^ ;1 = I + K^ (I ; P K^ );1 P K^ (I ; P K^ );1 2 RH1 :
;P I
(I ; P K^ );1 P
(I ; P K^ );1
Alternative characterizations of internal stability using coprime factorizations are also
presented.
Chapter 6 introduces some multivariable versions of the Bode's sensitivity integral
relations and Poisson integral formula. The sensitivity integral relations are used to
study the design limitations imposed by bandwidth constraint and the open-loop unstable poles, while the Poisson integral formula is used to study the design constraints
INTRODUCTION
imposed by the non-minimum phase zeros. For example, let S (s) be a sensitivity function, and let pi be the right half plane poles of the open-loop system and i be the
corresponding pole directions. Then we show that
1
This equality shows that the design limitations in multivariable systems are dependent
on the directionality properties of the sensitivity function as well as those of the poles
(and zeros), in addition to the dependence upon pole (and zero) locations which is
known in single-input single-output systems.
Chapter 7 considers the problem of reducing the order of a linear multivariable
dynamical system using the balanced truncation method. Suppose
2
A11 A12
A21 A22
C1 C2
G(s) = 4
B1
B2
D
2 RH1
Then the truncated system Gr (s) = AC11 BD1 is stable and satises an additive
1
error bound:
N
X
kG(s) ; Gr (s)k1 2
i :
i=r+1
is stable and minimum phase, and satises respectively the following relative and multiplicative error bounds:
G;1 (G ; Gr )
G;1 (G ; Gr )
1
1
N
Y
1 + 2i (
i=r+1
N
Y
i=r+1
1 + 2i (
1 + 2 + i )
1 + 2 + i )
i
;1
; 1:
Chapter 8 deals with the optimal Hankel norm approximation and its applications
in L1 norm model reduction. We show that for a given G(s) of McMillan degree n
there is a G^ (s) of McMillan degree r < n such that
N
X
G(s) ; G^ (s) ; D0 1
i=r+1
i :
Q2RH;1
Chapter 9 derives robust stability tests for systems under various modeling assumptions through the use of a small gain theorem. In particular, we show that an uncertain
system described below with an unstructured uncertainty 2 RH1 with kk1 < 1 is
robustly stable if and only if the transfer function from w to z has H1 norm no greater
than 1.
nominal system
-
z
z
w
-K
w
u
INTRODUCTION
10
Chapter 11 considers robust stability and performance for systems with multiple
sources of uncertainties. We show that an uncertain system is robustly stable for all
i 2 RH1 with ki k1 < 1 if and only if the structured singular value () of the
corresponding interconnection model is no greater than 1.
1
nominal system
Chapter 12 characterizes all controllers that stabilize a given dynamical system G(s)
using the state space approach. The construction of the controller parameterization is
done via separation theory and a sequence of special problems, which are so-called full
information (FI) problems, disturbance feedforward (DF) problems, full control (FC)
problems and output estimation (OE). The relations among these special problems are
established.
FI
dual
equivalent
FC
6
equivalent
DF
dual
OE
A B1 B2
11 (s) G12 (s) = 4 C D
G(s) = G
1
11 D12
G21 (s) G22 (s)
C2 D21 D22
3
5
we show that all stabilizing controllers can be parameterized as the transfer matrix from
y to u below where F and L are such that A + LC2 and A + B2 F are stable.
11
z
y
? ?
;
C2
D22
R
-A
66
c
B2
6
c
-F
- ;L
u1
y1
Chapter 13 studies the Algebraic Riccati Equation and the related problems: the
properties of its solutions, the methods to obtain the solutions, and some applications.
In particular, we study in detail the so-called stabilizing solution and its applications in
matrix factorizations. A solution to the following ARE
A X + XA + XRX + Q = 0
is said to be a stabilizing solution if A + RX is stable. Now let
H := ;AQ ;RA
and let X; (H ) be the stable H invariant subspace and
X
1
X; (H ) = Im X2
where X1 ; X2 2 C nn . If X1 is nonsingular, then X := X2 X1;1 is uniquely determined
by H , denoted by X = Ric(H ).
A key result of this chapter is the relationship between the spectral factorization of a
transfer function and the solution of a corresponding ARE. Suppose (A; B ) is stabilizable
and suppose either A has no eigenvalues on j!-axis or P is sign denite (i.e., P 0 or
P 0) and (P; A) has no unobservable modes on the j!-axis. Dene
P S
S R
(sI ; A);1 B :
I
INTRODUCTION
12
Then
;1 S
;BR;1B
H = ;A(P;;BR
;
1
SR S ) ;(A ; BR;1 S )
() 9 a solution X to
M = ;AF BI ; F = ;R;1(S + B X ):
Chapter 14 treats the optimal control of linear time-invariant systems with quadratic
performance criteria, i.e., LQR and H2 problems. We consider a dynamical system
described by an LFT with
3
A B1 B2
G(s) = 4 C1 0 D12 5 :
C2 D21 0
z
y
w
13
C1 B2
D12
D21 B1 C2
J2 := ;BA B ;0A ; ;BC2D
1 1
1 21
X2 := Ric(H2) 0; Y2 := Ric(J2 ) 0
C1 ); L2 := ;(Y2 C + B1 D ):
F2 := ;(B2 X2 + D12
2
21
Then the H2 optimal controller, i.e. the controller that minimizes kTzw k2 , is given by
Chapter 15 solves a max-min problem, i.e., a full information (or state feedback)
H1 control problem, which is the key to the H1 theory considered in the next chapter.
Consider a dynamical system
x_ = Ax + B1 w + B2 u
C1 D12 = 0 I :
z = C1 x + D12 u; D12
Then we show that sup umin
kz k2 <
if and only if H1 2 dom(Ric) and X1 =
w2BL2+ 2L2+
Ric(H1 ) 0 where
;2
H1 := ;CA C
B1 B;1A; B2 B2
1 1
G(s) as given in Chapter 14. We show that there exists an admissible controller such
that kTzw k1 <
i the following three conditions hold:
(i) H1 2 dom(Ric) and X1 := Ric(H1 ) 0;
(ii) J1 2 dom(Ric) and Y1 := Ric(J1 ) 0 where
;2
J1 := ;BA B
C1 C;1A; C2 C2 :
1 1
INTRODUCTION
14
Moreover, the set of all admissible controllers such that kTzw k1 <
equals the set of
all transfer matrices from y to u in
u
y
M1
A^1 ;Z1 L1 Z1 B2
4
M1 (s) = F1
0
I
I
0
;C2
3
5
F1 := ;B2 X1 ;
L1 := ;Y1 C2 ;
Z1 := (I ; ;2Y1 X1 );1 :
Chapter 17 considers again the standard H1 control problem but with some assumptions in the last chapter relaxed. We indicate how the assumptions can be relaxed
to accommodate other more complicated problems such as singular control problems.
We also consider the integral control in the H2 and H1 theory and show how the
general H1 solution can be used to solve the H1 ltering problem. The conventional
Youla parameterization approach to the H2 and H1 problems is also outlined. Finally,
the general state feedback H1 control problem and its relations with full information
control and dierential game problems are discussed.
Chapter 18 rst solves a gap metric minimization problem. Let P = M~ ;1 N~ be a
normalized left coprime factorization. Then we show that
K
stabilizing
= K inf
stabilizing
inf
K (I + PK );1 I P
I
K (I + PK );1M~
I
;1
q
1;
N~ M~
2
;1
~ N ~ M
1<
1 ;
N~ M~
2
H:
Using this stabilization result, a loop shaping design technique is proposed. The proposed technique uses only the basic concept of loop shaping methods and then a robust
15
stabilization controller for the normalized coprime factor perturbed system is used to
construct the nal controller.
Chapter 19 considers the design of reduced order controllers by means of controller
reduction. Special attention is paid to the controller reduction methods that preserve the
closed-loop stability and performance. In particular, two H1 performance preserving
reduction methods are proposed:
^
a) Let K0 be a stabilizing controller
such that
kF` (G; K0 )k1 <
. Then K is also a
stabilizing controller such that
F`(G; K^ )
<
if
1
W2;1 (K^ ; K0)W1;1
1 < 1
where W1 and W2 are some stable, minimum phase and invertible transfer matrices.
b) Let K0 = 12 ;221 be a central H1 controller such that kF`(G; K0 )k1 <
and
^ V^ 2 RH1 be such that
let U;
p
^
;1 I 0
U
12
;
1
2:
;
<
1
=
0 I
22
V^
1
Then K^ = U^ V^ ;1 is also a stabilizing controller such that kF`(G; K^ )k1 <
.
Thus the controller reduction problem is converted to weighted model reduction problems for which some numerical methods are suggested.
Chapter 20 brie
y introduces the Lagrange multiplier method for the design of xed
order controllers.
Chapter 21 discusses discrete time Riccati equations and some of their applications in
discrete time control. Finally, the discrete time balanced model reduction is considered.
16
INTRODUCTION
Linear Algebra
Some basic linear algebra facts will be reviewed in this chapter. The detailed treatment
of this topic can be found in the references listed at the end of the chapter. Hence we shall
omit most proofs and provide proofs only for those results that either cannot be easily
found in the standard linear algebra textbooks or are insightful to the understanding of
some related problems. We then treat a special class of matrix dilation problems which
will be used in Chapters 8 and 17; however, most of the results presented in this book
can be understood without the knowledge of the matrix dilation theory.
LINEAR ALGEBRA
18
such a basis for a subspace S is not unique but all bases for S have the same number
of elements. This number is called the dimension of S , denoted by dim(S ).
A set of vectors fx1 ; x2 ; : : : ; xk g in Fn are mutually orthogonal if xi xj = 0 for all
i 6= j and orthonormal if xi xj = ij , where the superscript denotes complex conjugate
transpose and ij is the Kronecker delta function with ij = 1 for i = j and ij = 0
for i 6= j . More generally, a collection of subspaces S1 ; S2 ; : : : ; Sk of Fn are mutually
orthogonal if x y = 0 whenever x 2 Si and y 2 Sj for i 6= j .
The orthogonal complement of a subspace S Fn is dened by
S ? := fy 2 Fn : y x = 0 for all x 2 S g:
We call a set of vectors fu1; u2 ; : : : ; uk g an orthonormal basis for a subspace S 2 Fn if
they form a basis of S and are orthonormal. It is always possible to extend such a basis
to a full orthonormal basis fu1 ; u2; : : : ; un g for Fn . Note that in this case
S ? = spanfuk+1 ; : : : ; un g;
and fuk+1 ; : : : ; un g is called an orthonormal completion of fu1; u2; : : : ; uk g.
Let A 2 Fmn be a linear transformation from Fn to Fm , i.e.,
A : Fn 7;! Fm :
(Note that a vector x 2 Fm can also be viewed as a linear transformation from F to Fm ,
hence anything said for the general matrix case is also true for the vector case.) Then
the kernel or null space of the linear transformation A is dened by
KerA = N (A) := fx 2 Fn : Ax = 0g;
and the image or range of A is
ImA = R(A) := fy 2 Fm : y = Ax; x 2 Fn g:
It is clear that KerA is a subspace of Fn and ImA is a subspace of Fm . Moreover, it can
be easily seen that dim(KerA) + dim(ImA) = n and dim(ImA) = dim(KerA)? . Note
that (KerA)? is a subspace of Fn .
Let ai ; i = 1; 2; : : :; n denote the columns of a matrix A 2 Fmn , then
ImA = spanfa1 ; a2 ; : : : ; an g:
The rank of a matrix A is dened by
rank(A) = dim(ImA):
It is a fact that rank(A) = rank(A ), and thus the rank of a matrix equals the maximal
number of independent rows or columns. A matrix A 2 Fmn is said to have full row
rank if m n and rank(A) = m. Dually, it is said to have full column rank if n m
and rank(A) = n. A full rank square matrix is called a nonsingular matrix. It is easy
19
to see that rank(A) = rank(AT ) = rank(PA) if T and P are nonsingular matrices with
appropriate dimensions.
A square matrix U 2 F nn whose columns form an orthonormal basis for Fn is called
an unitary matrix (or orthogonal matrix if F = R), and it satises U U = I = UU .
The following lemma is useful.
Lemma 2.1 Let D = d1 : : : dk 2 Fnk (n > k) be such that DD = I , so
d i ; i = 1; 2; : : : ; k are orthonormal. Then there exists a matrix D? 2 Fn(n;k) such that
D D? is a unitary matrix. Furthermore, the columns of D?, di ; i = k + 1; : : : ; n,
form an orthonormal completion of fd1 ; d2 ; : : : ; dk g.
The following results are standard:
Lemma 2.2 Consider the linear equation
where A 2 Fnl and
AX = B
equivalent:
(i) there exists a solution X 2 Flm .
(ii) the columns of B 2 ImA.
(iii) rank A B =rank(A).
(iv) Ker(A ) Ker(B ).
Furthermore, the solution, if it exists, is unique if and only if A has full column rank.
The following lemma concerns the rank of the product of two matrices.
Lemma 2.3 (Sylvester's inequality) Let A 2 Fmn and B 2 Fnk . Then
rank (A) + rank(B ) ; n rank(AB ) minfrank (A); rank(B )g:
For simplicity, a matrix M with mij as its i-th row and j -th column's element will
sometimes be denoted as M = [mij ] in this book. We will mostly use I as above to
denote an identity matrix with compatible dimensions, but from time to time, we will
use In to emphasis that it is an n n identity matrix.
Now let A = [aij ] 2 C nn , then the trace of A is dened as
Trace(A) :=
n
X
i=1
aii :
LINEAR ALGEBRA
20
Let A 2 C nn ; then the eigenvalues of A are the n roots of its characteristic polynomial
p() = det(I ; A). This set of roots is called the spectrum of A and is denoted by (A)
(not to be confused with singular values dened later). That is, (A) := f1 ; 2 ; : : : ; n g
if i is a root of p(). The maximal modulus of the eigenvalues is called the spectral
radius, denoted by
(A) := 1max
j j
in i
where, as usual, j j denotes the magnitude.
If 2 (A) then any nonzero vector x 2 C n that satises
Ax = x
is referred to as a right eigenvector of A. Dually, a nonzero vector y is called a left
eigenvector of A if
y A = y :
It is a well known (but nontrivial) fact in linear algebra that any complex matrix admits
a Jordan Canonical representation:
Theorem 2.4 For any square complex matrix A 2 C nn , there exists a nonsingular
matrix T such that
where
A = TJT ;1
J = diagfJ1 ; J2 ; : : : ; Jl g
Ji = diagfJi1 ; Ji2 ; : : : ; Jimi g
2
3
i 1
6
7
i 1
6
7
Jij = 666
... ...
7
7
7
5
2 C nij nij
i 1
i
Pl Pmi
with i=1 j=1 nij = n, and with fi : i = 1; : : : ; lg as the distinct eigenvalues of A.
The transformation T has the following form:
T = T1 T2 : : : Tl
Ti = Ti1 Ti2 : : : Timi
Tij = tij1 tij2 : : : tijnij
where tij1 are the eigenvectors of A,
Atij1 = i tij1 ;
4
21
A x1 x2 xn = x1 x2 xn
6
6
6
4
1
2
...
n
7
7
7
5
A=
where yi 2 C n is given by
2
6
6
6
4
y1
y2
..
.
yn
n
X
i=1
i xi yi
3
7
7
7
5
= x1 x2 xn ;1 :
In general, eigenvalues need not be real, and neither do their corresponding eigenvectors. However, if A is real and is a real eigenvalue of A, then there is a real eigenvector
corresponding to . In the case that all eigenvalues of a matrix A are real1 , we will
denote max (A) for the largest eigenvalue of A and min (A) for the smallest eigenvalue.
In particular, if A is a Hermitian matrix, then there exist a unitary matrix U and a
real diagonal matrix such that A = U U , where the diagonal elements of are the
eigenvalues of A and the columns of U are the eigenvectors of A.
The following theorem is useful in linear system theory.
Theorem 2.5 (Cayley-Hamilton) Let A 2 C nn and denote
det(I ; A) = n + a1 n;1 + + an :
Then
An + a1 An;1 + + an I = 0:
1
LINEAR ALGEBRA
22
This is obvious if A has distinct eigenvalues. Since
det(I ; A) = n + a1 n;1 + + an
where ai and Ri can be computed from the following recursive formulas:
a1
a2
=
=
..
.
=
=
; Trace A
; 21 Trace(R2 A)
R1 = I
R2 = R1 A + a1 I
..
.
an;1
; n;1 1 Trace(Rn;1 A)
Rn = Rn;1 A + an;1 I
0 = Rn A + an I:
an
; n1 Trace(Rn A)
The proof is left to the reader as an exercise. Note that the Cayley-Hamilton Theorem
follows from the fact that
0 = Rn A + an I = An + a1 An;1 + + an I:
11 A12
A := A
A21 A22
where A11 and A22 are also square matrices. Now suppose A11 is nonsingular, then A
has the following decomposition:
I
0
A11 A12 =
A21 A22
A21 A;111 I
A11 0
0
I A;111 A12
0
I
0
0 A22
I
0
A;221 A21 I
23
with := A11 ; A12 A;221 A21 , and A is nonsingular i is nonsingular. The matrix
() is called the Schur complement of A11 (A22 ) in A.
Moreover, if A is nonsingular, then
and
A11 A12 ;1 = A;111 + A;111 A12 ;1 A21 A;111 ;A;111 A12 ;1
A21 A22
;;1 A21 A;111
;1
#
"
;1
;;1 A12 A;221
A11 A12 ;1 =
:
A21 A22
;A;221 A21 ;1 A;221 + A;221 A21 ;1 A12 A;221
The above matrix inversion formulas are particularly simple if A is block triangular:
A11 0 ;1 =
A;111
0
;
1
;
1
A21 A22
;A22 A21 A11 A;221
A11 A12 ;1 = A;111 ;A;111 A12 A;221 :
0 A22
0
A;221
The following identity is also very useful. Suppose A11 and A22 are both nonsingular
matrices, then
(A11 ; A12 A;221 A21 );1 = A;111 + A;111 A12 (A22 ; A21 A;111 A12 );1 A21 A;111 :
As a consequence of the matrix decomposition formulas mentioned above, we can
calculate the determinant of a matrix by using its sub-matrices. Suppose A11 is nonsingular, then
det A = det A11 det(A22 ; A21 A;111 A12 ):
On the other hand, if A22 is nonsingular, then
det A = det A22 det(A11 ; A12 A;221 A21 ):
In particular, for any B 2 C mn and C 2 C nm , we have
det(In + xy ) = 1 + y x:
LINEAR ALGEBRA
24
@
@
@X F (X ) := @xij F (X ) :
Let A and B be constant complex matrices with compatible dimensions. Then the
following is a list of formulas for the derivatives2:
@
@X TracefAXB g
@
T
@X TracefAX B g
@
@X TracefAXBX g
@
T
@X TracefAXBX g
@
k
@X TracefX g
= AT B T
@
k
@X TracefAX g
@
;1
@X TracefAX B g
@
@X log det X
@
T
@X det X
@
k
@X detfX g
= BA
= AT X T B T + B T X T AT
= AT XB T + AXB
= k(X k;1 )T
k;1 X i AX k;i;1 T
i=0
P
= (X T );1
@ det X = (detX )(X T );1
= @X
so that all the rules applicable to a scalar function also apply here. In particular, we
have
d(AB ) = dA B + A dB
d
d
d
dA;1 = ;A;1 dA A;1 :
d
d
2 Note that transpose rather than complex conjugate transpose should be used in the list even if the
involved matrices are complex matrices.
25
Let A 2 C mn and B 2 C pq , then the Kronecker product of A and B is dened as
2
6
A
B := 664
..
.
..
.
3
7
7
7
5
2 C mpnq :
Furthermore, if the matrices A and B are square and A 2 C nn and B 2 C mm then
the Kronecker sum of A and B is dened as
A B := (A
Im ) + (In
B ) 2 C nmnm :
Let X 2 C mn and let vec(X ) denote the vector formed by stacking the columns of X
into one long vector:
2
3
vec(X ) :=
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
x11
x21
..
.
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
5
xm1
x12
x22 :
.
..
x1n
x2n
..
.
xmn
Then for any matrices A 2 C km , B 2 C nl , and X 2 C mn , we have
vec(AXB ) = (B T
A)vec(X ):
Consequently, if k = m and l = n, then
vec(AX + XB ) = (B T A)vec(X ):
Let A 2 C nn and B 2 C mm , and let fi ; i = 1; : : : ; ng be the eigenvalues of A
and fj ; j = 1; : : : ; mg be the eigenvalues of B . Then we have the following properties:
The eigenvalues of A
B are the mn numbers i j , i = 1; 2; : : :; n, j = 1; 2; : : :; m.
The eigenvalues of A B = (A
Im ) + (In
B ) are the mn numbers i + j ,
i = 1; 2; : : :; n, j = 1; 2; : : : ; m.
LINEAR ALGEBRA
26
(2:1)
where A 2 Fnn , B 2 Fmm , and C 2 Fnm are given matrices. There exists a unique
solution X 2 Fnm if and only if i (A)+j (B ) 6= 0; 8i = 1; 2; : : : ; n and j = 1; 2; : : : ; m.
In particular, if B = A , (2.1) is called the \Lyapunov Equation"; and the necessary
and sucient condition for the existence of a unique solution is that i (A) + j (A) 6=
0; 8i; j = 1; 2; : : : ; n
Proof. Equation (2.1) can be written as a linear matrix equation by using the Kronecker product:
(B T A)vec(X ) = vec(C ):
Now this equation has a unique solution i B T A is nonsingular. Since the eigenvalues
of B T A have the form of i (A) + j (B T ) = i (A) + j (B ), the conclusion follows. 2
The properties of the Lyapunov equations will be studied in more detail in the next
chapter.
27
let x1 ; x2 ; : : : ; xl be the corresponding eigenvector and the generalized eigenvectors obtained through the following equations:
(A ; 1 I )x1 = 0
(A ; 1 I )x2 = x1
..
.
(A ; 1 I )xl = xl;1 :
Then a subspace S with xt 2 S for some t l is an A-invariant subspace only if all lower
rank eigenvectors and generalized eigenvectors of xt are in S , i.e., xi 2 S; 81 i t.
This will be further illustrated in Example 2.1.
On the other hand, if S is a nontrivial subspace3 and is A-invariant, then there is
x 2 S and such that Ax = x.
An A-invariant subspace S C n is called a stable invariant subspace if all the
eigenvalues of A constrained to S have negative real parts. Stable invariant subspaces
will play an important role in computing the stabilizing solutions to the algebraic Riccati
equations in Chapter 13.
Example 2.1 Suppose a matrix A has the following Jordan canonical form
2
A x1 x2 x3 x4 = x1 x2 x3 x4
6
6
4
1 1
1
3
4
7
7
5
with Re1 < 0, 3 < 0, and 4 > 0. Then it is easy to verify that
S1 = spanfx1 g
S3 = spanfx3 g
S4 = spanfx4 g
S12 = spanfx1 ; x2 g
S13 = spanfx1 ; x3 g
S14 = spanfx1 ; x4 g
S123 = spanfx1 ; x2 ; x3 g
S124 = spanfx1 ; x2 ; x4 g
S34 = spanfx3 ; x4 g
are all A-invariant subspaces. Moreover, S1 ; S3 ; S12 ; S13 , and S123 are stable A-invariant
subspaces. However, the subspaces S2 = spanfx2 g, S23 = spanfx2 ; x3 g, S24 = spanfx2 ; x4 g,
and S234 = spanfx2 ; x3 ; x4 g are not A-invariant subspaces since the lower rank generalized eigenvector x1 of x2 is not in these subspaces. To illustrate, consider the subspace
S23 . Then by denition, Ax2 2 S23 if it is an A-invariant subspace. Since
Ax2 = x2 + x1 ;
Ax2 2 S23 would require that x1 be a linear combination of x2 and x3 , but this is
impossible since x1 is independent of x2 and x3 .
3
3
LINEAR ALGEBRA
28
kxkp :=
n
X
i=1
jxi jp
!1=p
; for 1 p 1:
kxk1 :=
kxk2 :=
n
X
i=1
jxi j;
v
u n
uX
t
i=1
jxi j2 ;
kxk1 := 1max
jx j:
in i
Clearly, norm is an abstraction and extension of our usual concept of length in 3dimensional Euclidean space. So a norm of a vector is a measure of the vector \length",
for example kxk2 is the Euclidean distance of the vector x from the origin. Similarly,
we can introduce some kind of measure for a matrix.
Let A = [aij ] 2 C mn , then the matrix norm induced by a vector p-norm is dened
as
kAxk
kAkp := sup kxk p :
x6=0
29
n
X
j =1
The matrix norms induced by vector p-norms are sometimes called induced p-norms.
This is because kAkp is dened by or induced from a vector p-norm. In fact, A can
be viewed as a mapping from a vector space C n equipped with a vector norm kkp to
another vector space C m equipped with a vector norm kkp . So from a system theoretical
point of view, the induced norms have the interpretation of input/output amplication
gains.
We shall adopt the following convention throughout the book for the vector and matrix
norms unless specied otherwise: let x 2 C n and A 2 C mn , then we shall denote the
Euclidean 2-norm of x simply by
kxk := kxk2
and the induced 2-norm of A by
kAk := kAk2 :
The Euclidean 2-norm has some very nice properties:
3. kxk kyk i there is a matrix 2 Fnm with kk 1 such that x = y.
Furthermore, kxk < kyk i kk < 1.
4. kUxk = kxk for any appropriately dimensioned unitary matrices U .
Another often used matrix norm is the so called Frobenius norm. It is dened as
p
kAkF := Trace(A A) =
v
u m n
uX X
t
i=1 j =1
jaij j2 :
LINEAR ALGEBRA
30
1. (A) kAk (This is also true for F norm and any induced matrix norm).
U=
"
p12
; p12
UA =
p2
0
p12
p12
0
0
75 =: [Aij ];
and let each Aij be an appropriately dimensioned matrix. Then for any induced matrix
p-norm
2
3
(2:2)
31
Proof. It is obvious that if the F -norm is used, then the right hand side of inequality (2.2) equals the left hand side. Hence only the induced p-norm cases, 1 p 1,
will be shown. Let a vector x be partitioned consistently with A as
2x
66 x12
x = 64 ..
.
xq
3
77
75 ;
2 kx1k
6 kx kp
2
kxkp =
664 .. p
.
kxq kp
3
77
75
:
p
Then
2 Pq A1j xj 3
6 Pjq=1 A x 7
2j j 7
k[Aij ]kp := sup k[Aij ] xkp = sup
664 j=1..
75
kxkp =1
kxkp=1
P .
qj=1 Amj xj
p
3
2
Pq
6
j=1 A1j xj
p 7
2 Pq kA1j k kxj k
P
66
q A2j xj
77
6 Pjq=1 kA kp kx kp
j
=1
sup
66 j=1 2.j p j p
p 7
= sup
666
7
.
..
77
kxkp=1
4 P
kxkp =1
6
..
q
4
Pq
j=1 kAmj kp kxj kp
5
j=1 Amj xj
p
p
2 kA11 k kA12k kA1q k 3 2 kx1k 3
6 kA kp kA kp kA kp 7 6 kx kp 7
21 p
22 p
2q p 7 6 2 p 7
= sup
664 ..
..
.. 75 64 .. 75
kxkp =1
.
.
.
kAm1 kp kAm2kp kA1.q kp kxq kp
p
h
i
sup
kAij kp
p kxkp
kxkp =1
h
i
=
kAij k
:
3
77
75
p
p p
LINEAR ALGEBRA
32
and
2 0
66 01 2
1 = 64 .. .. . .
.
. .
0
0
0
..
.
0 p
3
77
75
1 2 p 0; p = minfm; ng:
Proof. Let = kAk and without loss of generality assume m n. Then from the
denition of kAk, there exists a z 2 Fn such that
kAz k = kz k :
By Lemma 2.8, there is a matrix U~ 2 F mn such that U~ U~ = I and
~
Az = Uz:
Now let
We have Ax = y. Let
and
~
x = kzz k 2 Fn ; y =
Uz
2 Fm :
~
Uz
V = x V1 2 Fnn
U = y U1 2 Fmm
be unitary.4 Consequently, U AV has the following structure:
w
A1 := U AV = 0 B
4 Recall that it is always possible to extend an orthonormal set of vectors to an orthonormal basis
for the whole space.
33
U AV = :
The i is the i-th singular value of A, and the vectors ui and vj are, respectively,
the i-th left singular vector and the j -th right singular vector. It is easy to verify that
Avi = i ui
A ui = i vi :
The above equations can also be written as
A Avi = i2 vi
AA ui = i2 ui :
Hence i2 is an eigenvalue of AA or A A, ui is an eigenvector of AA , and vi is an
eigenvector of A A.
The following notations for singular values are often adopted:
LINEAR ALGEBRA
34
It is easy to see that A maps a unit disk to an ellipsoid with semi-axes of 1 and 2 .
Hence it is often convenient to introduce the following alternative denitions for the
largest singular value :
(A) := kmax
kAxk
xk=1
and for the smallest singular value of a tall matrix:
(A) := kmin
kAxk :
xk=1
Proof.
(i) By denition
(A + ) := kmin
k(A + )xk
xk=1
kmin
fkAxk ; kxkg
xk=1
kmin
kAxk ; kmax
kxk
xk=1
xk=1
= (A) ; ():
Hence ;() (A + ) ; (A). The other inequality (A + ) ; (A) ()
follows by replacing A by A + and by ; in the above proof.
(ii) This follows by noting that
(A) := kmin
kAxk
rxk=1
min x A Ax
=
kxk=1
(A) kmin
kxk
xk=1
= (A) ():
35
2
Some useful properties of SVD are collected in the following lemma.
Lemma 2.13 Let A 2 Fmn and
1 2 r > r+1 = = 0; r minfm; ng:
Then
1. rank(A) = r;
2. KerA = spanfvr+1; : : : ; vn g and (KerA)? = spanfv1 ; : : : ; vr g;
3. ImA = spanfu1; : : : ; ur g and (ImA)? = spanfur+1; : : : ; um g;
4. A 2 Fmn has a dyadic expansion:
A=
5.
6.
7.
8.
r
X
i=1
i ui vi = Ur r Vr
Proof. We shall only give a proof for part 8. It is easy to see that rank(Ak ) k and
kA ; Ak k = k+1 . Hence, we only need to show that min kA ; B k k+1 . Let
rank(B)k
B be any matrix such that rank(B ) k. Then
kA ; B k = k
U V ; B k = k ; U BV k
Ik+1 0 ( ; U BV ) Ik0+1
=
k+1 ; B^
where B^ = Ik+1 0 U BV Ik0+1 2 F(k+1)(k+1) and rank(B^ ) k. Let x 2 Fk+1
^ = 0 and kxk = 1. Then
be such that Bx
kA ; B k
k+1 ; B^
(k+1 ; B^ )x
= kk+1 xk k+1 :
Since B is arbitrary, the conclusion follows.
LINEAR ALGEBRA
36
;1
+ = 0r 00 :
37
B
C
0 ; C1 = V2 0
where B1 and C1 are full row rank. Then B1 and C1 have the same number of rows
and V3 := B1 C1 (C1 C1 );1 satises V3 V3 = I since B B = C C . Hence V3 is a unitary
matrix and V3 B1 = C1 . Finally let
U = V1 V03 V0 V2
4
for any suitably dimensioned V4 such that V4 V4 = I .
2
by
We can dene square root for a positive semi-denite matrix A, A1=2 = (A1=2 ) 0,
A = A1=2 A1=2 :
However, A;1=2 BA;1=2 and BA;1 are similar, and hence i (BA;1 ) = i (A;1=2 BA;1=2 ).
Therefore, the conclusion follows by the fact that
0 < I ; A;1=2 BA;1=2
i (A;1=2 BA;1=2 ) < 1 i (BA;1 ) < 1.
2
X
11
X12
X12 :
X22
LINEAR ALGEBRA
38
X22 = U1 U2
1
0
0 0
U
1
U2
I
X
I
X12
0
0
11
0 U
X12 X22
0 U 0
gives X12 U2 = 0. Hence, Ker X22 Ker X12 and now
X12 X22+ X22 = X12 U1 U1 = X12 U1 U1 + X12 U2 U2 = X12 :
The factorization follows easily.
X B
min
X
C A
(2:3)
X B
The matrix
is a dilation of its sub-matrices as indicated in the following
diagram:
C A
X B d
C A
C A
d
c
B
A
A
39
In this diagram, \c" stands for the operation of compression and \d" stands for dilation.
Compression is always norm non-increasing and dilation is always norm non-decreasing.
Sometimes dilation can be made to be norm preserving. Norm preserving dilations are
the focus of this section.
The simplest matrix dilation problem occurs when solving
X
min
X
A
:
(2:4)
Although (2.4) is a much simplied version of (2.3), we will see that it contains all the
essential features of the general problem. Letting
0 denote the minimum norm in (2.4),
it is immediate that
0 = kAk :
The following theorem characterizes all solutions to (2.4).
Theorem 2.17 8 0,
X
A
X = Y ( 2 I ; A A)1=2 :
Proof.
i
i
X
A
X X + A A 2 I
X X (
2 I ; A A):
Now suppose X X (
2 I ; A A) and let
Y := X ( 2 I ; A A)1=2
i+
LINEAR ALGEBRA
40
Remark 2.1 The theorem still holds if (
2I ; A A)1=2 is replaced by any matrix R
such that
2 I ; A A = R R.
~
A more restricted version of the above theorem is shown in the following corollary.
X
A
(<
)
2 ;1=2
X (
I ; A A)
1(< 1):
Theorem 2.19 8 0
X A
X = ( 2 I ; AA )1=2 Y:
X A
(<
)
2
(
I ; AA );1=2X
1 (< 1):
X B
0 := min
X
C A
:
(2:5)
The following so called Parrott's theorem will play an important role in many control
related optimization problems. The proof is the straightforward application of Theorem 2.17 and its dual, Theorem 2.19.
0 = max
C A
;
B
A
:
(2:6)
41
Proof. Denote by
^ the right hand side of the equation (2.6). Clearly,
0
^ since
compressions are norm non-increasing, and that
0
^ will be shown by using Theorem 2.17 and Theorem 2.19.
Suppose A 2 C nm and n m (the case for m > n can be shown in the same
fashion). Then A has the following singular value decomposition:
m
A=U 0
V ; U 2 C nn ; V 2 C mm :
n;m;m
Hence
^2 I ; A A = V (^ 2 I ; 2m )V
and
^2 I ; AA
Now let
=U
^2I ; 2
0
^2 In;m U :
and
2
2 1=2
0
(^
2 I ; AA )1=2 := U (^
I ;0m )
^In;m U :
Then it is easy to verify that
;A
(^
2 I ; A A)1=2
2
1
=
2
(^
I ; AA )
A
=
^2I
0
;A
(^
2 I ; A A)1=2
2
1
=
2
(^
I ; AA )
A
0
^2 I :
LINEAR ALGEBRA
42
Thus
^
0 , so
^ =
0 .
This theorem gives one solution to (2.3) and an expression for
0 . As in (2.4), there
may be more than one solution to (2.3), although the proof of theorem 2.21 exhibits
only one. Theorem 2.22 considers the problem of parameterizing all solutions. The
solution X^ = ;Y A Z is the \central" solution analogous to X = 0 in (2.4).
Theorem 2.22 Suppose
0. The solutions X such that
X B
C A
(2:7)
Proof. Since
0, again from Theorem 2.19 there exists a Z with kZ k 1 such
that
C = (
2 I ; AA )1=2 Z:
Note that using the above expression for C we have
2I ; C A C A
(I ; Z Z )1=2
Z Z )1=2
0
0
=
(I ;
;A Z
(
2 I ; A A)1=2
;A Z
(
2 I ; A A)1=2 :
Now apply
2.17 (Remark 2.1) to inequality (2.7) with respect to the partitioned
XTheorem
B
matrix C A to get
X B = W^
(I ; Z Z )1=2
0
;A Z
(
2 I ; A A)1=2
for some contraction W^ ,
W^
1. Partition W^ as W^ = W1 Y to obtain the
expression for X and B :
X = ;Y A Z +
W1 (I ; Z Z )1=2 ;
B = Y (
2 I ; A A)1=2 :
Then kY k 1 and the theorem follows by noting that
W1 Y
W , kW k 1, such that W1 = (I ; Y Y )1=2 W .
1 i there is a
The following corollary gives an alternative version of Theorem 2.22 when > 0 .
43
X B
C A
(<
)
(2:11)
(I ; Y Y );1=2 (X + Y A Z )(I ; Z Z );1=2
(<
)
(2:12)
where
Y = B (
2 I ; A A);1=2 ;
(2.13)
2
;
1
=
2
Z = (
I ; AA ) C:
(2.14)
Note that in the case of
>
0 , I ; Y Y and I ; Z Z are invertible since Corollary 2.18 and 2.20 clearly show that kY k < 1 and kZ k < 1. There are many alternative
characterizations of solutions to (2.11), although the formulas given above seem to be
the simplest.
As a straightforward application of the dilation results obtained above, consider the
following matrix approximation problem:
0 = min
kR + UQV k
Q
(2:15)
V
Proof. Let U? and V? be such that U U? and V? are unitary matrices.
Then
0 = min
kR + UQV k
Q
U U? (R + UQV ) VV
= min
Q
U RV + Q U RV
?
?
:
= min
Q
U? RV
U? RV?
44
LINEAR ALGEBRA
x_ = Ax + Bu; x(t0 ) = x0
y = Cx + Du
(3.1)
(3.2)
where x(t) 2 Rn is called the system state, x(t0 ) is called the initial condition of the
system, u(t) 2 Rm is called the system input, and y(t) 2 Rp is the system output.
The A; B; C; and D are appropriately dimensioned real constant matrices. A dynamical
system with single input (m = 1) and single output (p = 1) is called a SISO (single input
and single output) system, otherwise it is called MIMO (multiple input and multiple
45
46
A B
;1
C D := C (sI ; A) B + D
will be used. Other reasons for using this notation will be discussed in Chapter 10.
Note that
A B
C D
Zt
the state of a linear system at one time can be obtained from the state at another
through the transition matrix, we can assume without loss of generality that t0 = 0.
This will be assumed in the sequel.
The impulse matrix of the dynamical system is dened as
g(t) = L;1 fG(s)g = CeAt B 1+ (t) + D(t)
where (t) is the unit impulse and 1+ (t) is the unit step dened as
1; t 0;
1+ (t) := 0; t < 0:
47
The input/output relationship (i.e., with zero initial state: x0 = 0) can be described by
the convolution equation
y(t) = (g u)(t) :=
Z1
;1
Zt
;1
Denition 3.1 The dynamical system described by the equation (3.1) or the pair
(A; B ) is said to be controllable if, for any initial state x(0) = x0 , t1 > 0 and nal
state x1 , there exists a (piecewise continuous) input u() such that the solution of (3.1)
satises x(t1 ) = x1 . Otherwise, the system or the pair (A; B ) is said to be uncontrollable.
The controllability (or observability introduced next) of a system can be veried through
some algebraic or geometric criteria.
Wc (t) :=
Zt
0
eA BB eA d
C = B AB A2 B : : : An;1 B
P
has full row rank or, in other words, hA jImB i := ni=1 Im(Ai;1 B ) = Rn .
(iv) The matrix [A ; I; B ] has full row rank for all in C .
(v) Let and x be any eigenvalue and any corresponding left eigenvector of A, i.e.,
x A = x , then x B 6= 0.
(vi) The eigenvalues of A + BF can be freely assigned (with the restriction that complex
eigenvalues are in conjugate pairs) by a suitable choice of F .
48
Proof.
(i) , (ii): Suppose Wc (t1 ) > 0 for some t1 > 0, and let the input be dened as
u( ) = ;B eA(t1 ; )Wc (t1 );1 (eAt1 x0 ; x1 ):
Then it is easy to verify using the formula in (3.3) that x(t1 ) = x1 . Since x1 is
arbitrary, the pair (A; B ) is controllable.
To show that the controllability of (A; B ) implies that Wc (t) > 0 for any t > 0,
assume that
(A; B ) is controllable but Wc (t1 ) is singular for some t1 > 0. Since
eAtBB eAt 0 for all t, there exists a real vector 0 6= v 2 Rn such that
v eAtB = 0; t 2 [0; t1 ]:
Now let x(t1 ) = x1 = 0, and then from the solution (3.3), we have
0 = eAt1 x(0) +
Zt
49
A ; I B
does not have full row rank for some 2 C . Then there exists a vector x 2 C n
such that
x A ; I B = 0
i.e., x A = x and x B = 0. However, this will result in
x = x0
c
50
Denition 3.3 The dynamical system (3.1), or the pair (A; B), is said to be stabilizable
if there exists a state feedback u = Fx such that the system is stable, i.e., A + BF is
stable.
Therefore, it is more appropriate to call this stabilizability the state feedback stabilizability to dierentiate it from the output feedback stabilizability dened later.
The following theorem is a consequence of Theorem 3.1.
We now consider the dual notions of observability and detectability of the system
described by equations (3.1) and (3.2).
Denition 3.4 The dynamical system described by the equations (3.1) and (3.2) or
by the pair (C; A) is said to be observable if, for any t1 > 0, the initial state x(0) = x0
can be determined from the time history of the input u(t) and the output y(t) in the
interval of [0; t1]. Otherwise, the system, or (C; A), is said to be unobservable.
Wo (t) :=
is positive denite for any t > 0.
(iii) The observability matrix
Zt
0
eA C CeA d
2 C 3
66 CA 77
O = 666 CA. 2 777
4 .. 5
CAn;1
T
has full column rank or ni=1 Ker(CAi;1 ) = 0.
(iv) The matrix
A ; I
C
51
(v) Let and y be any eigenvalue and any corresponding right eigenvector of A, i.e.,
Ay = y, then Cy 6= 0.
(vi) The eigenvalues of A + LC can be freely assigned (with the restriction that complex
eigenvalues are in conjugate pairs) by a suitable choice of L.
(vii) (A ; C ) is controllable.
Proof. First, we will show the equivalence between conditions (i) and (iii). Once this
is done, the rest will follow by the duality or condition (vii).
(i) ( (iii): Note that given the input u(t) and the initial condition x0 , the output in
the time interval [0; t1] is given by
Zt
0
Since y(t) and u(t) are known, there is no loss of generality in assuming u(t) =
0; 8t. Hence,
y(t) = CeAt x(0); t 2 [0; t1 ]:
From this equation, we have
3 2 C 3
CA 77
77 666 CA
75 = 66 . 2 777 x(0)
4 .. 5
y(n;1) (0)
2 y(0)
66 y_ (0)
64 ...
CAn;1
where y(i) stands for the i-th derivative of y. Since the observability matrix O
has full column rank, there is a unique solution x(0) in the above equation. This
completes the proof.
(i) ) (iii): This will be proven by contradiction. Assume that (C; A) is observable but
that the observability matrix does not have full column rank, i.e., there is a vector
x0 such that Ox0 = 0 or equivalently CAi x0 = 0; 8i 0 by the Cayley-Hamilton
Theorem. Now suppose the initial state x(0) = x0 , then y(t) = CeAt x(0) = 0.
This implies that the system is not observable since x(0) cannot be determined
from y(t) 0.
Denition 3.5 The system, or the pair (C; A), is detectable if A + LC is stable for
some L.
52
A ; I
C
The conditions (iv) and (v) of Theorem 3.1 and Theorem 3.3 and the conditions
(ii) and (iii) of Theorem 3.2 and Theorem 3.4 are often called Popov-Belevitch-Hautus
(PBH) tests. In particular, the following denitions of modal controllability and observability are often useful.
Denition 3.6 Let be an eigenvalue of A or, equivalently, a mode of the system.
Then the mode is said to be controllable (observable) if x B 6= 0 (Cx 6= 0) for all left
(right) eigenvectors of A associated with , i.e., x A = x (Ax = x) and 0 6= x 2 C n .
Otherwise, the mode is said to be uncontrollable (unobservable).
It follows that a system is controllable (observable) if and only if every mode is controllable (observable). Similarly, a system is stabilizable (detectable) if and only if every
unstable mode is controllable (observable).
For example, consider the following 4th order system:
2 1 0 0 03
A B 66 01 1 1 0 1 77
6
7
C D = 64 00 00 01 02 1 75
0
with 1 6= 2 . Then, the mode 1 is not controllable if = 0, and 2 is not observable
if = 0. Note that if 1 = 2 , the system is uncontrollable and unobservable for any
and since in that case, both
1
203
6 7
x1 = 64 01 75
0
and
203
6 7
x2 = 64 00 75
1
53
with the last two rows independent of each other, then the system is controllable even
if 1 = 2 . For example, the reader may easily verify that the system with
20 03
6
7
B = 64 1 01 75
1 0
is controllable for any .
In general, for a system given in the Jordan canonical form, the controllability and
observability can be concluded by inspection. The interested reader may easily derive some explicit conditions for the controllability and observability by using Jordan
canonical form and the tests (iv) and (v) of Theorem 3.1 and Theorem 3.3.
54
c
x_
c
x_ c
y =
A A x B
c
12
c
c
0 Ac
xc + 0 u
C C xc + Du
c
c
xc
Proof. Since rank C = k1 < n, the pair (A; B) is not controllable. Now let q1; q2; : : : ; qk1
be any linearly independent columns of C . Let qi ; i = k1 +1; : : : ; n be any n ; k1 linearly
independent vectors such that the matrix
Q := q1 qk1 qk1 +1 qn
is nonsingular. Dene
T := Q;1 :
Then the transformation x = Tx will give the desired decomposition. To see that,
note that for each i = 1; 2; : : : ; k1 , Aqi can be written as a linear combination of qi ; i =
1; 2; : : : ; k1 since Aqi is a linear combination of the columns of C by the Cayley-Hamilton
Theorem. Therefore, we have
AT ;1 =
=
Aq Aq Aq
1
k
k +1 Aqn
A A
q q q
c
12
1
k
k +1 qn
0 Ac
A A
= T ;1
12
Ac
55
for some k1 k1 matrix Ac . Similarly, each column of the matrix B is a linear combination of qi ; i = 1; 2; : : :; k1 , hence
B = Q B0c = T ;1 B0c
for some Bc 2 C k1 m .
To show that (Ac ; Bc) is controllable, note that rank C = k1 and
B A B Ak1 ;1 B An;1 B
c
c c
c
c :
;
1
c
c
C=T
0
0
0
0
Since, for each j k1 , Ajc is a linear combination of Aic ; i = 0; 1; : : : ; (k1 ; 1) by CayleyHamilton theorem, we have
rank Bc Ac Bc Akc 1 ;1 Bc = k1 ;
i.e., (Ac ; Bc ) is controllable.
2
A numerically reliable way to nd a such transformation T is to use QR factorization.
For example, if the controllability matrix C has the QR factorization QR = C , then
T = Q;1 .
Corollary 3.7 If the system is stabilizable and the controllability matrix C has rank
k1 < n, then there exists a similarity transformation T such that
3
2
TAT ;1 TB 6 Ac A12 Bc 7
0 A 0 5
CT ;1 D = 4 c
Cc Cc D
with Ac 2 C k1 k1 , (Ac ; Bc ) controllable and with Ac stable.
Hence, the state space x is partitioned into two orthogonal subspaces
x
0
c
and
0
xc
with the rst subspace controllable from the input and second completely uncontrollable
from the input (i.e., the state xc are not aected by the control u). Write these subspaces
in terms of the original coordinates x, we have
x
;
1
x = T x = q1 qk1 qk1 +1 qn xc :
c
So the controllable subspace is the span of qi ; i = 1; : : : ; k1 or, equivalently, Im C . On the
other hand, the uncontrollable subspace is given by the complement of the controllable
subspace.
By duality, we have the following decomposition if the system is not completely
observable.
56
Theorem 3.8 If the observability matrix O has rank k2 < n, then there exists a similarity transformation T such that
3
2
TAT ;1 TB 6 Ao 0 Bo 7
A A B
CT ;1 D = 4 21 o o 5
Co
Corollary 3.9 If the system is detectable and the observability matrix C has rank k2 <
n, then there exists a similarity transformation T such that
2
3
TAT ;1 TB 6 Ao 0 Bo 7
A A B
CT ;1 D = 4 21 o o 5
Co
with Ao 2 C k2 k2 , (Co ; Ao ) observable and with Ao stable.
Similarly, we have
Theorem 3.10 Let an LTI dynamical system be described by the equations (3.1) and (3.2).
Then there exists a nonsingular coordinate transformation x = Tx such that
2 x_ 3
2 A 0 A 0 3 2 x 3 2 B 3
co
66 x_ co 77 = 66 A21co Aco A1323 A24 77 66 xcoco 77 + 66 Bcoco 77 u
4 x_ co 5
4 0 0 Aco 0 5 4 xco 5 4 0 5
_xco
0 0 A43 Aco
0
2 x 3 xco
co
6
7
x
6
y = Cco 0 Cco 0 4 xco 75 + Du
co
xco
or equivalently
TAT ;1
CT ;1
2 Aco 0
6 A A
TB = 66 021 0co
64 0 0
D
Cco
A13 0 Bco
A23 A24 Bco
Aco 0 0
A43 Aco 0
Cco 0 D
3
77
77
5
57
where the vector xco is controllable and observable, xco is controllable but unobservable, xco is observable but uncontrollable, and xco is uncontrollable and unobservable.
Moreover, the transfer matrix from u to y is given by
G(s) = Cco (sI ; Aco );1 Bco + D:
One important issue is that although the transfer matrix of a dynamical system
A B
C D
their internal behaviors are very dierent. In other words, while their input/output
behaviors are the same, their state space response with nonzero initial conditions are
very dierent. This can be illustrated by the state space response for the simple system
2 x_ 3
2 A 0 0 0 3 2 x 3 2 B 3
co
66 x_ co 77 = 66 0co Aco 0 0 77 66 xcoco 77 + 66 Bcoco 77 u
4 x_ co 5
4 0 0 Aco 0 5 4 xco 5 4 0 5
_xco
0 0 0 Aco
0
2 x 3 xco
co
6
7
x
6
y = Cco 0 Cco 0 4 xco 75
co
xco
with xco controllable and observable, xco controllable but unobservable, xco observable
but uncontrollable, and xco uncontrollable and unobservable.
The solution to this system is given by
2 x (t) 3
66 xcoco(t) 77
4 x (t) 5
co
2 Acot
R
e xco (0) + R0t eAco (t; )Bco u( )d
66 eAcotxco(0) + t eAco(t; )Bcou( )d
0
64
eAco t xco (0)
xco(t)
eAcot xco(0)
y(t) = Cco xco (t) + Cco xco ;
3
77
75
note that xco (t) and xco(t) are not aected by the input u, while xco(t) and xco(t) do
not show up in the output y. Moreover, if the initial condition is zero, i.e., x(0) = 0,
then the output
Zt
y(t) = Cco eAco(t; ) Bcou( )d:
0
58
However, if the initial state is not zero, then the response xco (t) will show up in the
output. In particular, if Aco is not stable, then the output y(t) will grow without bound.
The problems with uncontrollable and/or unobservable unstable modes are even more
profound than what we have mentioned. Since the states are the internal signals of the
dynamical system, any unbounded internal signal will eventually destroy the system.
On the other hand, since it is impossible to make the initial states exactly zero, any
uncontrollable and/or unobservable unstable mode will result in unacceptable system
behavior. This issue will be exploited further in section 3.7. In the next section, we will
consider how to place the system poles to achieve desired closed-loop behavior if the
system is controllable.
y ?e
R
6
e
u e v
59
Lemma 3.11 Let F be a constant matrix with appropriate dimension; then (A; B) is
u = Fx = ;1 ;1 x;
the system becomes
A + BF B 2 0
4
C + DF D = 1
0 1
0 0
1 0 0
3
5
x_ = Ax + Bu 7;! x_ = Ax + Bu + Ly
is called output injection which can be written as
A B A + LC B + LD
:
C D 7;!
C
D
By duality, the output injection does not change the system observability (detectability)
but may change the system controllability (stabilizability).
Remark 3.1 We would like to call attention to the diagram and the signals ow con-
vention used in this book. It may not be conventional to let signals
ow from the right
to the left, however, the reader will nd that this representation is much more visually
appealing than the traditional representation, and is consistent with the matrix manipulations. For example, the following diagram represents the multiplication of three
matrices and will be very helpful in dealing with complicated systems:
z = M1 M2 M3 w:
z
M1
M2
M3
w
The conventional signal block diagrams, i.e., signals
owing from left to right, will also
be used in this book.
~
60
We now consider some special state space representations of the dynamical system
described by equations (3.1) and (3.2). First, we will consider the systems with single
inputs.
Assume that a single input and multiple output dynamical system is given by
A b
G(s) = C d ; b 2 Rn ; C 2 Rpn ; d 2 Rp
2 ;a ;a ;a ;a 3
213
1
2
n;1
n
66 1 0 0
66 0 77
0 77
6
7
1
0
0 7 b1 := 66 0 77
A1 := 66 0.
64 ... 75
.
.
.. 75
..
4 .. ..
.
0
and
C = b Ab An;1 b
C1 = b1 A1 b1 An1 ;1 b1 :
Then it is easy to verify that both C and C1 are nonsingular. Moreover, the transformation
Tc = C1 C ;1
T AT ;1 T b A b
1
c c
c =
1
;1
;1
CTc
where
CTc
CTc;1 = 1 2 n;1 n
for some i 2 Rp . This state space representation is usually called controllable canonical
form or controller canonical form. It is also easy to show that the transfer matrix is
given by
n;1 2 sn;2 + + n;1 s + n
G(s) = C (sI ; A);1 b + d = 1 ssn + +
a1 sn;1 + + an;1 s + an + d;
which also shows that, given a column vector of transfer matrix G(s), a state space representation of the transfer matrix can be obtained as above. The quadruple (A; b; C; d)
is called a state space realization of the transfer matrix G(s).
61
x_ = A1 x + b1 u
and a state feedback control law
u = Fx = f1 f2 : : : fn;1 fn x:
Then the closed-loop system is given by
x_ = (A1 + b1 F )x
and det(I ; (A1 + b1 F )) = n +(a1 ; f1)n;1 + +(an ; fn ). It is clear that the zeros
of det(I ; (A1 + b1 F ) can be made arbitrary for an appropriate choice of F provided
that the complex zeros appear in conjugate pairs, thus showing that the eigenvalues of
A + bF can be freely assigned if (A; b) is controllable.
Dually, consider a multiple input and single output system
G(s) = Ac Bd ; B 2 Rnm ; c 2 Rn ; d 2 Rm ;
and assume that (c; A) is observable; then there is a transformation To such that
T AT ;1
o
cTo;1
2 ;a1
66 ;a2
ToB = 666 ...
d
66 ;an;1
4 ;an
1
and
1
0
..
.
0
0
0
0
1
..
.
0
0
0
n;1
0 1 3
0 2 77
. . 7
.. ..
1 n;1
0 n
0 d
77 ; i 2 Rm
77
5
n;2
n
G(s) = c(sI ; A);1 B + d = 1 ssn + a+ sn2;s 1 + + + a+ n;s 1+s +
a + d:
1
n;1
Lemma 3.12 If an m input system pair (A; B) is controllable and if A is cyclic, then
for almost all v 2 Rm , the single input pair (A; Bv) is controllable.
62
Proof. Without loss of generality, assume that A is in the Jordan canonical form and
that the matrix B is partitioned accordingly:
2J
66 1 J2
A = 64
...
2 i 1
66
i
6
Ji = 66
64
Jk
3
2B
77
66 B12
75 B = 64 ...
Bk
...
... ...
i 1
i
3
77
75
3
77
77
77
5
and i 6= j if i 6= j . By PBH test, the pair (A; B ) is controllable if and only if, for each
i = 1; : : : ; k, the last row of Bi is not zero. Let bi 2 Rm be the last row of Bi , and then
we only need to show that, for almost all v 2 Rm , bi v 6= 0 for each i = 1; : : : ; k which is
clear since for each i, the set v 2 Rm such that bi v = 0 has measure zero in Rm since
bi 6= 0.
2
The cyclicity assumption in this theorem is essential. Without this assumption, the
theorem does not hold. For example, the pair
1 0
1 0
A= 0 1 ; B= 0 1
is controllable but there is no v 2 R2 such that (A; Bv) is controllable since A is not
cyclic.
Since a matrix A with distinct eigenvalues is cyclic, by denition we have the following lemma.
Lemma 3.13 If (A; B) is controllable, then for almost any K 2 Rmn , all the eigenvalues of A + BK are distinct and, consequently, A + BK is cyclic.
A proof can be found in Brasch and Pearson [1970], Davison [1968], and Heymann
[1968].
Now it is evident that given a multiple input controllable pair (A; B ), there is a
matrix K 2 Rmn and a vector v 2 Rm such that A + BK is cyclic and (A + BK; Bv)
is controllable. Moreover, from the pole placement results for the single input system,
there is a matrix f 2 Rn so that the eigenvalues of (A + BK )+(Bv)f can be arbitrarily
assigned. Hence, the eigenvalues of A + BF can be arbitrarily assigned by choosing a
state feedback in the form of
u = Fx = (K + vf )x:
63
q_ = Mq + Nu + Hy
x^ = Qq + Ru + Sy
so that x^(t) ; x(t) ! 0 as t ! 1 for all initial states x(0), q(0) and for every input u().
Theorem 3.14 An observer exists i (C; A) is detectable. Further, if (C; A) is detectable, then a full order Luenberger observer is given by
q_ = Aq + Bu + L(Cq + Du ; y)
x^ = q
where L is any matrix such that A + LC is stable.
(3.5)
(3.6)
Proof. We rst show that the detectability of (C; A) is sucient for the existence
of an observer. To that end, we only need to show that the so-called Luenberger
observer dened in the theorem is indeed an observer. Note that equation (3.5) for q
is a simulation of the equation for x, with an additional forcing term L(Cq + Du ; y),
which is a gain times the output error. Equivalent equations are
q_ = (A + LC )q + Bu + LDu ; Ly
x^ = q:
64
These equations have the form allowed in the denition of an observer. Dene the error,
e := x^ ; x, and then simple algebra gives
e_ = (A + LC )e;
therefore e(t) ! 0 as t ! 1 for every x(0), q(0), and u(:).
To show the converse, assume that (C; A) is not detectable. Take the initial state
x(0) in the undetectable subspace and consider a candidate observer:
q_ = Mq + Nu + Hy
x^ = Qq + Ru + Sy:
Take q(0) = 0 and u(t) 0. Then the equations for x and the candidate observer are
x_ = Ax
q_ = Mq + HCx
x^ = Qq + SCx:
Since an unobservable subspace is an A-invariant subspace containing x(0), it follows
that x(t) is in the unobservable subspace for all t 0. Hence, Cx(t) = 0 for all t 0,
and, consequently, q(t) 0 and x^(t) 0. However, for some x(0) in the undetectable
subspace, x(t) does not converge to zero. Thus the candidate observer does not have
the required property, and therefore, no observer exists.
2
The above Luenberger observer has dimension n, which is the dimension of the state
x. It's possible to get an observer of lower dimension. The idea is this: since we can
measure y ; Du = Cx, we already know x modulo Ker C , so we only need to generate
the part of x in Ker C . If C has full row rank and p := dim y, then the dimension of
Ker C equals n ; p, so we might suspect that we can get an observer of order n ; p.
This is true. Such an observer is called a \minimal order observer". We will not pursue
this issue further here. The interested reader may consult Chen [1984].
Recall that, for a dynamical system described by the equations (3.1) and (3.2), if
(A; B ) is controllable and state x is available for feedback, then there is a state feedback
u = Fx such that the closed-loop poles of the system can be arbitrarily assigned.
Similarly, if (C; A) is observable, then the system observer poles can be arbitrarily
placed so that the state estimator x^ can be made to approach x arbitrarily fast. Now
let us consider what will happen if the system states are not available for feedback
so that the estimated state has to be used. Hence, the controller has the following
dynamics:
x^_ = (A + LC )^x + Bu + LDu ; Ly
u = F x^:
65
x_ A
=
BF
;LC A + BF + LC
x^_
e_ A + LC
=
x^_
0
A + BF
;LC
x
x^ :
e
x^ ;
and the closed-loop poles consist of two parts: the poles resulting from state feedback
(A + BF ) and the poles resulting from the state estimation (A + LC ). Now if (A; B ) is
controllable and (C; A) is observable, then there exist F and L such that the eigenvalues
of A + BF and A + LC can be arbitrarily assigned. In particular, they can be made to
be stable. Note that a slightly weaker result can also result even if (A; B ) and (C; A)
are only stabilizable and detectable.
The controller given above is called an observer-based controller and is denoted as
u = K (s)y
and
A + BF + LC + LDF ;L
K (s) =
:
F
A B
;
G=
C D
y
u
66
Suppose that G1 and G2 are two subsystems with state space representations:
B1
G1 = AC1 D
1
1
B2 :
G2 = AC2 D
2
2
Then the series or cascade connection of these two subsystems is a system with the
output of the second subsystem as the input of the rst subsystem as shown in the
following diagram:
G1
G2
This operation in terms of the transfer matrices of the two subsystems is essentially the
product of two transfer matrices. Hence, a representation for the cascaded system can
be obtained as
G1 G2 =
=
A B A B
2
1
2
1
C1 D1
C2 D2
2 A BC BD 3 2 A 0
1
1 2
2
1 2
4 0 A2
B2 5 = 4 B1 C2 A1
C1 D1 C2
B2
B1 D2
D1 D2
3
5:
D1 C2 C1
Similarly, the parallel connection or the addition of G1 and G2 can be obtained as
3
A B A B 2 A1 0
B1
B2 5 :
G1 + G2 = C1 D1 + C2 D2 = 4 0 A2
1
1
2
2
C1 C2 D1 + D2
Next we consider a feedback connection of G1 and G2 as shown below:
y
r
f
G1
D1 D2
6;
- G2
Then the closed-loop transfer matrix from r to y is given by
2 A ; B D R;1C
;1 3
;B1 R21;1 C2
B1 R21
1
1 2 12 1
;1 C1
;1 C2 B2 D1 R;1 5
B2 R12
A2 ; B2 D1 R21
T =4
;1
;1
;21
1
R12 C1
;R12 D1 C2
D1 R21
where R12 = I + D1 D2 and R21 = I + D2 D1 . Note that these state space representations
may not be necessarily controllable and/or observable even if the original subsystems
G1 and G2 are.
For future reference, we shall also introduce the following denitions.
67
Denition 3.7 The transpose of a transfer matrix G(s) or the dual system is dened
as
G 7;! GT (s) = B (sI ; A );1 C + D
A B A C
7 ! B D :
C D ;
or equivalently
A B ;A ;C
7 ! B D :
C D ;
or equivalently
Proof. The right inverse case will be proven and the left inverse case follows by duality.
Suppose DDy = I . Then
GGy =
=
2 A BDy C
4 0 A ; BDy C
yC
2 CA DD
BDy C
4 0 A ; BDy C
C
C
I I
BDy
;BDy 5
DDy
3
BDy
;BDy 5 :
I
68
B ;
G(s) = CA D
a realization of G(s).
We note that if the transfer matrix is either single input or single output, then the
formulas in Section 3.4 can be used to obtain a controllable or observable realization.
The realization for a general MIMO transfer matrix is more complicated and is the focus
of this section.
Theorem 3.16 A state space realization (A; B; C; D) of G(s) is minimal if and only if
(A; B ) is controllable and (C; A) is observable.
Proof. We shall rst show that if (A; B; C; D) is a minimal realization of G(s), then
(A; B ) must be controllable and (C; A) must be observable. Suppose, on the contrary,
that (A; B ) is not controllable and/or (C; A) is not observable. Then from Kalman
decomposition, there is a smaller dimensioned controllable and observable state space
realization that has the same transfer matrix; this contradicts the minimality assumption. Hence (A; B ) must be controllable and (C; A) must be observable.
Next we show that if an n-th order realization (A; B; C; D) is controllable and observable, then it is minimal. But supposing it is not minimal, let (Am ; Bm ; Cm ; D) be
a minimal realization of G(s) with order k < n. Since
we have
This implies that
CAi B = Cm Aim Bm ; 8i 0:
OC = Om Cm
(3:7)
69
where C and O are the controllability and observability matrices of (A; B ) and (C; A),
respectively, and
Cm :=
Om :=
B A B An;1B
m m
2 mC m 3m
66 CmmAm 77
64 ... 75 :
Cm Anm;1
By Sylvester's inequality,
rank C + rank O ; n rank (OC ) minfrank C ; rank Og;
and, therefore, we have rank (OC ) = n since rank C = rank O = n by the controllability
and observability assumptions. Similarly, since (Am ; Bm ; Cm ; D) is minimal, (Am ; Bm )
is controllable and (Cm ; Am ) is observable. Moreover,
rank Om Cm = k < n;
which is a contradiction since rank OC = rank Om Cm by equality (3.7).
The following property of minimal realizations can also be veried, and this is left
to the reader.
Theorem 3.17 Let (A1; B1; C1 ; D) and (A2 ; B2; C2 ; D) be two minimal realizations of
a real rational transfer matrix G(s), and let C1 , C2 , O1 , and O2 be the corresponding controllability and observability matrices, respectively. Then there exists a unique
nonsingular T such that
G (s) G (s)
2
G(s) = 1
G3 (s) G4 (s)
A B
Gi (s) = Ci Di ; i = 1; : : : ; 4:
i
i
70
Note that Gi (s) may itself be a multiple input and multiple output transfer matrix. In
particular, if Gi (s) is a column or row vector of transfer functions, then the formulas
in Section 3.4 can be used to obtain a controllable or observable realization for Gi (s).
Then a realization for G(s) can be given by
2A 0 0 0 B 0 3
66 01 A2 0 0 01 B2 77
6
7
G(s) = 66 00 00 A03 A04 B03 B04 77 :
64 C C 0 0 D D 75
1
2
1
2
0
0 C3 C4 D3 D4
Alternatively, if the transfer matrix G(s) can be factored into the product and/or
the sum of several simply realized transfer matrices, then a realization for G can be
obtained by using the cascade or addition formulas in the last section.
A problem inherited with these kinds of realization procedures is that a realization
thus obtained will generally not be minimal. To obtain a minimal realization, a Kalman
controllability and observability decomposition has to be performed to eliminate the
uncontrollable and/or unobservable states. (An alternative numerically reliable method
to eliminate uncontrollable and/or unobservable states is the balanced realization method
which will be discussed later.)
We will now describe one factorization procedure that does result in a minimal
realization by using partial fractional expansion (The resulting realization is sometimes
called Gilbert's realization due to Gilbert).
Let G(s) be a p m transfer matrix and write it in the following form:
G(s) = Nd((ss))
with d(s) a scalar polynomial. For simplicity, we shall assume that d(s) has only real
and distinct roots i 6= j if i 6= j and
d(s) = (s ; 1 )(s ; 2 ) (s ; r ):
Then G(s) has the following partial fractional expansion:
G(s) = D +
r W
X
i
s ; i :
i=1
Suppose
rank Wi = ki
and let Bi 2 Rki m and Ci 2 Rpki be two constant matrices such that
Wi = Ci Bi :
71
2I
66 1 k
G(s) = 64
C1
...
3
.. 77
. 7:
Br 5
B1
r Ikr
Cr D
It follows immediately from PBH tests that this realization is controllable and observable. Hence, it is minimal.
An immediate consequence of this minimal realization is that a transfer matrix with
an r-th order polynomial denominator does not necessarily have an r-th order state
space realization unless Wi for each i is a rank one matrix.
This approach can, in fact, be generalized to more complicated cases where d(s) may
have complex and/or repeated roots. Readers may convince themselves by trying some
simple examples.
72
A B
be a state space realization of a (not necessarily stable)
Lemma 3.20 Let
C D
P = P =
with P1 nonsingular such that
P
0
0 0
1
AP + PA + BB = 0:
Now partition the realization (A; B; C; D) compatibly with P as
2A A B 3
4 A1121 A2212 B12 5 :
C1 C2 D
73
A
Then
11
C1
B1
D
1 B1 P1 A21 + B1 B2 ;
0 = AP + PA + BB = A11 PA1 +PP1+A11B +BB
B2 B2
21 1
2 1
which gives B2 = 0 and A21 = 0 since P1 is nonsingular. Hence, part of the realization
is not controllable:
2A A B 3 2A A B 3
4 A2111 A2212 B12 5 = 4 011 A1222 01 5 = AC11 BD1 :
1
C1
C2 D
C1
C2 D
Finally, it follows from Lemma 3.18 that (A11 ; B1 ) is controllable if A11 is stable.
We also have
Lemma 3.21 Let CA DB be a state space realization of a (not necessarily stable)
transfer matrix G(s). Suppose that there exists a symmetric matrix
Q = Q = Q01 00
with Q1 nonsingular such that
QA + A Q + C C = 0:
Now partition the realization (A; B; C; D) compatibly with Q as
2A A B 3
4 A1121 A2212 B12 5 :
C1
Then
A
C2 D
11
C1
B1
D
74
B be a stable realization.
1. Let G(s) = CA D
2. Compute the controllability Gramian P 0 from
AP + PA + BB = 0:
1
3. Diagonalize P to get P = U1 U2
0
U U unitary.
1
2
U AU U B
4. Then G(s) =
CU1
0
0
U1 U2 with 1 > 0 and
is a controllable realization.
2 ;1 ;4= 1 3
G(s) = 4 4 ;2 2 5
;1 2=
where is any nonzero number. It is easy to check that the controllability Gramian of
the realization is given by
0:5
P=
2 :
Since the last diagonal term of P can be made arbitrarily small by making small,
the controllability of the corresponding state can be made arbitrarily weak. If the state
corresponding to the last diagonal term of P is removed, we get a transfer function
;1 1 = ;1
G^ = ;
1 0
s+1
which is not close to the original transfer function in any sense. The problem may be
easily detected if one checks the observability Gramian Q, which is
Q=
0:5
1=2 :
75
Since 1=2 is very large if is small, this shows that the state corresponding to the
last diagonal term is strongly observable. This example shows that controllability (or
observability) Gramian alone can not give an accurate indication of the dominance of
the system states in the input/output behavior.
This motivates the introduction of a balanced realization which gives balanced
Gramians for controllability
A B and observability.
Suppose G = C D is stable, i.e., A is stable. Let P and Q denote the controllability Gramian and observability Gramian, respectively. Then by Lemma 3.18, P and
Q satisfy the following Lyapunov equations
AP + PA + BB = 0
(3:9)
A Q + QA + C C = 0;
(3:10)
and P 0, Q 0. Furthermore, the pair (A; B ) is controllable i P > 0, and (C; A) is
observable i Q > 0.
Suppose the state is transformed by a nonsingular T to x^ = Tx to yield the realization
" ^ ^#
;1 TB
A B
TAT
G = ^ ^ = CT ;1 D :
C D
Then the Gramians are transformed to P^ = TPT and Q^ = (T ;1 ) QT ;1. Note that
P^ Q^ = TPQT ;1, and therefore the eigenvalues of the product of the Gramians are
invariant under state transformation.
Consider the similarity transformation T which gives the eigenvector decomposition
PQ = T ;1 T; = diag(1 ; : : : ; n ):
Then the columns of T ;1 are eigenvectors of PQ corresponding to the eigenvalues fi g.
Later, it will be shown that PQ has a real diagonal Jordan form and that 0, which
are consequences of P 0 and Q 0.
Although the eigenvectors are not unique, in the case of a minimal realization they
can always be chosen such that
P^ = TPT = ;
Q^ = (T ;1) QT ;1 = ;
where = diag(1 ; 2 ; : : : ; n ) and 2 = . This new realization with controllability
and observability Gramians P^ = Q^ = will be referred to as a balanced realization
(also called internally balanced realization). The decreasingly order numbers, 1
2 : : : n 0, are called the Hankel singular values of the system.
More generally, if a realization of a stable system is not minimal, then there is a transformation such that the controllability and observability Gramians for the transformed
realization are diagonal and the controllable and observable subsystem is balanced. This
is a consequence of the following matrix fact.
76
Theorem 3.22 Let P and Q be two positive semidenite matrices. Then there exists
a nonsingular matrix T such that
2
1
6
2
TPT = 64
0
3
2
77 ; (T ;1) QT ;1 = 66 1
5
4
0
respectively, with 1 ; 2 ; 3 diagonal and positive denite.
3
3
77 ;
5
Proof. Since P is a positive semidenite matrix, there exists a transformation T1 such
that
Now let
T1PT1 = I0 00 :
(T );1 QT ;1 =
1
Q
11
Q12
Q12 ;
Q22
2
(T2);1 (T1);1 QT1;1(T2 );1 = 4
2
(T3);1 = 4
21
0
0
0
Q^ 121 Q^ 122
Q^ 121
Q^ 122 5 :
Q22
0
^
;Q121 ;1 2
0 0
I 0 5;
0 I
giving
2 2
1
(T3);1 (T2);1 (T1);1 QT1;1(T2 );1 (T3 );1 = 4 0
0
0
0
0
0 0 Q22 ; Q^ 121 ;1 2 Q^ 121
3
5:
77
2 ;1=2
1
0
(T4 );1 = 4 0 I
Dene
Then
0 U2
and let
0
0
3
5
T = T4 T3 T2 T1:
2
1
6
TPT = 64
2
3
2
77 ; (T );1QT ;1 = 66 1
5
4
3
3
77
5
2
with 2 = I .
Corollary 3.23 The product of two positive semi-denite matrices is similar to a positive semi-denite matrix.
Proof. Let P and Q be any positive semi-denite matrices. Then it is easy to see that
with the transformation given above
TPQT ;1 =
2
0
0 0 :
1
A B
,
T such that G =
Gramian Q given by
TAT ;1 TB
CT ;1 D
2
1
6
P = 64
2
3
2
77 ; Q = 66 1
5
4
3
3
77 ;
5
A B
is a minimal realization, a balanced realization
In the special case where
C D
78
;1
TB
4. Let T ;1 = RU ;1=2 . Then TPT = (T );1 QT ;1 = and TAT
CT ;1 D
is balanced.
Assume that the Hankel singular values of the system is decreasingly ordered so that
= diag(1 ; 2 ; : : : ; n ) and 1 2 : : : n and suppose r r+1 for some r then
the balanced realization implies that those states corresponding to the singular values
of r+1 ; : : : ; n are less controllable and observable than those states corresponding to
1 ; : : : ; r . Therefore, truncating those less controllable and observable states will not
lose much information about the system. These statements will be made more concrete
in Chapter 7 where error bounds will be derived for the truncation error.
Two other closely related realizations are called input normal realization with P = I
and Q = 2 , and output normal realization with P = 2 and Q = I . Both realizations
can be obtained easily from the balanced realization by a suitable scaling on the states.
y
s;1
1
s;1 u
s+1
1
1 1
g(s) = ss ;
+ 1 s ; 1 = s + 1;
is stable and has a rst order minimal realization. On the other hand, let
x1 = y
x2 = u ; :
Then a state space description for this dynamical system is given by
x_
1 ;1 x 1
1
1
x_ 2 = 0 ;1
x2 + 2 u
y =
1
x
1
0
x2
79
and is a second order system. Moreover, it is easy to show that the unstable mode 1 is
uncontrollable but observable. Hence, the output can be unbounded if the initial state
x1 (0) is not zero. We should also note that the above problem does not go away by
changing the interconnection order:
y
s;1
s+1
s;1
1
In the later case, the unstable mode 1 becomes controllable but unobservable. The
unstable mode can still result in the internal signal unbounded if the initial state
(0) is not zero. Of course, there are fundamental dierences between these two types
of interconnections as far as control design is concerned. For instance, if the state is
available for feedback control, then the latter interconnection can be stabilized while
the former cannot be.
This example shows that we must be very careful in canceling unstable modes in
the procedure of forming a transfer function in control designs; otherwise the results
obtained may be misleading and those unstable modes become hidden modes waiting
to blow. One observation from this example is that the problem is really caused by the
unstable zero of the subsystem ss;+11 . Although the zeros of an SISO transfer function
are easy to see, it is not quite so for an MIMO transfer matrix. In fact, the notion of
\system zero" cannot be generalized naturally from the scalar transfer function zeros.
For example, consider the following transfer matrix
2 1
1 3
6 s + 1 s + 2 77
G(s) = 64
1 5
2
s+2 s+1
which is stable and each element of G(s) has no nite zeros. Let
2 s+2
s +p1 3
p
;
K =4 s; 2 s; 2 5
0
1
which is unstable. However,
3
2
p
s
+
2
6 ; (s + 1)(s + 2) 0 77
KG = 664
7
2
1 5
s+2
s+1
p
is stable. This implies that G(s) must have an unstable zero at 2 that cancels the
unstable pole of K . This leads us to the next topic: multivariable system poles and
zeros.
80
Let R[s] denote the polynomial ring with real coecients. A matrix is called a polynomial matrix if every element of the matrix is in R[s]. A square polynomial matrix is
called a unimodular matrix if its determinant is a nonzero constant (not a polynomial
of s). It is a fact that a square polynomial matrix is unimodular if and only if it is
invertible in R[s], i.e., its inverse is also a polynomial matrix.
Denition 3.11 Let Q(s) 2 R[s] be a (p m) polynomial matrix. Then the normal
rank of Q(s), denoted normalrank (Q(s)), is the rank in R[s] or, equivalently, is the
maximum dimension of a square submatrix of Q(s) with nonzero determinant in R[s].
In short, sometimes we say that a polynomial matrix Q(s) has rank(Q(s)) in R[s] when
we refer to the normal rank of Q(s).
To show the dierence between the normal rank of a polynomial matrix and the
rank of the polynomial matrix evaluated at certain point, consider
Q(s) = ss2 11 :
Then Q(s) has normal rank 2 since det Q(s) = s ; s2 6 0. However, Q(0) has rank 1.
It is a fact in linear algebra that any polynomial matrix can be reduced to a socalled Smith form through some pre- and post- unimodular operations. [cf. Kailath,
1984, pp.391].
Lemma 3.25 (Smith form) Let P (s) 2 R[s] be any polynomial matrix, then there
exist unimodular matrices U (s); V (s) 2 R[s] such that
2
(s) 0 0 0 3
66 10
2 (s) 0 0 77
.. . .
.
.. 77
U (s)P (s)V (s) = S (s) := 666 ...
. ..
.
. 7
4 0
0
r (s) 0 5
0
0 0 0
and
i (s) divides
i+1 (s).
S (s) is called the Smith form of P (s). It is also clear that r is the normal rank of P (s).
We shall illustrate the procedure of obtaining a Smith form by an example. Let
2 s + 1 (s + 1)(2s + 1) s(s + 1) 3
P (s) = 4 s + 2 (s + 2)(s2 + 5s + 3) s(s + 2) 5 :
1
2s + 1
s
The polynomial matrix P (s) has normal rank 2 since
s + 1 (s + 1)(2s + 1)
det(P (s)) 0; det s + 2 (s + 2)(s2 + 5s + 3) = (s + 1)2 (s + 2)2 6 0:
81
First interchange the rst row and the third row and use row elementary operation to
zero out the s+1 and s+2 elements of P (s). This process can be done by pre-multiplying
a unimodular matrix U to P (s):
20 0 1 3
U = 4 0 1 ;(s + 2) 5 :
1 0 ;(s + 1)
Then
21
3
2s + 1
s
P1 (s) := U (s)P (s) = 4 0 (s + 1)(s + 2)2 0 5 :
0
0
0
Next use column operation to zero out the 2s + 1 and s terms in P1 . This process can
be done by post-multiplying a unimodular matrix V to P1 (s):
2 1 ;(2s + 1) ;s 3
1
0 5
V (s) = 4 0
0
0
1
and
21
3
0
0
P1 (s)V (s) = 4 0 (s + 1)(s + 2)2 0 5 :
0
0
0
Then we have
21
3
0
0
S (s) = U (s)P (s)V (s) = 4 0 (s + 1)(s + 2)2 0 5 :
0
0
0
Similarly, let Rp (s) denote the set of rational proper transfer matrices.1 Then any
real rational transfer matrix can be reduced to a so-called McMillan form through some
pre- and post- unimodular operations.
Lemma 3.26 (McMillan form) Let G(s) 2 Rp(s) be any proper real rational transfer
matrix, then there exist unimodular matrices U (s); V (s) 2 R[s] such that
2 1 (s) 0 0 0 3
66 10(s) 2((ss)) 0 0 77
2
6
.. . .
.. .. 777
U (s)G(s)V (s) = M (s) := 66 ...
.
.
. . 7
64
r
0
0 r ((ss)) 0 5
0
0 0 0
and i (s) divides i+1 (s) and i+1 (s) divides i (s).
1 It is obvious that a similar notion of normal rank can be extended to a matrix in R ( ). In
particular, let ( ) 2 R ( ) be a rational matrix and write ( ) = ( ) ( ) such that ( ) is a
p s
G s
p s
G s
N s =d s
d s
polynomial and ( ) is a polynomial matrix, then ( ) is said to have normal rank if ( ) has
normal rank .
N s
G s
N s
82
Proof. If we write the transfer matrix G(s) as G(s) = N (s)=d(s) such that d(s) is a
scalar polynomial and N (s) is a p m polynomial matrix and if let the Smith form of
N (s) be S (s) = U (s)N (s)V (s), the conclusion follows by letting M (s) = S (s)=d(s). 2
Denition 3.12 The number Pi deg(i (s)) is called the McMillan degree of G(s)
where deg(i (s)) denotes the degree of the polynomial i (s), i.e., the highest power
of s in i (s).
The McMillan degree of a transfer matrix is closely related to the dimension of a
minimal realization of G(s). In fact, it can be shown that the dimension of a minimal
realization of G(s) is exactly the McMillan degree of G(s).
Denition 3.13 The roots of all the polynomials i(s) in the McMillan form for G(s)
are called the poles of G.
Let (A; B; C; D) be a minimal realization of G(s). Then it is fairly easy to show that
a complex number is a pole of G(s) if and only if it is an eigenvalue of A.
Denition 3.14 The roots of all the polynomials i(s) in the McMillan form for G(s)
are called the transmission zeros of G(s). A complex number z0 2 C is called a blocking
zero of G(s) if G(z0 ) = 0.
It is clear that a blocking zero is a transmission zero. Moreover, for a scalar transfer
function, the blocking zeros and the transmission zeros are the same.
We now illustrate the above concepts through an example. Consider a 3 3 transfer
matrix:
3
2
2s + 1
s
1
66 (s + 1)(s + 2) (s + 1)(s + 2) (s + 1)(s + 2) 77
77
66
1
s2 + 5s + 3
s
77 :
6
G(s) = 6 (s + 1)2
(s + 1)2
(s + 1)2
77
66
5
4
1
2s + 1
s
(s + 1)2 (s + 2) (s + 1)2 (s + 2) (s + 1)2 (s + 2)
Then G(s) can be written as
s + 1 (s + 1)(2s + 1) s(s + 1)
G(s) = (s + 1)12 (s + 2) 4 s + 2 (s + 2)(s2 + 5s + 3) s(s + 2)
1
2s + 1
s
3
5 := N (s) :
d(s)
83
Since N (s) is exactly the same as the P (s) in the previous example, it is clear that the
G(s) has the McMillan form
3
2
1
0
0
77
66 (s + 1)2(s + 2)
7
66
s + 2 0 77
M (s) = U (s)G(s)V (s) = 66
0
s + 1 77
6
and G(s) has McMillan degree of 4. The poles of the transfer matrix are f;1; ;1; ;1; ;2g
and the transmission zero is f;2g. Note that the transfer matrix has pole and zero at
the same location f;2g; this is the unique feature of multivariable systems.
To get a minimal state space realization for G(s), note that G(s) has the following
partial fractional expansion:
213
203
20 0 03
1
1
G(s) = 4 0 1 0 5 + s + 1 4 0 5 1 ;1 ;1 + s + 1 4 1 5 0 3 1
0
0
0 0 0
2 3
203
1
5 ;1 3 2 +
4 5 1 ;1 ;1
(s + 1)2 11
2 ;1 3
1
+ s + 2 4 0 5 1 ;3 ;2
0
+ s +1 1 4 0
1
1
Since there are repeated poles at ;1, the Gilbert's realization procedure described in the
last section cannot be used directly. Nevertheless, a careful inspection of the fractional
expansion results in a 4-th order minimal state space realization:
2 ;1 0 1 0 0 3 1 3
66 0 ;1 1 0 ;1 3 2 77
66 0 0 ;1 0 1 ;1 ;1 77
G(s) = 66 0 0 0 ;2 1 ;3 ;2 77 :
66 0 0 1 ;1 0 0 0 77
4 1 0 0 0 0 1 0 5
0 1 0 1 0 0 0
We remind readers that there are many dierent denitions of system zeros. The
denitions introduced here are the most common ones and are useful in this book.
Lemma 3.27 Let G(s) be a p m proper transfer matrix with full column normal rank.
Then z0 2 C is a transmission zero of G(s) if and only if there exists a 0 =
6 u0 2 C m
such that G(z0 )u0 = 0.
84
Proof. We shall outline a proof of this lemma. We shall rst show that there is a
vector u0 2 C m such that G(z0 )u0 = 0 if z0 2 C is a transmission zero. Without loss of
generality, assume
2
66
6
G(s) = U1 (s) 66
64
1 (s)
1 (s)
0
..
.
0
0
2 (s)
2 (s)
..
.
0
0
...
0
0
..
.
m (s)
m (s)
3
77
77 V (s)
77 1
5
for some unimodular matrices U1 (s) and V1 (s) and suppose z0 is a zero of 1 (s), i.e.,
1 (z0 ) = 0. Let
where e1 = [1; 0; 0; : : :] 2 Rm . Then it is easy to verify that G(z0 )u0 = 0. On the other
hand, suppose there is a u0 2 C m such that G(z0 )u0 = 0. Then
2
66
6
U1 (z0 ) 66
64
Dene
Then
1 (z0 )
1 (z0 )
0
..
.
0
0
2 (z0 )
2 (z0 )
..
.
0
0
0
0
..
.
...
m (z0 )
m (z0 )
3
77
77 V (z )u = 0:
77 1 0 0
5
2u 3
66 u12 77
64 ... 75 = V1 (z0)u0 6= 0:
um
2 (z )u
66 12(z00)u12
64 ...
m (z0 )um
3
77
75 = 0:
Note that the lemma may not be true if G(s) does not have full column normal rank.
This can be seen from the following example. Consider
G(s) = s +1 1 11 11 ; u0 = ;11 :
85
It is easy to see that G has no transmission zero but G(s)u0 = 0 for all s. It should
also be noted that the above lemma applies even if z0 is a pole of G(s) although G(z0 )
is not dened. The reason is that G(z0 )u0 may be well dened. For example,
s;1 0
1
G(s) = s+1
0 ss;+21 ; u0 = 0 :
Corollary 3.30 Let G(s) be a square m m proper transfer matrix and det G(s) 6 0.
Suppose zo 2 C is not a pole of G(s). Then z0 2 C is a transmission zero of G(s) if
and only if det G(z0 ) = 0.
Using thepabove corollary, we can conrm that the example in the last section does have
a zero at 2 since
2 1
1 3
6 s + 1 s + 2 77 = 2 ; s2 :
det 64
2
1 5 (s + 1)2 (s + 2)2
s+2 s+1
Note that the above corollary may not be true if z0 is a pole of G. For example,
s;1
0
G(s) = s+1
s
+2
0 s;1
A B
C D
86
Denition 3.15 The eigenvalues of A are called the poles of the realization of G(s).
To dene zeros, let us consider the following system matrix
B :
Q(s) = A ;C sI D
Denition 3.16 A complex number z0 2 C is called an invariant zero of the system
realization if it satises
A;z I B
A ; sI B
0
rank
C
D < normalrank
C D :
The invariant zeros are not changed by constant state feedback since
A + BF ; z I B
A ; z I B I 0
A;z I B
0
0
0
rank
C + DF
D = rank
C
D
F I = rank
C
D :
It is also clear that invariant zeros are not changed under similarity transformation.
The following lemma is obvious.
A ; sI B
Lemma 3.31 Suppose
has full column normal rank.
Then z0 2 C is
D
an invariant zero of a realization (A; B; C; D) if and only if there exist 0 6= x 2 C n and
u 2 C m such that
A ; z I B x
0
C
D
u = 0:
Moreover, if u = 0, then z0 is also a non-observable mode.
x
since A ; sI B has full column normal rank.
C
x
x
B
A ; sI B has full
D u = 0 or u = 0 since
C D
87
A;z I
C
x=0
A ; sI B
Lemma 3.32 Suppose
has full row normal rank.
Then z0 2 C is an
D
invariant zero of a realization (A; B; C; D) if and only if there exist 0 6= y 2 C n and
v 2 C p such that
y v A ; z0I B = 0:
C
D
Moreover, if v = 0, then z0 is also a non-controllable mode.
A ; sI B
Lemma 3.33 G(s) has full column (row) normal rank if and only if
C D
and
C (A ; sI );1 I
A ; sI B
0
G(s)
A ; sI B
normalrank
C D = n + normalrank(G(s)):
Theorem 3.34 Let G(s) be a real rational proper transfer matrix and let (A; B; C; D)
Proof. We will give a proof only for the case that the transmission zero is not a pole
of G(s). Then, of course, z0 is not an eigenvalue of A since the realization is minimal.
Note that
A ; sI B
A ; sI B
I
0
C D = C (A ; sI );1 I
0
G(s) :
Since we have assumed that z0 is not a pole of G(s), we have
A;z I B
0
rank
C
D = n + rank G(z0 ):
88
Hence
A;z I B
A ; sI B
0
rank
C
D < normalrank
C D
if and only if rank G(z0 ) < normalrank G(s). Then the conclusion follows from
Lemma 3.29.
2
Note that the minimality assumption is essential for the converse statement. For
example,
A 0 consider a transfer matrix G(s) = D (constant) and a realization of G(s) =
where A is any square matrix with any dimension and C is any matrix with
C D
compatible dimension. Then G(s) has no poles or zeros but every eigenvalue of A is an
A 0 .
invariant zero of the realization C
D
Lemma 3.36 Let G(s) 2 Rp(s) be a p m transfer matrix and let (A; B; C; D) be a
minimal realization. If the system input is of the form u(t) = u0 et , where 2 C is not
a pole of G(s) and u0 2 C m is an arbitrary constant vector, then the output due to the
input u(t) and the initial state x(0) = (I ; A);1 Bu0 is y(t) = G()u0 et ; 8t 0.
Proof. The system response with respect to the input u(t) = u0et and the initial
condition x(0) = (I ; A);1 Bu0 is (in terms of Laplace transform):
Y (s) = C (sI ; A);1 x(0) + C (sI ; A);1 BU (s) + DU (s)
= C (sI ; A);1 x(0) + C (sI ; A);1 Bu0 (s ; );1 + Du0 (s ; );1
= C (sI ; A);1 (x(0) ; (I ; A);1 Bu0 ) + G()u0 (s ; );1
= G()u0 (s ; );1 :
Hence y(t) = G()u0 et .
Combining the above two lemmas, we have the following results that give a dynamical
interpretation of a system's transmission zero.
Corollary 3.37 Let G(s) 2 Rp(s) be a p m transfer matrix and let (A; B; C; D) be
a minimal realization. Suppose that z0 2 C is a transmission zero of G(s) and is not
a pole of G(s). Then for any nonzero vector u0 2 C m the output of the system due to
the initial state x(0) = (z0 I ; A);1 Bu0 and the input u = u0 ez0 t is identically zero:
y(t) = G(z0 )u0 ez0 t = 0.
89
The following lemma characterizes the relationship between zeros of a transfer function and poles of its inverse.
Lemma 3.38 Suppose that G = CA DB is a square transfer matrix with D nonsingular, and suppose z0 is not an eigenvalue of A (note that the realization is not
necessarily minimal). Then there exists x0 such that
G(z0 )u0 = 0:
D;1
has a pole at z0 which is observable. Then, by denition, there exists x0 such that
(A ; BD;1 C )x0 = z0 x0
and
Cx0 6= 0:
2
The above lemma implies that z0 is a zero of an invertible G(s) if and only if it is a
pole of G;1 (s).
90
PERFORMANCE SPECIFICATIONS
92
1
X
kxkp :=
i=0
jxi jp
!1=p
l1 [0; 1) space:
l1 [0; 1) consists of all bounded sequences x = (x0 ; x1 ; : : :), and the l1 norm is
dened as
Lp (I ) spaces for 1 p 1:
For each 1 p < 1, Lp (I ) consists of all Lebesgue measurable functions x(t)
dened on an interval I R such that
kxkp :=
and
Z
1=p
jx(t)jp dt
93
Note that if each component or function is itself a vector or matrix, then the corresponding Banach space can also be formed by replacing the absolute value j j of each
component or function with its spatially normed component or function. For example,
consider a vector space with all sequences in the form of
x = (x0 ; x1 ; : : :)
where each component xi is a k m matrix and each element of xi is bounded. Then
xi is bounded in any matrix norm, and the vector space becomes a Banach space if the
following norm is dened
kxk1 := sup i (xi )
i
where i (xi ) := kxi k is any matrix norm. This space will also be denoted by l1 .
Let V1 and V2 be two vector spaces and let T be an operator from S V1 into
V2 . An operator T is said to be linear if for any x1 ; x2 2 S and scalars ; 2 C , the
following holds:
T (x1 + x2 ) = (Tx1 ) + (Tx2 ):
Moreover, let V0 be a linear subspace in V1 . Then the operator T0 : V0 7;! V2 dened by
T0 x = Tx for every x 2 V0 is called the restriction of T to V0 and is denoted as T jV0 = T0 .
On the other hand, a linear operator T : V1 7;! V
2 coinciding
with T0 on V0 V1 is
x
1
n
n+m ,
called an extension of T0. For example, let V0 :=
0 : x1 2 C V1 = C
and let
2 2 C (n+m)(n+m) ; T = A1 0 2 C (n+m)(n+m) :
T = A01 A
0
A3
0 0
Then T jV0 = T0 .
Denition 4.1 Two normed spaces V1 and V2 are said to be linearly isometric, denoted
by V1
= V2 , if there exists a linear operator T of V1 onto V2 such that
kTxk = kxk
for all x in V1 . In this case, the mapping T is said to be an isometric isomorphism.
hx; yi := x y =
n
X
i=1
2x 3
2y 3
1
1
xi yi 8x = 64 ... 75 ; y = 64 ... 75 2 C n :
xn
yn
PERFORMANCE SPECIFICATIONS
94
Note that many important metric notions and geometrical properties such as length,
distance, angle, and the energy of physical systems can be deduced from this inner
product. For instance, the length of a vector x 2 C n is dened as
x; y
x; y
i=
x y
95
A Hilbert space is a complete inner product space with the norm induced by its
inner product. Obviously, a Hilbert space is also a Banach space. For example, C n
with the usual inner product is a (nite dimensional) Hilbert space. More generally, it
is straightforward to verify that C nm with the inner product dened as
hA; B i := Trace A B =
n X
m
X
i=1 j =1
i=;1
jxi j2 < 1
hx; yi :=
1
X
i=;1
xi yi
for x; y 2 l2 (;1; 1). The subspaces l2 (;1; 0) and l2 [0; 1) of l2 (;1; 1) are
dened similarly and consist of sequences of the form x = (: : : ; x;2 ; x;1 ) and
x = (x0 ; x1 ; x2 ; : : :), respectively.
Note that we can also dene a corresponding Hilbert space even if each component
xi is a vector or a matrix; in fact, the following inner product will suce:
hx; yi :=
1
X
i=;1
Trace(xi yi ):
L2 (I ) for I R:
L2 (I ) consists of all square integrable and Lebesgue measurable functions dened
on an interval I R with the inner product dened as
PERFORMANCE SPECIFICATIONS
96
Some very often used spaces in this book are L2 [0; 1); L2 (;1; 0]; L2 (;1; 1).
More precisely, they are dened as
L2 = L2 (;1; 1): Hilbert space of matrix-valued functions on R, with inner
product
hf; gi :=
Z1
;1
L2+ = L2 [0; 1): subspace of L2 (;1; 1) with functions zero for t < 0.
L2; = L2 (;1; 0]: subspace of L2 (;1; 1) with functions zero for t > 0.
Let H be a Hilbert space and M H a subset. Then the orthogonal complement of
M , denoted by H M or M ?, is dened as
M ? = fx : hx; yi = 0; 8y 2 M; x 2 Hg :
It can be shown that M ? is closed (hence M ? is also a Hilbert space). For example,
let M = L2+ L2 , then M ? = L2; is a Hilbert space.
Let M and N be subspaces of a vector space V . V is said to be the direct sum of
M and N , written V = M N , if M \ N = f0g, and every element v 2 V can be
expressed as v = x + y with x 2 M and y 2 N . If V is an inner product space and
M and N are orthogonal, then V is said to be the orthogonal direct sum of M and N .
As an example, it is easy to see that L2 is the orthogonal direct sum of L2; and L2+ .
Similarly, l2 (;1; 1) is the orthogonal direct sum of l2 (;1; 0) and l2 [0; 1).
The following is a version of the so-called orthogonal projection theorem:
Theorem 4.2 Let H be a Hilbert space, and let M be a closed subspace of H. Then for
each vector v 2 H, there exist unique vectors x 2 M and y 2 M ? such that v = x + y,
i.e., H = M M ?. Moreover, x 2 M is the unique vector such that d(v; M ) = kv ; xk.
Let H1 and H2 be two Hilbert spaces, and let A be a bounded linear operator from
H1 into H2 . Then there exists a unique linear operator A : H2 7;! H1 such that for
all x 2 H1 and y 2 H2
hAx; yi = hx; A yi:
A is called the adjoint of A. Furthermore, A is called self-adjoint if A = A .
Let H be a Hilbert space and M H be a closed subspace. A bounded operator P
mapping from H into itself is called the orthogonal projection onto M if
P (x + y) = x; 8 x 2 M and y 2 M ? :
97
Let S C be an open set, and let f (s) be a complex valued function dened on S :
f (s) : S 7;! C :
Then f (s) is said to be analytic at a point z0 in S if it is dierentiable at z0 and also
at each point in some neighborhood of z0 . It is a fact that if f (s) is analytic at z0 then
f has continuous derivatives of all orders at z0 . Hence, a function analytic at z0 has
a power series representation at z0 . The converse is also true, i.e., if a function has a
power series at z0 , then it is analytic at z0 . A function f (s) is said to be analytic in S if
it has a derivative or is analytic at each point of S . A matrix valued function is analytic
in S if every element of the matrix is analytic in S . For example, all real rational stable
transfer matrices are analytic in the right-half plane and e;s is analytic everywhere.
A well known property of the analytic functions is the so-called Maximum Modulus
Theorem.
Theorem 4.3 If f (s) is dened and continuous on a closed-bounded set S and analytic
on the interior of S , then the maximum of jf (s)j on S is attained on the boundary of
S , i.e.,
max
jf (s)j = smax
jf (s)j
s2S
2@S
where @S denotes the boundary of S .
Next we consider some frequently used complex (matrix) function spaces.
L2 (j R) Space
L2 (j R) or simply L2 is a Hilbert space of matrix-valued (or scalar-valued) func-
tions on j R and consists of all complex matrix functions F such that the integral
below is bounded, i.e.,
Z1
;1
kF k2 := hF; F i:
For example, all real rational strictly proper transfer matrices with no poles on the
imaginary axis form a subspace (not closed) of L2 (j R) which is denoted by RL2 (j R) or
simply RL2 .
PERFORMANCE SPECIFICATIONS
98
H2 Space2
H2 is a (closed) subspace of L2 (j R) with matrix functions F (s) analytic in Re(s) >
0 (open right-half plane). The corresponding norm is dened as
Z1
kF k22 := sup 21
Trace [F ( + j!)F ( + j!)] d! :
>0
;1
kF k22 = 21
Z1
;1
Hence, we can compute the norm for H2 just as we do for L2 . The real rational
subspace of H2 , which consists of all strictly proper and real rational stable transfer
matrices, is denoted by RH2 .
H2? Space
H2? is the orthogonal complement of H2 in L2 , i.e., the (closed) subspace of functions in L2 that are analytic in the open left-half plane. The real rational subspace
of H2? , which consists of all strictly proper rational transfer matrices with all poles
in the open right half plane, will be denoted by RH?2 . It is easy to see that if G
is a strictly proper, stable, and real rational transfer matrix, then G 2 H2 and
G 2 H2? . Most of our study in this book will be focused on the real rational
case.
The L2 spaces dened above in the frequency domain can be related to the L2 spaces
dened in the time domain. Recall the fact that a function in L2 space in the time
domain admits a bilateral Laplace (or Fourier) transform. In fact, it can be shown that
this bilateral Laplace (or Fourier) transform yields an isometric isomorphism between
the L2 spaces in the time domain and the L2 spaces in the frequency domain (this is
what is called Parseval's relations or Plancherel Theorem in complex analysis):
L2 (;1; 1)
= L2 (j R)
H2
L2 [0; 1) =
L2 (;1; 0]
= H2? :
As a result, if g(t) 2 L2 (;1; 1) and if its Fourier (or bilateral Laplace) transform is
G(j!) 2 L2 (j R), then
kGk2 = kgk2 :
2 The H space and H space dened below together with the H spaces, 1, which will not be
1
p
2
introduced in this book, are usually called Hardy spaces named after the mathematician G. H. Hardy
(hence the notation of H).
3 See Francis [1987].
p
99
Hence, whenever there is no confusion, the notation of functions in the time domain
and in the frequency domain will be used interchangeably.
Dene an orthogonal projection
P+ : L2 (;1; 1) 7;! L2 [0; 1)
such that, for any function f (t) 2 L2 (;1; 1), we have g(t) = P+ f (t) with
t 0;
g(t) := f (0t;); for
for t < 0:
In this book, P+ will also be used to denote the projection from L2 (j R) onto H2 .
Similarly, dene P; as another orthogonal projection from L2 (;1; 1) onto L2 (;1; 0]
(or L2 (j R) onto H2? ). Then the relationships between L2 spaces and H2 spaces can be
shown as in Figure 4.1.
Laplace Transform L2 [0; 1)
H2
Inverse Transform
P+
L2 (;1; 1)
P;
L2 (;1; 0]
P+
L2 (j R)
P;
Laplace Transform
Inverse Transform
?
H2?
PERFORMANCE SPECIFICATIONS
100
H1 Space
H1 is a (closed) subspace in L1 with functions that are analytic in the open
right-half plane and bounded on the imaginary axis. The H1 norm is dened as
kF k1 := sup [F (s)] = ess sup [F (j!)] :
Re(s)>0
! 2R
; Space
H1
H1; is a (closed) subspace in L1 with functions that are analytic in the open
; norm is dened as
left-half plane and bounded on the imaginary axis. The H1
kF k1 := sup [F (s)] = ess sup [F (j!)] :
Re(s)<0
! 2R
; is denoted by RH;
The real rational subspace of H1
1 which consists of all proper
rational transfer matrices with all poles in the open right half plane.
Denition 4.3 A transfer matrix G(s) 2 H1; is usually said to be antistable or anticausal.
Remark 4.1 The notation for L1 is somewhat unfortunate; it should be clear to the
reader that the L1 space in the time domain and in the frequency domain denote
completely dierent spaces. The L1 space in the time domain is usually used to denote
signals, while the L1 space in the frequency domain is usually used to denote transfer
functions and operators.
~
Let G(s) 2 L1 be a p q transfer matrix. Then a multiplication operator is dened
as
MG : L2 7;! L2
MG f := Gf:
101
In writing the above mapping, we have assumed that f has a compatible dimension. A
more accurate description of the above operator should be
MG : Lq2 7;! Lp2
i.e., f is a q-dimensional vector function with each component in L2 . However, we shall
suppress all dimensions in this book and assume all objects have compatible dimensions.
It is easy to verify that the adjoint operator MG = MG .
A useful fact about the multiplication operator is that the norm of a matrix G in
L1 equals the norm of the corresponding multiplication operator.
Theorem 4.4 Let G 2 L1 be a p q transfer matrix. Then kMGk = kGk1 .
Remark 4.2 It is also true that this operator norm equals the norm of the operator
restricted to H2 (or H2? ), i.e.,
kMGk = kMGjH2 k := sup fkGf k2 : f 2 H2 ; kf k2 1g :
This will be clear in the proof where an f 2 H2 is constructed.
~
Z1
kGk21 21
kf (j!)k2 d!
;1
= kGk21kf k22:
To show that kGk1 is the least upper bound, rst choose a frequency !o where [G(j!)]
is maximum, i.e.,
[G(j!0 )] = kGk1
and denote the singular value decomposition of G(j!0 ) by
r
X
i=2
2 ej
66 12 ej
v1 (j!0 ) = 64 ..
.
1
2
q ejq
3
77
75
PERFORMANCE SPECIFICATIONS
102
where i 2 R is such that i 2 [;; 0) and q is the column dimension of G. Now let
i 0 be such that
i ; j!0
i = \ + j! ;
and let f be given by
2 ;s
66 12 +;+sss
f (s) = 66 ..
4 . ;s
1
1
2
2
q qq +s
3
77
77 f^(s)
5
If !0 = 1, then G(j!0 ) is real and v1 is real. In this case, the conclusion follows by
letting f (s) = v1 f^(s).
2
ZT
;T
if the limit exists and is nite for all . It is easy to see from the denition that Ruu ( ) =
(; ). Assume further that the Fourier transform of the signal's autocorrelation
Ruu
matrix function exists (and may contain impulses). This Fourier transform is called the
spectral density of u, denoted Suu (j!):
Suu (j!) :=
Z1
;1
103
Thus Ruu ( ) can be obtained from Suu (j!) by performing an inverse Fourier transform:
Z1
Ruu ( ) := 21
Suu (j!)ej! d!:
;1
We will call signal u(t) a power signal if the autocorrelation matrix Ruu ( ) exists and
is nite for all , and moreover, if the power spectral density function Suu (j!) exists
(note that Suu (j!) need not be bounded and may include impulses).
The power of the signal is dened as
ZT
p
1
kukP = Tlim
ku(t)k2 dt = Trace [Ruu (0)]
!1 2T ;T
where kk is the usual Euclidean norm and the capital script P is used to dierentiate
this power semi-norm from the usual Lebesgue Lp norm. The set of all signals having
nite power will be denoted by P .
The power semi-norm of a signal can also be computed from its spectral density
function
Z1
1
2
Trace[Suu (j!)]d!:
kukP = 2
;1
This expression implies that, if u 2 P , Suu is strictly proper in the sense that Suu (1) =
0. We note that if u 2 P and ku(t)k1 := supt ku(t)k < 1, then kukP kuk1.
However, not every signal having nite 1-norm is a power signal since the limit in the
denition may not exist. For example, let
2k < t < 22k+1 ; for k = 0; 1; 2; : : :
u(t) = 10 2otherwise
:
Then limT !1 21T ;TT u2 dt does not exist. Note also that power signals are persistent
signals in time such as sines or cosines; clearly, a time-limited signal has zero power, as
does an L2 signal. Thus k kP is only a semi-norm, not a norm.
Now let G be a linear system transfer matrix with convolution kernel g(t), input u(t), and output z (t). Then Rzz ( ) = g( ) Ruu ( ) g (; ) and Szz (j!) =
G(j!)Suu (j!)G (j!). These properties are useful in establishing some input and output relationships in the next section.
A signal u(t) is said to have bounded power spectral density if kSuu (j!)k1 < 1.
The set of signals having bounded spectral density is denoted as
S := fu(t) 2 Rm : kSuu (j!)k1 < 1g:
p
The quantity kuks := kSuu (j!)k is called the spectral density norm of u(t). The set
S can be used to model signals with xed spectral characteristics by passing white noise
signals through a weighting lter. Similarly, P could be used to model signals whose
spectrum is not known but which are nite in power.
PERFORMANCE SPECIFICATIONS
104
z
We will further assume that G(s) is strictly proper, i.e, G(1) = 0 although most of
the results derived here hold for the non-strictly proper case. In the time-domain an
input-output model for such a system has the form of a convolution equation,
z =gu
i.e.,
Zt
z (t) = g(t ; )u( ) d
0
where the p q real matrix g(t) is the convolution kernel. Let the convolution kernel
and the corresponding transfer matrix be partitioned as
2 g (t) g (t) 3 2 g (t) 3
11
1q
1
6
7
6
.
.
. 7
.
.
g(t) = 4 .
. 5 = 4 .. 5 ;
gp1 (t) gpq (t)
gp (t)
Gp (s)
where gi (t) is a q-dimensional row vector of the convolution kernel and Gi (s) is a row
vector of the transfer matrix. Now if G is considered as an operator from the input space
to the output space, then a norm is induced on G, which, loosely speaking, measures
the size of the output for a given input u. These norms can determine the achievable
performance of the system for dierent classes of input signals
The various input/output relationships, given dierent classes of input signals, are
summarized in two tables below. Table 4.1 summarizes the results for xed input signals.
Note that the (t) in this table denotes the unit impulse and u0 2 Rq is a constant vector
indicating the direction of the input signal.
We now prove Table 4.1.
105
Input u(t)
Output z (t)
kz k2 = kGu0 k2 = kgu0k2
u(t) = u0 (t); u0 2 Rq
kz k1 = kgu0k1
kz kP = 0
kz k2 = 1
u(t) = u0 sin !0 t; u0 2 Rq
kz kP = p1 kG(j!0 )u0 k
2
Proof.
u = u0(t): If u(t) = u0(t), then z (t) = g(t)u0, so kz k2 = kgu0k2 and kz k1 =
kgu0k1 . But by Parseval's relation, kgu0k2 = kGu0k2 . On the other hand,
ZT
1
ug (t)g(t)u0 dt
P = Tlim
!1 2T ;T 0
Z1
1
Tracefg(t)g(t)g dt ku0 k2
Tlim
!1 2T ;1
1 kGk2 ku k2 = 0:
= Tlim
2 0
!1 2T
kz k2
u = u0 sin !0 t: With the input u(t) = u0 sin(!0 t), the i-th output as t ! 1 is
zi (t) = jGi (j!0 )u0 j sin[!0 t + argfGi (j!0 )u0 g]:
(4:1)
PERFORMANCE SPECIFICATIONS
106
The 2-norm of this signal is innite as long as Gi (j!0 )u0 6= 0, i.e., the system's
transfer function does not have a zero in every channel in the input direction at
the frequency of excitation.
The amplitude of the sinusoid (4.1) equals jGi (j!0 )u0 j. Hence
lim sup max
jz (t)j = max
jGi (j!0 )u0 j = kG(j!0 )u0 k1
i i
i
and
t!1
v
u
p
uX
lim sup kz (t)k = t jGi (j!0 )u0 j2 = kG(j!0 )u0 k :
t!1
i=1
It is interesting to see what this table signies from the control point of view. To
focus our discussion, let us assume that u(t) = u0 sin !0 t is a disturbance or a command
signal on a feedback system, and z (t) is the tracking error. Then we say that the system
has good tracking behavior if z (t) is small in some sense, for instance, lim supt!1 kz (t)k
is small. Note that
lim sup kz (t)k = kG(j!0 )u0 k
t!1
for any given !0 and u0 2 Rq . Now if we want to track signals from various channels,
that is if u0 can be chosen to be any direction, then we would require that (G(j!0 ))
be small. Furthermore, if, in addition, we want to track signals of many dierent
frequencies, we then would require that (G(j!0 )) be small at all those frequencies.
This interpretation enables us to consider the control system in the frequency domain
even though the specications are given in the time domain.
Table 4.2 lists the maximum possible system gains when the input signal u is not
required to be a xed signal; instead it can be any signal in a unit ball in some function
space.
Now we give a proof for Table 4.2. Note that the rst row (L2 7;! L2 ) has been
shown in Section 4.3.
107
Signal Norms
kuk22 =
L2
L2
kuk2P = 21
kuk2P = 21
L1
Z1
0
Induced Norms
kuk2 dt
kGk1
kuk2S = kSuu k1
Z1
;1
Z1
;1
kGk1
TracefSuu (j!)gd!
kGk2
TracefSuu (j!)gd!
kGk1
max
kgi k1
i
L1
Z1
Proof.
S 7;! S : If u 2 S , then
Now suppose
[G(j!0 )] = kGk1
and take a signal u such that Suu (j!0 ) = I . Then
kSzz (j!)k1 = kGk21 :
Z1
1
P = 2 ;1 TracefG(j!)Suu (j!)G (j!)g d!:
kz k2
kg(t)k dt
PERFORMANCE SPECIFICATIONS
108
Now let u be white with unit spectral density, i.e., Suu = I . Then
Z1
1
2
kz kP = 2
TracefG(j!)G (j!)g d! = kGk22 :
;1
P 7;! P :
kz kP kGk1 kukP :
To achieve the equality, assume that !0 is such that
[G(j!0 )] = kGk1
and denote the singular value decomposition of G(j!0 ) by
r
X
i=2
2 ej
66 12ej
v1 (j!) = 64 ..
.
1
2
q ejq
3
77
75
2 ;s
66 21 +;+sss
u(t) = 66 ..
4 . ;s
1
1
2
2
q qq +s
3
77
77 u^(t)
5
109
2 ;j!
66 21 +;+j!j!j!
Suu (j!) = 66
4 ... ;j!
1
1
2
2
q qq +j!
3
2 ;j!
77
66 12 +;+j!j!j!
77 Su^u^ (j!) 66 ..
5
4 . ;j!
1
1
2
2
q qq +j!
3
77
77
5
kukP = 1:
Finally,
jzi (t)j =
=
Z t
Z
gi( )u(t ; ) d t jgi( )u(t ; )j d
0
Z 0t X
Z 1X
q
q
jgij ( )j d kuk1
0 j =1
kgi k1 kuk1
0 j =1
and
kz k1 := max
sup jz (t)j
i t i
max
kgi k1 kuk1 :
i
jgij ( )j d kuk1
PERFORMANCE SPECIFICATIONS
110
That maxi kgi k1 is the least upper bound can be seen as follows: Assume
that the maximum maxi kgi k1 is achieved for i = 1. Let t0 be given and set
Note that since g1( ) is a vector function, the sign function sign(g1 ( )) is a
component-wise operation. Then kuk1 = 1 and
Zt
z1 (t0 ) =
g1 ( )u(t0 ; ) d
Zt X
q
0
0 j =1
= kg1 k1 ;
jg1j ( )j d
Z 1X
q
t0 j =1
jg1j ( )j d:
kz (t)k =
Z t
g( )u(t ; )d
Z t0
Zt
0
0
kg( )k ku(t ; )k d
kg( )k d kuk1 :
R
And, therefore, kz k1 01 kg( )k d kuk1 .
Next we shall derive some simple and useful bounds for the H1 norm and the L1
norm of a stable system. Suppose
G(s) = CA B0 2 RH1
is a balanced realization, i.e., there exists
= diag(1 ; 2 ; : : : ; n ) 0
with 1 2 : : : n 0, such that
A + A + BB = 0
A + A + C C = 0:
111
Theorem 4.5
1 kGk1
Z1
0
kg(t)k dt 2
n
X
i=1
i
Remark 4.3 It should be clear that the inequalities stated in the theorem do not
depend on a particular state space realization of G(s). However, use of the balanced
realization does make the proof simple.
~
Proof. The inequality 1 kGk1 follows from the Nehari Theorem of Chapter 8.
We will now show the other inequalities. Since
G(s) :=
Z1
0
kGk1
Z 1
;
st
= sup
g(t)e dt
Re(s)>0 0
Z 1
g(t)e;st
dt
sup
Re(s)>0 0
Z1
kg(t)k dt:
To prove the last inequality, let ui be the ith unit vector. Then
1
n
if i = j and X
ui ui = I:
0 if i =
6 j
i=1
Dene
i (t) = ui eAt=2 B and i (t) = CeAt=2 ui . It is easy to verify that
ui uj = ij =
k
i k22 =
k i k22 =
Z1
0
Z1
0
Z1
0
kg(t)k dt =
X
Z 1
X
n
n Z1
i
i
dt 0 k i
ik dt
0
i=1
i=1
n
n
X
X
i=1
k i k2 k i k2 2
i=1
i :
PERFORMANCE SPECIFICATIONS
112
2
It should be clear from the above two tables that many system performance criteria
can be stipulated as requiring a certain closed loop transfer matrix have small H2 norm
or H1 norm or L1 norm. Moreover, if L1 performance is satised, then the H1 norm
performance is also satised. We will be most interested in H2 and H1 performance in
this book.
kGk2 :=
s Z1
1
2 ;1 TracefG (j!)G(j!)g d!
= kgk2
=
sZ 1
;1
Tracefg(t)g(t)g dt
A B
G(s) =
C 0
113
(4:2)
where Lo and Lc are observability and controllability Gramians which can be obtained
from the following Lyapunov equations
ALc + LcA + BB = 0 A Lo + LoA + C C = 0:
kGk22
=
=
Z1
Z0
0
Tracefg(t)g(t)g dt =
Z1
0
Tracefg(t)g(t) g dt
TracefB eA t C CeAt B g dt =
Z1
0
TracefCeAt BB eA t C g dt:
The lemma follows from the fact that the controllability Gramian of (A; B ) and the
observability Gramian of (C; A) can be represented as
Lo =
Z1
0
Z1
0
A Lo + LoA + C C = 0:
To compute the L2 norm of a rational transfer function, G(s) 2 L2 , using state space
approach. Let G(s) = [G(s)]+ + [G(s)]; with G+ 2 RH2 and G; 2 RH?2 . Then
kGk22 = k[G(s)]+ k22 + k[G(s)]; k22
where k[G(s)]+ k2 and k[G(s)]; k2 = k[G(;s)]+ k2 can be computed using the above
lemma.
Still another useful characterization of the H2 norm of G is in terms of hypothetical
input-output experiments. Let ei denote the ith standard basis vector of Rm where m
is the input dimension of the system. Apply the impulsive input (t)ei ((t) is the unit
impulse) and denote the output by zi (t)(= g(t)ei ). Assume D = 0, and then zi 2 L2+
and
m
X
kGk22 = kzi k22 :
i=1
Note that this characterization of the H2 norm can be appropriately generalized for
nonlinear time varying systems, see Chen and Francis [1992] for an application of this
norm in sampled-data control.
PERFORMANCE SPECIFICATIONS
114
We shall rst consider, as in the L2 case, how to compute the 1 norm of an L1 transfer
matrix. Let G(s) 2 L1 and recall that the L1 norm of a transfer function G is dened
as
kGk1 := ess sup fG(j!)g:
!
f!1; ; !N g:
Then an estimate for kGk1 is
max fG(j!k )g:
1kN
This value is usually read directly from a Bode singular value plot. The L1 norm can
also be computed in state space if G is rational.
B 2 RL :
G(s) = CA D
1
(4:3)
Then kGk1 <
if and only if (D) <
and H has no eigenvalues on the imaginary
axis where
A + BR;1DC
;1 B
BR
H := ;C (I + DR;1D )C ;(A + BR;1 D C )
(4.4)
and R =
2I ; D D.
Proof. Let (s) =
2I ; G(s)G(s). Then it is clear that kGk1 <
if and only if
(j!) > 0 for all ! 2 R. Since (1) = R > 0 and since (j!) is a continuous function
of !, (j!) > 0 for all ! 2 R if and only if (j!) is nonsingular for all ! 2 R [ f1g,
i.e., (s) has no imaginary axis zero. Equivalently, ;1 (s) has no imaginary axis pole.
It is easy to compute by some simple algebra that
2
;1 (s) = 4
BR;1 3
H
;C DR;1 5 :
;
1
;
1
R DC R B
R ;1
115
Thus the conclusion follows if the above realization has neither uncontrollable modes
nor unobservable modes on the imaginary axis. Assume that j!0 is an eigenvalue
of H but not a pole of ;1 (s). Then j!0 must be either anunobservable mode of
;1
( R;1 D C R;1 B ; H ) or an uncontrollable mode of (H; ;CBR
DR;1 ). Now
suppose j!
x0 is an unobservable mode of ( R;1DC R;1B ; H ). Then there exists
an x0 = x1 6= 0 such that
2
Bisection Algorithm
Lemma 4.7 suggests the following bisection algorithm to compute RL1 norm:
(a) select an upper bound
u and a lower bound
l such that
l kGk1
u ;
(b) if (
u ;
l )=
l specied level, stop; kGk (
u +
l )=2. Otherwise go to next
step;
(c) set
= (
l +
u )=2;
(d) test if kGk1 <
by calculating the eigenvalues of H for the given
;
(e) if H has an eigenvalue on j R set
l =
; otherwise set
u =
; go back to step (b).
Of course, the above algorithm applies to H1 norm computation as well. Thus L1
norm computation requires a search, over either
or !, in contrast to L2 (H2 ) norm
computation, which does not. A somewhat analogous situation occurs for constant
matrices with the norms kM k22 = trace(M M ) and kM k1 = [M ]. In principle, kM k22
can be computed exactly with a nite number of operations, as can the test for whether
(M ) <
(e.g.
2 I ; M M > 0), but the value of (M ) cannot. To compute (M ),
we must use some type of iterative algorithm.
116
PERFORMANCE SPECIFICATIONS
Remark 4.4 It is clear that kGk1 < i ;1G 1 < 1. Hence, there is no loss of
118
uncertainty. Indeed, this requirement was the original motivation for the development
of feedback systems. Feedback is only required when system performance cannot be
achieved because of uncertainty in system characteristics. The more detailed treatment
of model uncertainties and their representations will be discussed in Chapter 9.
For the moment, assuming we are given a model including a representation of uncertainty which we believe adequately captures the essential features of the plant, the
next step in the controller design process is to determine what structure is necessary
to achieve the desired performance. Preltering input signals (or open loop control)
can change the dynamic response of the model set but cannot reduce the eect of uncertainty. If the uncertainty is too great to achieve the desired accuracy of response,
then a feedback structure is required. The mere assumption of a feedback structure,
however, does not guarantee a reduction of uncertainty, and there are many obstacles
to achieving the uncertainty-reducing benets of feedback. In particular, since for any
reasonable model set representing a physical system uncertainty becomes large and the
phase is completely unknown at suciently high frequencies, the loop gain must be small
at those frequencies to avoid destabilizing the high frequency system dynamics. Even
worse is that the feedback system actually increases uncertainty and sensitivity in the
frequency ranges where uncertainty is signicantly large. In other words, because of the
type of sets required to reasonably model physical systems and because of the restriction
that our controllers be causal, we cannot use feedback (or any other control structure)
to cause our closed-loop model set to be a proper subset of the open-loop model set.
Often, what can be achieved with intelligent use of feedback is a signicant reduction
of uncertainty for certain signals of importance with a small increase spread over other
signals. Thus, the feedback design problem centers around the tradeo involved in reducing the overall impact of uncertainty. This tradeo also occurs, for example, when
using feedback to reduce command/disturbance error while minimizing response degradation due to measurement noise. To be of practical value, a design technique must
provide means for performing these tradeos. We will discuss these tradeos in more
detail later in section 5.5 and in Chapter 6.
To focus our discussion, we will consider the standard feedback conguration shown
in Figure 5.1. It consists of the interconnected plant P and controller K forced by
r-e
;6
di
- ?e up-
- ?e y?en
119
command r, sensor noise n, plant input disturbance di , and plant output disturbance
d. In general, all signals are assumed to be multivariable, and all transfer matrices are
assumed to have appropriate dimensions.
Assume that the plant P and the controller K in Figure 5.1 are xed real rational
proper transfer matrices. Then the rst question one would ask is whether the feedback
interconnection makes sense or is physically realizable. To be more specic, consider a
simple example where
; 1; K = 1
P = ; ss +
2
In order to be consistent with the notation used in the rest of the book, we shall
denote
K^ := ;K
and regroup the external input signals into the feedback loop as w1 and w2 and regroup
the input signals of the plant and the controller as e1 and e2 . Then the feedback loop
with the plant and the controller can be simply represented as inFigure 5.2 and the
1
system is well-posed if and only if the transfer matrix from w
w2 to e1 exists and is
proper.
Lemma 5.1 The feedback system in Figure 5.2 is well-posed if and only if
I ; K^ (1)P (1)
(5:1)
is invertible.
120
- e e1
+6
P
K^
+ w
e2 ?
e+ 2
Proof. The system in the above diagram can be represented in equation form as
^ 2
e1 = w1 + Ke
e2 = w2 + Pe1 :
Then an expression for e1 can be obtained as
^ )e1 = w1 + Kw
^ 2:
(I ; KP
^ );1 exists and is proper.
Thus well-posedness is equivalent to the condition that (I ; KP
But this is equivalent to the condition that the constant term of the transfer function
^ is invertible.
I ; KP
2
It is straightforward to show that (5.1) is equivalent to either one of the following
two conditions:
I ;K^ (1)
is invertible;
(5:2)
;P (1)
I
A B
P=
"
C D
A^
K^ = ^
C
B^
:
D^
(5:3)
(5:4)
Fortunately, in most practical cases we will have D = 0, and hence well-posedness for
most practical control systems is guaranteed.
121
x_
e2
x^_
e1
=
=
=
=
Ax + Be1
Cx + De1
^2
A^x^ + Be
^ 2:
C^ x^ + De
(5.6)
(5.7)
(5.8)
(5.9)
Denition 5.2 The system of Figure 5.2 is said to be internally stable if the origin
(x; x^) = (0; 0) is asymptotically stable, i.e., the states (x; x^) go to zero from all initial
states when w1 = 0 and w2 = 0.
Note that internal stability is a state space notion. To get a concrete characterization
of internal stability, solve equations (5.7) and (5.9) for e1 and e2 :
e I ;D^ ;1
1 =
x
0 C^
e2
x^ :
;D I
C 0
Note that the existence of the inverse is guaranteed by the well-posedness condition.
Now substitute this into (5.6) and (5.8) to get
x_
~ x
x^_ = A x^
where
0 C^ :
;D I
C 0
Thus internal stability is equivalent to the condition that A~ has all its eigenvalues in
the open left-half plane. In fact, this can be taken as a denition of internal stability.
A~ = A0 A0^ + B0 B0^
I ;D^ ;1
Lemma 5.2 The system of Figure 5.2 with given stabilizable and detectable realizations
for P and K^ is internally stable if and only if A~ is a Hurwitz matrix.
It is routine to verify that the above denition of internal stability depends only on
P and K^ , not on specic realizations of them as long as the realizations of P and K^ are
both stabilizable and detectable, i.e., no extra unstable modes are introduced by the
realizations.
The above notion of internal stability is dened in terms of state-space realizations
of P and K^ . It is also important and useful to characterize internal stability from the
122
transfer matrix point of view. Note that the feedback system in Figure 5.2 is described,
in term of transfer matrices, by
I ;K^ e w
1
1
(5:10)
e = w :
;P I
2
Now it is intuitively clear that if the system in Figure 5.2 is internally stable, then for all
bounded inputs (w1 ; w2 ), the outputs (e1 ; e2) are also bounded. The following lemma
shows that this idea leads to a transfer matrix characterization of internal stability.
Lemma 5.3 The system in Figure 5.2 is internally stable if and only if the transfer
matrix
(I ; P K^ );1 P
from (w1 ; w2 ) to (e1 ; e2 ) belongs to RH1 .
;P
(I ; P K^ );1
(5:11)
A B " A^ B^ #
Proof. As above let C D and ^ ^ be stabilizable and detectable realizaC D
tions of P and K^ , respectively. Let y1 denote the output of P and y2 the output of K^ .
Then the state-space equations for the system in Figure 5.2 are
x_
_
yx^
y2
e1
e2
=
=
=
A
x B 0 e
1
x^ + 0 B^
e2
x D 0 e
1
x^ + 0 D^
e2
w 0 I y
1 +
1 :
0
0 A^
C 0
0 C^
w2
I 0
y2
x w
1
x^ + w2 :
Now suppose that this system is internally stable. Then the well-posedness condition
implies that (I ; DD^ ) = (I ; P K^ )(1) is invertible. Hence, (I ; P K^ ) is invertible.
Furthermore, since the eigenvalues of
A~ = A0 A0^ + B0 B0^
I ;D^ ;1
;D
0 C^
C 0
are in the open left-half plane, it follows that the transfer matrix from (w1 ; w2 ) to (e1 ; e2 )
given in (5.11) is in RH1 .
123
;D
e1
1
is nonsingular. Now routine calculations give the transfer matrix from w
w2 to e2
in terms of the state space realizations:
;D
0 C^ (sI ; A~);1 B 0
C 0
0 B^
124
Remark 5.1 It should be noted that internal stability is a basic requirement for a
practical feedback system. This is because all interconnected systems may be unavoidably subject to some nonzero initial conditions and some (possibly small) errors, and
it cannot be tolerated in practice that such errors at some locations will lead to unbounded signals at some other locations in the closed-loop system. Internal stability
guarantees that all signals in a system are bounded provided that the injected signals
(at any locations) are bounded.
~
However, there are some special cases under which determining system stability is
simple.
Corollary 5.4 Suppose K^ 2 RH1. Then the system in Figure 5.2 is internally stable
i (I ; P K^ );1 P 2 RH1 .
Proof. The necessity is obvious. To prove the suciency, it is sucient to show that
(I ; P K^ );1 2 RH1 . But this follows from
(I ; P K^ );1 = I + (I ; P K^ );1 P K^
and (I ; P K^ );1 P; K^ 2 RH1 .
2
This corollary is in fact the basis for the classical control theory where the stability
is checked only for one closed-loop transfer function with the implicit assumption that
the controller itself is stable. Also, we have
Corollary 5.5 Suppose P 2 RH1. Then the system in Figure 5.2 is internally stable
i K^ (I ; P K^ );1 2 RH1 .
Corollary 5.6 Suppose P 2 RH1 and K^ 2 RH1 . Then the system in Figure 5.2 is
internally stable i (I ; P K^ );1 2 RH1 .
To study the more general case, dene
125
Proof. It is easy to show that P K^ and (I ; P K^ );1 have the following realizations:
3
2
A B C^ B D^
P K^ = 64 0 A^ B^ 75
C DC^ DD^
"
A B
(I ; P K^ );1 =
C D
where
A =
B =
A BC^ BD^
^ );1 C DC^
+
(
I
;
D
D
^
^
0 A
B
BD^
^ ;1
B^ (I ; DD)
C = (I ; DD^ );1 C DC^
D = (I ; DD^ );1 :
It is also easy to see that A = A~. Hence, the system is internally stable i A is stable.
Now suppose that the system is internally stable, then (I ; P K^ );1 2 RH1 . This
implies that all zeros of det(I ; P (s)K^ (s)) must be in the left-half plane. So we only
need to show that given condition (ii), condition (i) is necessary and sucient for the
B ) is stabilizable i
internal stability. This follows by noting that (A;
A BC^ BD^
; B^
(5:12)
0 A^
A) is detectable i
is stabilizable; and (C;
A BC^
^
;
(5:13)
C DC
0 A^
is detectable. But conditions (5.12) and (5.13) are equivalent to condition (i), i.e., P K^
has no unstable pole/zero cancelations.
2
With this observation, the MIMO version of the Nyquist stability theorem is obvious.
Theorem 5.8 (Nyquist Stability Theorem) The system is internally stable if and only
if condition (i) in Theorem 5.7 is satised and the Nyquist plot of (j!) for ;1 !
1 encircles the origin, (0; 0), nk + np times in the counter-clockwise direction.
Proof. Note that by SISO Nyquist stability theorem, (s) has all zeros in the open
left-half plane if and only if the Nyquist plot of (j!) for ;1 ! 1 encircles the
origin, (0; 0), nk + np times in the counter-clockwise direction.
126
Recall that two polynomials m(s) and n(s), with, for example, real coecients, are said
to be coprime if their greatest common divisor is 1 (equivalent, they have no common
zeros). It follows from Euclid's algorithm1 that two polynomials m and n are coprime
i there exist polynomials x(s) and y(s) such that xm + yn = 1; such an equation is
called a Bezout identity. Similarly, two transfer functions m(s) and n(s) in RH1 are
said to be coprime over RH1 if there exists x; y 2 RH1 such that
xm + yn = 1:
The more primitive, but equivalent, denition is that m and n are coprime if every
common divisor of m and n is invertible in RH1 , i.e.,
h; mh;1; nh;1 2 RH1 =) h;1 2 RH1 :
More generally, we have
Denition 5.3 Two matrices M and N in RH1 are right coprime over RH1 if they
have the same number of columns and if there exist matrices Xr and Yr in RH1 such
that
X Y M = X M + Y N = I:
r r
r
r
N
Similarly, two matrices M~ and N~ in RH1 are left coprime over RH1 if they have the
same number of rows and if there exist matrices Xl and Yl in RH1 such that
M~ N~ Xl = MX
~ l + NY
~ l = I:
Yl
M
Note that these denitions are equivalent to saying that the matrix N is left
invertible in RH1 and the matrix M~ N~ is right-invertible in RH1 . These two
equations are often called Bezout identities.
Now let P be a proper real-rational matrix. A right-coprime factorization (rcf)
of P is a factorization P = NM ;1 where N and M are right-coprime over RH1 .
Similarly, a left-coprime factorization (lcf) has the form P = M~ ;1N~ where N~ and M~
are left-coprime over RH1 . A matrix P (s) 2 Rp (s) is said to have double coprime
factorization if there exist a right coprime factorization P = NM ;1 , a left coprime
factorization P = M~ ;1N~ , and Xr ; Yr ; Xl ; Yl 2 RH1 such that
X Y M ;Y
r
r
l
N Xl = I:
;N~ M~
(5:14)
Of course implicit in these denitions is the requirement that both M and M~ be square
and nonsingular.
1
127
A B
P=
C D
is a stabilizable and detectable realization. Let F and L be such that A + BF and A + LC
are both stable, and dene
M ;Y 2 A + BF B ;L 3
l
5
4
N Xl = C +FDF DI 0I
X Y 2 A + LC ;(B + LD) L 3
r
r
4 F
I
0 5:
;N~ M~ =
;D
(5:15)
(5:16)
Then P = NM ;1 = M~ ;1N~ are rcf and lcf, respectively, and, furthermore, (5.14) is
satised.
128
BF B
N (s) = CA +
+ DF D :
Therefore
u = Mv; y = Nv
so that y = NM ;1 u, i.e., P = NM ;1 .
We shall now see how coprime factorizations can be used to obtain alternative characterizations of internal stability conditions. Consider again the standard stability analysis
diagram in Figure 5.2. We begin with any rcf's and lcf's of P and K^ :
P = NM ;1 = M~ ;1 N~
(5:17)
~
K^ = UV ;1 = V~ ;1 U:
(5:18)
Lemma 5.10 Consider the system in Figure 5.2. The following conditions are equivalent:
1. The feedback system is internally stable.
M U
2. N V is invertible in RH1 .
V~ ;U~
I K^
=
P I
I K^ ;1
2 RH1 :
P I
M ;1
I
UV ;1 = M U
;1
NM
N V
I K^ ;1 M
=
P I
M U ;1
:
V
N V
0
(5:19)
0
V ;1
M
129
0
M U
0 V ; N V
are right-coprime (this fact is left as an exercise for the reader), (5.19) holds i
M U ;1
2 RH1 :
N V
This proves the equivalence of conditions 1 and 2. The equivalence of 1 and 3 is proved
similarly.
The conditions 4 and 5 are implied by 2 and 3 from the following equation:
V~ ;U~ M U V~ M ; UN
~
0
:
~ ; NU
~
N V =
;N~ M~
0
MV
Since the left hand side of the above equation is invertible in RH1 , so is the right hand
side. Hence, conditions 4 and 5 are satised. We only need to show that either condition
4 or condition 5 implies condition 1. Let us show condition 5 ! 1; this is obvious since
I K^ ;1
I V~ ;1U~ ;1
= NM ;1
I
P I
=
M
0 I
V~ M U~ ;1 V~
N
~ ~ ;1
if VNM UI
2 RH1 or if condition 5 is satised.
0
0 I 2 RH1
Combining Lemma 5.10 and Theorem 5.9, we have the following corollary.
Corollary 5.11 Let P be a proper real-rational matrix and P = NM ;1 = M~ ;1N~ be
corresponding rcf and lcf over RH1 . Then there exists a controller
K^ 0 = U0 V0;1 = V~0;1 U~0
with U0 ; V0 ; U~0 , and V~0 in RH1 such that
V~ ;U~ M U I 0
0
0
0
(5:20)
N V0 = 0 I :
;N~ M~
Furthermore, let F and L be such that A + BF and A + LC are stable. Then a particular
set of state space realizations for these matrices can be given by
M U 2 A + BF B ;L 3
0
5
4
(5:21)
N V0 = C +FDF DI I0
V~ ;U~ 2 A + LC ;(B + LD) L 3
0
0
4 F
(5:22)
I
0 5:
;N~ M~ =
;D
130
Proof. The idea behind the choice of these matrices is as follows. Using the observer
theory, nd a controller K^ 0 achieving internal stability; for example
K^ 0 := A + BF +FLC + LDF ;0L :
(5:23)
Perform factorizations
K^ 0 = U0 V0;1 = V~0;1 U~0
which are analogous to the ones performed on P . Then Lemma 5.10 implies that each
of the two left-hand side block matrices of (5.20) must be invertible in RH1 . In fact,
(5.20) is satised by comparing it with the equation (5.14).
2
r-e
;6
di
- ?e up-
- ?e y?en
131
Ti = I ; Si = Li (I + Li );1
To = I ; So = Lo(I + Lo);1 ;
respectively. (The word complementary is used to signify the fact that T is the complement of S , T = I ; S .) The matrix I + Li is called input return dierence matrix and
I + Lo is called output return dierence matrix.
It is easy to see that the closed-loop system, if it is internally stable, satises the
following equations:
y
r;y
u
up
=
=
=
=
To (r ; n) + So Pdi + So d
So (r ; d) + Ton ; So Pdi
KSo (r ; n) ; KSod ; Ti di
KSo (r ; n) ; KSod + Si di :
(5.24)
(5.25)
(5.26)
(5.27)
These four equations show the fundamental benets and design objectives inherent in
feedback loops. For example, equation (5.24) shows that the eects of disturbance d
on the plant output can be made \small" by making the output sensitivity function So
small. Similarly, equation (5.27) shows that the eects of disturbance di on the plant
input can be made small by making the input sensitivity function Si small. The notion
of smallness for a transfer matrix in a certain range of frequencies can be made explicit
using frequency dependent singular values, for example, (So ) < 1 over a frequency
range would mean that the eects of disturbance d at the plant output are eectively
desensitized over that frequency range.
Hence, good disturbance rejection at the plant output (y) would require that
;
(So ) = (I + PK );1 = (I +1 PK ) ; (for disturbance at plant output, d)
;
(So P ) = (I + PK );1P = (PSi ); (for disturbance at plant input, di )
be made small and good disturbance rejection at the plant input (up ) would require
that
;
(Si ) = (I + KP );1 = (I +1 KP ) ; (for disturbance at plant input, di )
;
(Si K ) = K (I + PK );1 = (KSo); (for disturbance at plant output, d)
be made small, particularly in the low frequency range where d and di are usually
signicant.
Note that
132
then
1
(
S
)
o
(PK ) + 1
(PK ) ; 1 ; if (PK ) > 1
1
1
(KP ) + 1 (Si ) (KP ) ; 1 ; if (KP ) > 1:
These equations imply that
(So ) 1 () (PK ) 1
(Si ) 1 () (KP ) 1:
Now suppose P and K are invertible, then
133
high frequency range, which in turn implies that the loop gain, (Lo ), should be small
at those frequencies.
Still another tradeo is with the sensor noise error reduction. The con
ict between
the disturbance rejection and the sensor noise reduction is evident in equation (5.24).
Large (Lo(j!)) values over a large frequency range make errors due to d small. However, they also make errors due to n large because this noise is \passed through" over
the same frequency range, i.e.,
y = To (r ; n) + So Pdi + So d (r ; n):
Note that n is typically signicant in the high frequency range. Worst still, large loop
gains outside of the bandwidth of P , i.e., (Lo (j!)) 1 or (Li (j!)) 1 while
(P (j!)) 1, can make the control activity (u) quite unacceptable, which may cause
the saturation of actuators. This follows from
u = KSo(r ; n ; d) ; Ti di = Si K (r ; n ; d) ; Ti di P ;1 (r ; n ; d) ; di :
Here, we have assumed P to be square and invertible for convenience. The resulting
equation shows that disturbances and sensor noise are actually amplied at u whenever
the frequency range signicantly exceeds the bandwidth of P , since for ! such that
(P (j!)) 1, we have
[P ;1 (j!)] = [P (1j!)] 1:
Similarly, the controller gain, (K ), should also be kept not too large in the frequency
range where the loop gain is small in order to not saturate the actuators. This is because
for small loop gain (Lo (j!)) 1 or (Li (j!)) 1
u = KSo (r ; n ; d) ; Ti di K (r ; n ; d):
Therefore, it is desirable to keep (K ) not too large when the loop gain is small.
To summarize the above discussion, we note that good performance requires in some
frequency range, typically some low frequency range (0; !l ):
(PK ) 1; (KP ) 1; (K ) 1
and good robustness and good sensor noise rejection require in some frequency range,
typically some high frequency range (!h ; 1)
(PK ) 1; (KP ) 1; (K ) M
where M is not too large. These design requirements are shown graphically in Figure 5.4.
The specic frequencies !l and !h depend on the specic applications and the knowledge
one has on the disturbance characteristics, the modeling uncertainties, and the sensor
noise levels.
134
HHH
@ H6H
@ H
XXXX @@ @@@(L)
;;;;;;;ZZ @@ @
;;;;;;;;ZZ @ @
;;; ;;;;;Z S @
;; ;;;;; ; ZS @Z
;;;;; ;;; SH Z
!h
;;;; ;;;;;
Z
H
;; ;;; ; !l HHH ZZ ;;;;;;;; log !
;;;;;
H;;;;;;;;;;
(L)SS @@HH
;
SS @@HH;;H;;H;;
S
@@ H
S
Figure 5.4: Desired Loop Gain
135
Although the above loop shaping design can be eective in some of applications, there
are severe intrinsic limitations. Some of these limitations are listed below:
The loop shaping technique described above can only eectively deal with problems with uniformly specied performance and robustness specications. More
specically, the method can not eectively deal with problems with dierent specications in dierent channels and/or problems with dierent uncertainty characteristics in dierent channels without introducing signicant conservatism. To
illustrate this diculty, consider an uncertain dynamical system
P = (I + )P
where P is the nominal plant and is the multiplicative modeling error. Assume
that can be written in the following form
~ < 1:
= ~ Wt ; ()
Then, for robust stability, we would require (To ) = (~ Wt To) < 1 or (Wt To)
1. If a uniform bound is required on the loop gain to apply the loop shaping technique, we would need to overbound (Wt To ):
(Wt To ) (Wt ) (To ) (Wt ) 1 ;(L(oL) ) ; if (Lo ) < 1
o
and the robust stability requirement is implied by
1 ; if (L ) < 1:
(Lo ) (W1) + 1 (W
o
t
t)
136
6
(Ws )
;
;;;;;;;;;Z;ZZ
; ; ; ; ; ; ;ZZ
;;;;;;;;;;; ;;; ZZ
;; ; ;;;;;;;;;;;Z
;;;;;; ; ;@ ;@ ;@ ;@;;@ @ @ @ @ ;;;;;;;;;P;;@P;P;@;;@ @ @ @ @ @@ log !
;; ; ; ; ; ; ;@PP@P@P@ @ @ @ @
P@P @@ @@ @
; ; ; ; ; ; ;;7
1
(Wt )
2
Ws = 4
s+1
(s+1)(s+2)
s+1
3
2
5 ; Wt = 4
s+2
s+10
(s+1)
s+10
s+2
s+10
3
5:
137
It is easy to show that for large , the weighting functions are as shown in Figure 5.5, and thus the above loop shaping technique cannot be applied. However,
it is also easy to show that the system with controller
K = I2
gives
0 ; W T = s+10 0
Ws S = s+2
t
1
1
0 s+2
0 s+10
jws (j!)s(j!)j 1; 8!
with
;1 8! ! ;
0
jws (j!)j = ;1
8! > !0 :
138
6u~
d~i
Wu
- Wr
r-e
;6
d~
Wi
di
- ?e -
Wd
d
- ?e y- We ee?
n
Wn
n~
139
these weights may be treated as tuning parameters which are chosen by the designer
to achieve the best compromise between the con
icting objectives. The selection of the
weighting matrices should be guided by the expected system inputs and the relative
importance of the outputs.
Hence, control design may be regarded as a process of choosing a controller K such
that certain weighted signals are made small in some sense. There are many dierent
ways to dene the smallness of a signal or transfer matrix, as we have discussed in
the last chapter. Dierent denitions lead to dierent control synthesis methods, and
some are much harder than others. A control engineer should make a judgment of the
mathematical complexity versus engineering requirements.
Below, we introduce two classes of performance formulations: H2 and H1 criteria.
For the simplicity of presentation, we shall assume di = 0 and n = 0.
H2 Performance
E kek22 = E
Z 1
0
kek22 + 2 ku~k22
o
We SoWd
=
W KS W
u o d
2
2
140
with some appropriate choice of weighting matrix Wu and scalar . The parameter
clearly denes the tradeo we discussed earlier between good disturbance rejection at
the output and control eort (or disturbance and sensor noise rejection at the actuators).
Note that can be set to = 1 by an appropriate choice of Wu . This problem can
be viewed as minimizing the energy consumed by the system in order to reject the
disturbance d.
This type of problem was the dominant paradigm in the 1960's and 1970's and is
usually referred to as Linear Quadratic Gaussian Control or simply as LQG. (They
will also be referred to as H2 mixed sensitivity problems for the consistency with the
H1 problems discussed next.) The development of this paradigm stimulated extensive
research eorts and is responsible for important technological innovation, particularly
in the area of estimation. The theoretical contributions include a deeper understanding
of linear systems and improved computational methods for complex systems through
state-space techniques. The major limitation of this theory is the lack of formal treatment of uncertainty in the plant itself. By allowing only additive noise for uncertainty,
the stochastic theory ignored this important practical issue. Plant uncertainty is particularly critical in feedback systems.
H1 Performance
Although the H2 norm (or L2 norm) may be a meaningful performance measure and
although LQG theory can give ecient design compromises under certain disturbance
and plant assumptions, the H2 norm suers a major deciency. This deciency is due
to the fact that the tradeo between disturbance error reduction and sensor noise error
reduction is not the only constraint on feedback design. The problem is that these
performance tradeos are often overshadowed by a second limitation on high loop gains
{ namely, the requirement for tolerance to uncertainties. Though a controller may be
designed using FDLTI models, the design must be implemented and operated with a real
physical plant. The properties of physical systems, in particular the ways in which they
deviate from nite-dimensional linear models, put strict limitations on the frequency
range over which the loop gains may be large.
A solution to this problem would be to put explicit constraints on the loop gain in
the penalty function. For instance, one may chose to minimize
sup kek2 = kWe So Wd k1 ;
kd~k2 1
kd~k2 1
kd~k2 1
o
2
So Wd
:
kek22 + 2 ku~k22 =
WWueKS
o Wd
1
141
This problem can also be regarded as minimizing the maximum power of the error
subject to all bounded power disturbances: let
e
e^ := u~
W S W W S W
Se^e^ = W eKSo Wd Sd~d~ W eKSo Wd
u o d
u o d
then
and
kdkP 1
kdkP 1
1 Z1
We SoWd
2 ;1 Trace Se~e~(j!) d! =
Wu KSo Wd
2
:
1
Alternatively, if the system robust stability margin is the major concern, the weighted
complementary sensitivity has to be limited. Thus the whole cost function may be
W S W
We To Wd
1 o 2
1
where W1 and W2 are the frequency dependent uncertainty scaling matrices. These
design problems are usually called H1 mixed sensitivity problems. For a scalar system,
an H1 norm minimization problem can also be viewed as minimizing the maximum
magnitude of the system's steady-state response with respect to the worst case sinusoidal
inputs.
142
Performance Limitations
This chapter introduces some multivariable versions of the Bode's sensitivity integral relations and Poisson integral formula. The sensitivity integral relations are used to study
the design limitations imposed by bandwidth constraints and the open-loop unstable
poles, while the Poisson integral formula is used to study the design constraints imposed
by the non-minimum phase zeros. These results display that the design limitations in
multivariable systems are dependent on the directionality properties of the sensitivity
function as well as those of the poles and zeros, in addition to the dependence upon pole
and zero locations which is known in single-input single-output systems. These integral
relations are also used to derive lower bounds on the singular values of the sensitivity
function which display the design tradeos.
6.1 Introduction
One important problem that arises frequently is concerned with the level of performance
that can be achieved in feedback design. It has been shown in the previous chapters that
the feedback design goals are inherently con
icting, and a tradeo must be performed
among dierent design objectives. It is also known that the fundamental requirements
such as stability and robustness impose inherent limitations upon the feedback properties irrespective of design methods, and the design limitations become more severe in
the presence of right-half plane zeros and poles in the open-loop transfer function.
An important tool that can be used to quantify feedback design constraints is furnished by the Bode's sensitivity integral relation and the Poisson integral formula. These
integral formulae express design constraints directly in terms of the system's sensitivity
143
PERFORMANCE LIMITATIONS
144
A ; zI B
=0
C
holds, where = 1, and is called the input zero direction associated with z . Analogously, a transmission zero z of L(s) satises the relation
x w A ; zI B = 0;
C D
where x and w are some vectors with w satisfying the condition w w = 1. The vector
w is called the output zero direction associated with z . Note also that p 2 C is a pole of
L(s) if and only if it is a zero of L;1(s). By a slight abuse of terminology, we shall call
the input and output zero directions of L;1 (s) the input and output pole directions of
L(s), respectively.
In the sequel we shall preclude the possibility that z is both a zero and pole of L(s).
Then, by Lemma 3.27 and 3.28, z is a zero of L(s) if and only if L(z ) = 0 for some
vector , = 1, or w L(z ) = 0 for some vector w, w w = 1. Similarly, p is a pole of
L(s) if and only if L;1(p) = 0 for some vector , = 1, or w L;1(p) = 0 for some
vector w, w w = 1.
It is well-known that a non-minimum phase transfer function admits a factorization
that consists of a minimum phase part and an all-pass factor. Let zi 2 C + , i = 1; ; k,
be the non-minimum phase zeros of L(s) and let i ; i i = 1; be the input directions
generated from the following iterative procedure
145
i
R!1
sup
s2C+
jsj R
R (L(s)) = 0
PERFORMANCE LIMITATIONS
146
Each of these assumptions has important implications. Assumption 6.1 implies that
the sensitivity function S (s) is analytic in C + . Assumption 6.2 states that the open-loop
transfer function has a rollo rate of more than one pole-zero excess. Note that most
of practical systems require a rollo rate of more than one pole-zero excess in order to
maintain a sucient stability margin. One instance for this assumption to hold is that
each element of L has a rollo rate of more than one pole-zero excess.
Suppose that the open-loop transfer function L(s) has poles pi in the open right-half
plane with input pole directions i , i = 1; ; k, which are obtained through a similar
iterative procedure as in the last section.
Lemma 6.1 Let L(s) and S (s) be dened by (6.1). Then p 2 C is a zero of S (s) with
zero direction if and only if it is a pole of L(s) with pole direction .
and L;1(p) = 0. However, S (s) = (I + L(s));1 = I + L;1 (s) L;1 (s). Hence,
S (p) = 0. This establishes the suciency part. The proof for necessity follows by
reversing the above procedure.
2
Then the sensitivity function S (s) can be factorized as
Z1
0
0k
1
X
ln (S (j!))d! max @ (Repj )j j A :
j =1
(6:4)
Z1
0
(6:5)
Z1
0
ln (S (j!))d!
k
X
i=1
Repi :
(6:6)
147
Z1
0
ln jS (j!)jd! =
k
X
i=1
Repi
Z1
0
Rep 0
1
1
1 2
0 Rep2
2
Rep 0
1
1
1 2
max
Re0p Re0p2 21
= amx
= max
1 2
2 1 1
Rep1
(Rep1 )1 2 :
(Rep2 )2 1
Rep2
0
Rep2
PERFORMANCE LIMITATIONS
148
The utility of this corollary is clear. This result fully characterizes the limitation
imposed by a pair of open-loop unstable poles on the sensitivity reduction properties.
This limitation depends not only on the relative distances of the poles to the imaginary
axis, but also on the principal angle between the two pole directions.
Next, we investigate the design constraints imposed by open-loop non-minimum
phase zeros upon sensitivity properties. The results below may be considered to be a
matrix extension of the Poisson integral relation.
max
R!1 2[;=2;=2]
ln (S (Rej ))
R
= 0:
(6:8)
Then, for any non-minimum phase zero z = x0 + jy0 2 C + of L(s) with output direction
w, w w = 1,
Z1
Z1
(6:9)
(6:10)
Note that the condition (6.8) is satised if L(s) is a proper rational transfer matrix.
Furthermore, for single-input single-output systems (n = 1),
Yk
w Bi;1(z) = Yk z + pi
i=1
i=1 z ; pi
and
Z1
Yk + pi
ln jS (j!)j x2 + (!x0; y )2 d! = ln zz ;
pi :
0
;1
0
i=1
The more interesting result, however, is the inequality (6.10). This result again suggests
that the multivariable sensitivity properties are closely related to the pole and zero
directions. This result implies that the sensitivity reduction ability of the system may
be severely limited by the open-loop unstable poles and non-minimum phase zeros,
especially when these poles and zeros are close to each other and the angles between
their directions are small.
149
Corollary 6.5 Let the assumption in Theorem 6.4 holds. In addition, suppose that the
condition (6.11) is satised. Then, for each open right-half plane zero z 2 C + of L(s)
with output direction w,
; z z
w Bk;1(z) B1;1 (z) ; z ;
kS (s)k1 M
L
and
( )
( )
kS (s)k1 M
L
( )
; z z
( )
( )
(6:12)
(6:13)
Z1
Then the inequality (6.12) follows by applying inequality (6.10) and inequality (6.13)
follows by applying inequality (6.9)
2
The interpretation of Corollary 6.5 is similar to that in single-input single-output
systems. Roughly stated, this result shows that for a non-minimum phase system, its
PERFORMANCE LIMITATIONS
150
sensitivity must increase beyond one at certain frequencies if the sensitivity reduction is
to be achieved at other frequencies. Of particular importance here is that the sensitivity
function will in general exhibit a larger peak in multivariable systems than in singleinput single-output systems, due to the fact that (S (z )) 1.
The design tradeos and limitations on the sensitivity reduction which arise from
bandwidth constraints as well as open-loop unstable poles can be studied using the
extended Bode integral relations. However, these integral relations by themselves do
not mandate a meaningful tradeo between the sensitivity reduction and the sensitivity
increase, since the sensitivity function can be allowed to exceed one by an arbitrarily
small amount over an arbitrarily large frequency range so as not to violate the Bode
integral relations. However, bandwidth constraints in feedback design typically require
that the open-loop transfer function be small above a specied frequency, and that it roll
o at a rate of more than one pole-zero excess above that frequency. These constraints
are commonly needed to ensure stability robustness despite the presence of modeling
uncertainty in the plant model, particularly at high frequencies. One way of quantifying
such bandwidth constraints is by requiring the open-loop transfer function to satisfy
(6:14)
where !H > !L , and MH > 0, k > 0 are some given constants. With the bandwidth
constraint given as such, the following result again shows that the sensitivity reduction
specied by (6.11) can be achieved only at the expense of increasing the sensitivity at
certain frequencies.
Corollary 6.6 Suppose the Assumptions 6.1-6.2 hold. In addition, suppose that the
conditions (6.11) and (6.14) are satised for some !H and !L such that !H > !L .
Then
1 !H!;L!L
!H
(6:15)
max (S (j!)) e
(1 ; ) k(!H ;!L)
ML
!2[!L ;!H ]
where
max
Pk
i=1 (Repi ) i i
!H ; !L
:
Z1
1 Z 1 1 M i
X
M
H
H d!
; ln 1 ; !1+k d! =
1+
k
!H
i=1 !H i !
1 ; !M1+Hk
151
!i
1 1
X
!H
MH
=
1+k
i
i
(1
+
k
)
;
1
!
H
i=1
!
1 1 M !i
X
!
!
M
H
H
H
H
k
1+k = ; k ln 1 ; ! 1+k
H
i=1 i !H
!
H
; k ln(1 ; ):
Then
Z1
0
ln (S (j!))d! =
Z !L
0
ln (S (j!))d! +
Z !H
!L
ln (S (j!))d! +
Z1
Z1
!H
ln (S (j!))d!
!L ln ML + (!H ; !L ) !2max
ln (S (j!)) ;
ln 1 ; !M1+Hk d!
[!L ;!H ]
!H
!
!L ln ML + (!H ; !L ) !2max
ln (S (j!)) ; kH ln(1 ; )
[! ;! ]
L H
The above lower bound shows that the sensitivity can be very signicant in the
transition band.
Z 1 d ln jLj
Yk j!0 + zi
j
j
1
ln coth 2 d + \ j! ; z
\L(j!0 ) =
;1 d
i=1 0 i
where zi 's are assumed to be the right-half plane zeros of L(s) and := ln(!=!0 ). Note
L(j!)j is the slope of the Bode plot which is almost always negative. It follows
that d ln jd
that \L(j!0 ) will be large if the gain L attenuates slowly and small if it attenuates
rapidly. The behavior of \L(j!) is particularly important near the crossover frequency
!c where jL(j!c )j = 1 since + \L(j!c) is the phase margin of the feedback system,
and further the return dierence is given by
+ \L(j! )
c
;
1
j1 + L(j!c )j = j1 + L (j!c )j = 2 sin
2
PERFORMANCE LIMITATIONS
152
which must not be too small for good stability robustness. If + \L(j!c ) is forced to
be very small by rapid gain attenuation, the feedback system will amplify disturbances
0 + z i 0 for each i,
and exhibit little uncertainty tolerance at and near !c . Since \ j!
j!0 ; zi
a non-minimum phase zero contributes an additional phase lag and imposes limitations
upon the rollo rate of the open-loop gain. The con
ict between attenuation rate
and loop quality near crossover is thus clearly evident. A thorough discussion of the
limitations these relations impose upon feedback control design is given by Bode [1945],
Horowitz [1963], and Freudenberg and Looze [1988]. See also Freudenberg and Looze
[1988] for some multivariable generalizations.
In the classical feedback theory, it has been common to express design goals in
terms of the \shape" of the open-loop transfer function. A typical design requires that
the open-loop transfer function have a high gain at low frequencies and a low gain at
high frequencies while the transition should be well-behaviored. The same conclusion
applies to multivariable system where the singular value plots should be well-behaviored
between the transition band.
f s
f s
f s
f s
jy
f s
@ f s
@ f s
@x
@y
G = Gr + a
and a is small in some norm. This model reduction is usually called an additive model
reduction problem. On the other hand, one may also desire that the approximation be
in relative form
Gr = G(I + rel )
so that rel is small in some norm. This is called a relative model reduction problem.
We shall be only interested in L1 norm approximation in this book. Once the norm is
chosen, the additive model reduction problem can be formulated as
inf
deg(Gr )r
kG ; Gr k1
153
154
with appropriate choice of Wi and Wo . We shall see in this chapter how the balanced
realization can give an eective approach to the above model reduction problems.
A B
is a balanced realizaConsider a stable system G 2 RH1 and suppose G =
C D
tion, i.e., its controllability and observability Gramians are equal and diagonal. Denote
the balanced Gramians by , then
A + A + BB = 0
(7:1)
A + A + C C = 0:
(7:2)
0
Now partition the balanced Gramian as = 01 and partition the system
2
accordingly as
2
3
A11 A12
G = 4 A21 A22
C1 C2
B1
B2
D
5:
Then (7.1) and (7.2) can be written in terms of their partitioned matrices as
A11 1 + 1 A11 + B1 B1 = 0
(7.3)
1 A11 + A11 1 + C1 C1 = 0
(7.4)
A21 1 + 2 A12 + B2 B1 = 0
(7.5)
2 A21 + A12 1 + C2 C1 = 0
(7.6)
A22 2 + 2 A22 + B2 B2 = 0
(7.7)
2 A22 + A22 2 + C2 C2 = 0:
(7.8)
The following theorem characterizes the properties of these subsystems.
Theorem 7.1 Assume that 1 and 2 have no diagonal entries in common. Then both
subsystems (Aii ; Bi ; Ci ); i = 1; 2 are asymptotically stable.
155
Proof. It is clearly sucient to show that A11 is asymptotically stable. The proof for
the stability of A22 is similar.
By Lemma 3.20 or Lemma 3.21, 1 can be assumed to be positive denite without
loss of generality. Then it is obvious that i (A11 ) 0 by Lemma 3.19. Assume that
A11 is not asymptotically, then there exists an eigenvalue at j! for some !. Let V be
a basis matrix for Ker(A11 ; j!I ). Then we have
(A11 ; j!I )V = 0
which gives
(7:9)
V (A11 + j!I ) = 0:
(7.10)
(7.11)
Multiplication of (7.11) from the right by V and from the left by V gives V C1 C1 V = 0,
which is equivalent to
C1 V = 0:
Multiplication of (7.11) from the right by V now gives
(A11 + j!I )1 V = 0:
Analogously, rst multiply (7.10) from the right by 1 V and from the left by V 1 to
obtain
B1 1 V = 0:
Then multiply (7.10) from the right by 1 V to get
(A11 ; j!I )21 V = 0:
It follows that the columns of 21 V are in Ker(A11 ; j!I ). Therefore, there exists a
matrix 1 such that
21 V = V 21 :
Since 21 is the restriction of 21 to the space spanned by V , it follows that it is possible
to choose V such that 21 is diagonal. It is then also possible to choose 1 diagonal and
such that the diagonal entries of 1 are a subset of the diagonal entries of 1 .
Multiply (7.5) from the right by 1 V and (7.6) by V to get
156
This is a Sylvester equation in (A21 V ). Because 21 and 22 have no diagonal entries in
common it follows from Lemma 2.7 that
A21 V = 0
(7:12)
A
11
A21
A12
A22
V
0
V
= j! 0
which means that the A-matrix of the original system has an eigenvalue at j!. This
contradicts the fact that the original system is asymptotically stable. Therefore A11
must be asymptotically stable.
2
Corollary 7.2 If has distinct singular values, then every subsystem is asymptotically
stable.
2 ;2 ;2:8284 ;2 3
(s ; 1)(s ; 2) = 4 0
;1 ;1:4142 5
(s + 1)(s + 2)
2
1:4142
1
1:4142 1:4142
s2 ; s + 2 = 4 ;1;:4142
0
0 5
s2 + s + 2
;1:4142 0
1
is also a balanced realization with = I but one of the subsystems is not stable.
2 A A
11
12
G(s) = 4 A21 A22
C1 C2
B1
B2
D
3
5
157
Proof. The stability of Gr follows from Theorem 7.1. We shall now prove the error
bound for the case si = 1 for all i. The case where the multiplicity of i is not equal to
one is more complicated and an alternative proof is given in the next chapter. Hence,
we assume si = 1 and N = n.
Let
(s) := (sI ; A11 );1
(s) := sI ; A22 ; A21 (s)A12
B~ (s) := A21 (s)B1 + B2
C~ (s) := C1 (s)A12 + C2
then using the partitioned matrix results of section 2.3,
G(s) ; Gr (s) = C (sI ; A);1 B ; C1 (s)B1
=
158
These expressions for B~ (j!)B~ (j!) and C~ (j!)C~ (j!) are then substituted into (7.13)
to obtain
=2 + ;1 (j! ) (j! ) + ; (j! ) (j! ) :
[G(j!) ; Gr (j!)] = 1max
2
2
2
2
=2
[G(j!) ; Gr (j!)] = n 1max
1 + ;1(j!) [1 + (j!)]
(7:14)
where := ; (j!) (j!) = ; is an \all pass" scalar function. (This is the only
place we need the assumption of si = 1) Hence j(j!)j = 1.
Using triangle inequality we get
(7:15)
G(s) ; Gr (s) =
by the denition of Ek (s), we have
[G(j!) ; Gr (j!)]
NX
;1
k =r
NX
;1
k=r
Ek (s)
[Ek (j!)] 2
NX
;1
k=r
k+1 :
159
with > 0. Then kG(s)k1 = G(0) = n and G(s) has the following state space
realization
2 2
3
66
G = 666
4
;4
...
;2n
n
77
77
7
n 5
..
.
0
and the controllability and observability Gramians of the realization are given by
i+j
P = Q = 2i+ 2j
160
a very small perturbation to the balanced realization then the perturbed system has
a balanced realization with distinct singular values and P = Q I . This perturbed
system will have the desired properties and this is exactly how the above example is
constructed.
0
1
2
3
2
1.996 1.991 1.9904
Bounds: 2 4i=r+1 i 7.9772 5.9772 3.9818 1.9904
2r+1
2
1.9954 1.9914 1.9904
kG ; GPr k1
kWo (G ; Gr )Wi k1
is made as small as possible. Assume that G; Wi , and Wo have the following state space
realizations
A B
A B
A B
i
i
G = C 0 ; Wi = C D ; Wo = Co Do
i
i
o
o
2 A 0 BC BD 3
" A B #
i
i
6
7
B
C
A
0
0
o
o
Wo GWi = 64 0 0 Ai Bi 75 =:
:
C 0
Do C Co
(7.16)
(7.17)
Then the input weighted Gramian P and the output weighted Gramian Q are dened
by
I
I
n
P := In 0 P 0 ; Q := In 0 Q 0n :
161
It can be shown easily that P and Q satisfy the following lower order equations
A
0
BC
A
P P12
P12 P22
+
A
P P12
P12 P22
BC
A
BD BD
i
=0
(7 18)
Q Q12
Q12 Q22
A 0
BC A
o
+
A 0
BC A
o
Q Q12
Q12 Q22
C D C D
C
C
o
=0
(7 19)
PA + AP + BB = 0
(7:20)
QA + A Q + C C = 0:
Now let T be a nonsingular matrix such that
(7:21)
TPT = (T ;1 ) QT ;1 = 1
2
(i.e., balanced) with 1 = diag(1 Is1 ; : : : ; r Isr ) and 2 = diag(r+1 Isr+1 ; : : : ; n Isn )
and partition the system accordingly as
Gr =
A
11
C1
C2
B1 :
0
Unfortunately, there is generally no known a priori error bound for the approximation
error and the reduced order model Gr is not guaranteed to be stable either.
Gr = G(I + rel )
162
and krel k1 is made as small as possible. rel is usually called the relative error. In
the case where G is square and invertible, this problem can be simply formulated as
G;1 (G ; G )
:
min
r 1
degG r
r
A ; BD;1C ;BD;1
;1
;1 .
D C
D
(a) Then the weighted Gramians P and Q for the frequency weighted balanced realization of G can be obtained as
PA + AP + BB = 0
(7:22)
Q(A ; BD;1 C ) + (A ; BD;1 C ) Q + C (D;1 ) D;1 C = 0:
(7:23)
(b) Suppose the realization for G is weighted balanced, i.e.,
P = Q = diag(1 Is1 ; : : : ; r Isr ; r+1 Isr+1 ; : : : ; N IsN ) = diag(1 ; 2 )
with 1 > 2 > : : : > N 0 and let the realization of G be partitioned compatibly
with 1 and 2 as
2A A B 3
1
11
12
G(s) = 4 A21 A22 B2 5 :
C1
Then
Gr (s) =
A
C2 D
krel k1
kmul k1
N
Y
i=r+1
N
Y
i=r+1
1 + 2i (
1 + 2i (
B1
D
11
C1
1 + 2 + i )
;1
1 + 2 + i )
; 1:
163
Proof. Since the input weighting matrix Wi = I , it is obvious that the input weighted
Gramian is given by P . Now the output weighted transfer matrix is given by
G;1 (G ; D) =
2 A
3 "
0
B
B #
A
;
1
;
1
4 ;BD C A ; BD C 0 5 =:
:
C 0
;1
;1
D C
D C
Q
Q := Q
Q Q
B
G(s) = CA D
and assume that D has full row rank. Without loss of generality, we shall also assume
that D
such that DD = I . Let D? be a matrix with full row rank such
isDnormalized
that D is square and unitary.
?
Lemma 7.6 A complex number z 2 C is a zero of G(s) if and only if z is an uncontrollable mode of (A ; BD C; BD? ).
A B
Proof. Since D has full row rank and G(s) = C D is a minimal realization, z
is a transmission zero of G(s) if and only if
A ; zI B
C
A ; zI B I
C
2 I 0
0
4 ;C I
[D ; D ]
?
0
0
0 I
3
5
164
BD
? :
= A ; BD0 C ; zI BD
I
0
A ; zI B
C
A ; BD C ; zI BD
?
does not have full row rank. By PBH test, this implies that z is a zero of G(s) if and
only if it is an uncontrollable mode of (A ; BD C; BD? ).
2
Corollary 7.7 There exists a matrix C~ such that the augmented system
A B 2 A B 3 G(s)
Ga := C D = 4 C D 5 = G~ (s)
a
a
C~ D?
is minimum phase.
D ;1 C
~
A;B D
C~ = A ; BD C ; BD? C:
?
If the previous model reduction algorithms are applied to the augmented system Ga ,
the corresponding P and Q equations are given by
PA + AP + BB = 0
Q(A ; BDa;1 Ca ) + (A ; BDa;1 Ca ) Q + Ca (Da;1 ) Da;1 Ca = 0:
Moreover, we have
" ^ #
"
#
G(s) = G(s) (I + ); G(s) = G^ (s) (I + )
rel
mul
G~ (s)
G~ (s)
G^~ (s)
G^~ (s)
and
G^ (s) = G(s)(I + rel ); G(s) = G^ (s)(I + mul ):
However, there are in general innitely many choices of C~ and the model reduction
results will in general depend on the specic choice. Hence an appropriate choice of C~
is important. To motivate our choice, note that the equation for Q can be rewritten as
Q(A;BD C )+(A;BD C ) Q;QBD? D? B Q+C C +(C~ ;D? B Q) (C~ ;D?B Q) = 0
165
C~ = D?B Q:
G^ mp = Gmp (I + rel )
and
Gmp = G^ mp (I + mul):
Then G^ := Gap G^ mp has exactly the same right half plane poles and zeros as that of G
and
G^ = G(I + rel )
G = G^ (I + mul):
Unfortunately, this approach may be conservative if the transfer matrix has many nonminimum phase zeros or unstable poles.
An alternative relative/multiplicative model reduction approach, which does not
require that the transfer matrix be minimum phase but does require solving an algebraic
Riccati equation, is the so-called Balanced Stochastic Truncation (BST) method. Let
G(s) 2 RH1 be a square transfer matrix with a state space realization
B
G(s) = CA D
and det(D) 6= 0. Let W (s) 2 RH1 be a minimum phase left spectral factor of
G(s)G (s), i.e,
W (s)W (s) = G(s)G (s):
Then W (s) can be obtained as
A B
W
W (s) =
CW D
166
with
BW = PC + BD
CW = D;1 (C ; BW X )
where P is the controllability Gramian given by
AP + PA + BB = 0
(7.24)
(7:25)
X is stable. The
with AW := A ; BW (DD );1 C such that AW + BW (DD );1 BW
realization G is said to be a balanced stochastic realization if
2I
66 1 s 2Is
P = X = 64
1
...
n Isn
3
77
75
with 1 > 2 > : : : > n 0. i is in fact the i-th Hankel singular value of the so-called
\phase matrix" (W (s));1 G(s).
Theorem 7.8 Let G(s) 2 RH1 have the following balanced stochastic realization
A B 2 A11 A12 B1 3
G(s) = C D = 4 A21 A22 B2 5
C1
C2 D
G^ = A11 B1
C1 D
is stable and
n 1+
;1
Y
i ;1
G (G ; G^)
1
1
;
i
i=r+1
n 1+
;1
Y
i
G^ (G ; G^ )
1
1 ; i ; 1:
i=r+1
167
It is easy to show that the Hankel singular values of the phase function is given by
Y3 1 + i
i=2 1 ; i
; 1 = 2:51522024046226:
168
3
X
i
i=2 1 ; i
= 2:34052280324021
which is smaller than the actual error. Other weighted model reduction methods can
be found in Al-Saggaf and Franklin [1988], Glover [1986,1989], Glover, Limebeer and
Hung [1992], Hung and Glover [1986] and references therein. Discrete time balance
model reduction results can be found in Al-Saggaf and Franklin [1987], Hinrichsen and
Pritchard [1990], and references therein.
where 1 > 2 : : : > k+1 : : : > n are the Hankel singular values of G(s). Moreover,
we show that a square stable transfer function G(s) can be represented as
G(s) = D0 + 1 E1 (s) + 2 E2 (s) + : : : + n En (s)
where Ek (s) are all-pass functions and the partial sum D0 + 1 E1 (s) + 2 E2 (s) + : : : +
k Ek (s) have McMillan degrees k. This representation is obtained by reducing the order
one dimension at a time via optimal Hankel norm approximations. This representation
also gives that
kGk1 2(1 + : : : + n )
and further that there exists a constant D0 such that
kG(s) ; D0 k1 (1 + : : : + n ):
169
170
The above bounds are then used to show that the k-th order optimal Hankel norm
approximation, G^ (s), together with some constant matrix D0 satises
G(s) ; G^ (s) ; D0
1 (k+1 + : : : + n ):
We shall also provide an alternative proof for the error bounds derived in the last chapter
for the truncated balanced realizations using the results obtained in this chapter.
Finally we consider the Hankel operator in discrete time and oer some alternative
proof of the well-known Nehari's theorem.
Let G(s) 2 L1 be a matrix function. The Hankel operator associated with G will be
denoted by ;G and is dened as
;G : H2? 7;! H2
MG
H2?
>
;G
L2
P+
H2
There is a corresponding Hankel operator in the time domain. Let g(t) denote the
inverse (bilateral) Laplace transform of G(s). Then the time domain Hankel operator
is
;g : L2 (;1; 0] 7;! L2 [0; 1)
;g f := P+ (g f ); for f 2 L2 (;1; 0]:
Thus
R 0 g(t ; )f ( )d; t 0;
(;g f )(t) = ;1
0;
t < 0:
Because of the isometric isomorphism property between the L2 spaces in the time domain and in the frequency domain, we have
k;g k = k;G k :
171
Hence, in this book we will use the time domain and the frequency domain descriptions
for Hankel operators interchangeably.
For the interest of this book, we will now further restrict G to be rational, i.e.,
G(s) 2 RL1 . Then G can be decomposed into strictly causal part and anticausal part,
i.e., there are Gs (s) 2 RH2 and Gu (s) 2 RH?2 such that
Z0
;1
y(t) = ;g u(t); t 0
based on the past input u(t); t 0.
In the state space representation, the Hankel operator can be more specically decomposed as the composition of maps from the past input to the initial state and then
172
u(t)
y(t)
Z0
;1
o : C n 7;! L2 [0; 1)
o x0 := CeAt x0 ; t 0:
(If all the data are real, then the two operators become c : L2 (;1; 0] 7;! Rn and
o : Rn 7;! L2 [0; 1).) Clearly, x0 = cu(t) for u(t) 2 L2 (;1; 0] is the system state
at t = 0 due to the past input and y(t) = o x0 ; t 0, is the future output due to the
initial state x0 with the input set to zero.
It is easy to verify that
;g = o c:
L2 (;1; 0]
>
Cn
;g
ZZ
ZZo
ZZ~
- L2[0; 1)
The adjoint operators of c and o can also be obtained easily from their denitions
as follows: let u(t) 2 L2 (;1; 0], x0 2 C n , and y(t) 2 L2 [0; 1), then
h c u; x0 iC n
Z0
;1
h o x0 ; yiL2 [0;1) =
Z1
0
173
Z1
0
where h; iX denotes the inner product in the Hilbert space X . Therefore, we have
c : C n 7;! L2 (;1; 0]
c x0 = B e;A x0 ; 0
and
o : L2 [0; 1) 7;! C n
o y(t) =
Z1
0
eA t C y(t)dt:
Z1
0
B eA(t; )C y(t)dt; 0:
Let Lc and Lo be the controllability and observability Gramians of the system, i.e.,
Lc =
Lo =
Then we have
Z1
Z01
0
eAt BB eA t dt
eAt C CeAt dt:
c c x0 = Lcx0
o ox0 = Lox0
for every x0 2 C n . Thus Lc and Lo are the matrix representations of the operators
c c and o o.
Theorem 8.1 The operator ;g ;g (or ;Gp;G) and the matrix LcLo have the same
nonzero eigenvalues. In particular k;g k = (Lc Lo).
Proof. Let 2 6= 0 be an eigenvalue of ;g ;g , and let 0 6= u 2 L2 (;1; 0] be a corresponding eigenvector. Then by denition
;g ;g u = c o o cu = 2 u:
Pre-multiply (8.1) by c and dene x = cu 2 C n to get
LcLo x = 2 x:
(8:1)
(8:2)
174
v := 1 ;g u 2 L2 [0; 1):
;g u = v
;g v = u:
This pair of vectors (u; v) are called a Schmidt pair of ;g . The proof given above suggests
a way to construct this pair: nd the eigenvalues and eigenvectors of LcLo , i.e., i2 and
xi such that
Lc Loxi = i2 xi :
Then the pairs (ui ; vi ) given below are the corresponding Schmidt pairs:
u = ( 1 L x ) 2 L (;1; 0]; v = x 2 L [0; 1):
i
c i o i
o i
Remark 8.2 As seen in various literature, there are some alternative ways to write a
Hankel operator. For comparison, let us examine some of the alternatives below:
(i) Let v(t) = u(;t) for u(t) 2 L2 (;1; 0], and then v(t) 2 L2 [0; 1). Hence, the
Hankel operator can be written as
;g : L2 [0; 1) 7;! L2 [0; 1) or ;G : H2 7;! H2
(;g v)(t) =
=
175
;^ Gv = P; (Gv); for v 2 H2 :
Now for any v 2 H2 and u 2 H2? , we have
Hence, ;^ G = ;G .
Denition 8.1 The inertia of a general complex, square matrix A denoted In(A) is the
triple ((A); (A); (A)) where
(A) = number of eigenvalues of A in the open right half-plane.
(A) = number of eigenvalues of A in the open left half-plane.
(A) = number of eigenvalues of A on the imaginary axis.
176
Theorem 8.2 Given complex n n and n m matrices A and B, and hermitian matrix
P = P satisfying
AP + PA + BB = 0
(8:3)
then
(1) If (P ) = 0 then (A) (P ); (A) (P ).
(2) If (A) = 0 then (P ) (A); (P ) (A).
21
22
A~ P01 00 + P01 00 A~ + B~ B~ =
(8:4) ) B2 B2 = 0 ) B2 =
(8:4); (8:5) ) A21 P1 = 0 ) A21 =
(8:4) ) A11 P1 + P1 A11 + B1 B1 =
) (by part (1)) (A11 )
(A11 )
but since (A11 ) = (P1 ) = 0
(P1 ) = (A11 ) (A)
(P1 ) = (A11 ) (A):
(8.4)
0
0
0
(8.5)
(8.6)
(8.7)
(P1 )
(P1 )
Theorem 8.3 Given a realization (A; B; C ) (not necessarily stable) with A 2 C nn ,
B 2 C nm , C 2 C mn , then
177
(1) If (A; B; C ) is completely controllable and completely observable the following two
statements are equivalent:
(a) there exists a D such that GG = 2 I where G(s) = D + C (sI ; A);1 B .
(b) there exist P; Q 2 C nn such that
(i) P = P , Q = Q
(ii) AP + PA + BB = 0
(iii) A Q + QA + C C = 0
(iv) PQ = 2 I
(2) Given that part (1b) is satised then there exists a D satisfying
D D = 2 I
D C + BQ = 0
DB + CP = 0
and any such D will satisfy part (1a) (note, observability and controllability are
not assumed).
= G =
B
D
These two transfer functions are identical and both minimal (since (A; B; C ) is assumed
to be minimal), and hence there exists a similarity transformation T relating the statespace descriptions, i.e.,
;A = T (A ; BD C )T ;1
(8.8)
C = TBD
(8.9)
;
1
B = D CT :
(8.10)
Further
(8:9) ) B = D C (T );1
(8.11)
(8:10) ) C = T BD
(8.12)
;
1
(8:8) ) ;A = ;C DB + (T A T )
= T (A ; (T );1 C DB T )(T );1
(8:9) and (8:10) ) = T (A ; BD C )(T );1 :
(8.13)
178
Hence, T and T satisfy identical equations, (8.8) to (8.10) and (8.11) to (8.13), and
minimality implies these have a unique solution and hence T = T .
Now setting
Q = ;T
P = ;T ;1
(8.14)
(8.15)
clearly satises part (1b), equations (i) and (iv). Further, (8.8) and (8.9) imply
TA + A T ; C C = 0
(8:16)
AT ;1 + T ;1A ; T ;1 C CT ;1 = 0
(8:17)
QBB Q = C C
(8:19)
DB Q = ;C
DB = ;CQ;1 = ;CP:
(8.20)
(8.21)
Equations (8.20) and (8.21) imply that the conditions of part (2) are satised. Now
note that
G(s)G = I:
Part (2) follows immediately from the proof of (1b) ) (1a) above.
The following theorem dilates a given transfer function to an all-pass function.
179
Theorem 8.4 Let G(s) = CA DB with A 2 C nn ; B 2 C nm ; C 2 C mn ; D 2
C mm
satisfy
AP + PA + BB = 0
A Q + QA + C C = 0
(8.22)
(8.23)
P = P = diag(1 ; Ir )
Q = Q = diag(2 ; Ir )
(8.24)
(8.25)
for
11 A12 ; B = B1 ; C = C C
A= A
1
2
A21 A22
B2
^ ^
and dene W (s) := A^ B^
C D
A^
B^
C^
D^
=
=
=
=
(8:26)
with
D ; U
(8.27)
(8.28)
(8.29)
(8.30)
B2 = ;C2 U
(8.31)
; = 1 2 ; 2 I:
(8.32)
and
Also dene the error system
A B
E (s) = G(s) ; W (s) = e e
Ce De
with
Then
C ;C^ ; D = D ; D:
^
Ae = A0 A0^ ; Be = B
;
C
=
e
e
^
B
(8:33)
180
(1) (Ae ; Be ; Ce ) satisfy
Ae Pe + Pe Ae + Be Be = 0
Ae Qe + Qe Ae + Ce Ce = 0
with
Pe =
4
2
Qe =
(8.34)
(8.35)
3
1 0
I
0 Ir
0 5
I 0 2 ;;1
3
2 0 ;;
0 Ir 0 5
;; 0 1 ;
Pe Qe = 2 I
(8.36)
(8.37)
(8.38)
Proof. For notational convenience it will be assumed that = 1 and this can be done
(8.39)
(8.40)
and hence B2 B2 = C2 C2 and by Lemma 2.14 there exists a unitary U satisfying (8.31).
(1) The proof of equations (8.34) to (8.38) is by a straightforward calculation, as
follows. To verify (8.34) and (8.36) we need (8.22) which is assumed, together with
A11 + A^ + B1 B^ = 0
(8.41)
A21 + B2 B^ = 0
(8.42)
;
1
;
1
^
^
^
^
A2 ; + 2 ; A + B B = 0
(8.43)
181
B2 (B1 2 + U C1 );;1
(B2 B1 2 ; C2 C1 );;1
((;A21 1 ; A12 )2 + A12 2 + A21 );;1
;A21 ) (8:42)
(8.44)
(8.45)
(8.46)
where B2 B1 and C2 C1 were substituted in (8.45) using the (2,1) blocks of (8.22) and
(8.23), respectively. To verify (8.41)
182
183
2 105s + 250
G(s) = 39(ss ++2)(
s + 5)2 :
3
5
184
In general the error E (j!) will have modulus equal to but E (s) will contain unstable
poles.
3
Example 8.3 Let us nally complete the analysis of this G(s) by permuting the second
and third states in the balanced realization of the last example to obtain 1 = diag(2; 1),
= 21 . We will nd
"
1
15 ;4
4
^
^
^
A^ = ;
;20 ;6 ; B = 4 ; C = 15 3 ; D = ; 2
2
A^ B^
s + 110)
W (s) = ^ ^ = 21 (;(ss2 ++ 123
21s + 10)
C D
2)(;s + 5)2 (s2 ; 21s + 10) :
E (s) = G(s) ; W (s) = ; 12 (;(ss +
+ 2)(s + 5)2 (s2 + 21s + 10)
Note that ;1 ; = diag(;15=2; ;3=4) so that A^ is stable by Theorem 8.4, part (3b).
^ C^ ) is minimal by Theorem 8.4, part (3d).
jE (j!)j = 12 by Theorem 8.4, part (2), (A;^ B;
W (s) is in fact an optimal second-order Hankel norm approximation to G(s).
3
and then Theorem 8.6 shows that the construction of Theorem 8.4 can be used to achieve
this lower bound.
185
Lemma 8.5 Given a stable, rational, p m, transfer function matrix G(s) with Hankel
singular values 1 2 : : : k k+1 k+2 : : : n > 0, then for all G^ (s) stable
and of McMillan degree k
i (G(s) ; G^ (s)) i+k (G(s)); i = 1; : : : ; n ; k;
(8.61)
^
i+k (G(s) ; G(s)) i (G(s)); i = 1; : : : ; n:
(8.62)
In particular,
(8:63)
Proof. We shall prove (8.61) only and the inequality (8.62) follows from (8.61) by
setting
G(s) = (G(s) ; G^ (s)) ; (;G^ (s)):
^ B;
^ C^ ) be a minimal state space realization of G^ (s), then (Ae ; Be ; Ce ) given by
Let (A;
(8.33) will be a state space realization of G(s) ; G^ (s). Now let P = P and Q = Q
satisfy (8.34) and (8.35) respectively (but not necessary (8.36) and (8.37) and write
11 Q12 ; P ; Q 2 Rnn :
P = PP11 PP12 ; Q = Q
11 11
12 Q22
Q
22
12
P = RR
12
R = R011 R
R22
with
= P11 ; R12 R
R22 = P221=2 ; R12 = P12 P22;1=2 ; R11 R11
12
^ B;
^ C^ ) is a minimal realization.)
(P22 > 0 since (A;
i (G(s) ; G^ (s)) = i (PQ) = i (RR Q) = i (R QR)
i In 0 R QR I0n
R
11
R 0 Q
=
i
11
11
11
(8.64)
186
where (8.64) follows from the fact that X is an n k matrix () rank(XX ) k). 2
We can now give a solution to the optimal Hankel norm approximation problem for
square transfer functions.
Theorem 8.6 Given a stable, rational, m m, transfer function G(s) then
^ (s) ; F (s)
, McMillan degree (G^ ) k.
(1) k+1 (G(s)) =
inf
G(s) ; G
G^ 2H1 ;F 2H;1
(2) If G(s) has Hankel singular values 1 2 : : : k > k+1 = k+2 : : : = k+r >
k+r+1 : : : n > 0 then G^ (s) of McMillan degree k is an optimal Hankel norm
; (whose McMillan
approximation to G(s) if and only if there exists F (s) 2 H1
degree can be chosen n + k ; 1) such that E (s) := G(s) ; G^ (s) ; F (s) satises
E (s)E (s) = k2+1 I:
(8:65)
In which case
^ (s)
= k+1 :
(8:66)
G(s) ; G
H
(3) Let G(s) be as in (2) above, then an optimal Hankel norm approximation of McMillan degree k, G^ (s), can be constructed as follows. Let (A; B; C ) be a balanced
realization of G(s) with corresponding
= diag(1 ; 2 ; : : : ; k ; k+r+1 ; : : : ; n ; k+1 ; : : : ; k+r );
^ B;
^ C;
^ D^ ) from equations (8.26) to (8.31). Then
and dene (A;
#
"
A^ B^
^
(8:67)
G(s) + F (s) = ^ ^
C D
; with the McMillan degree of G^ (s) = k and the
where G^ (s) 2 H1 and F (s) 2 H1
McMillan degree of F (s) = n ; k ; r.
Proof. By the denition of L1 norm, for all F (s) 2 H1; , and G^(s) of McMillan degree
k
=
sup
sup
sup
f 2H?2 ;kf k2 1
f 2H?2 ;kf k2 1
f 2H?2 ;kf k2 1
G ; G^
H
k+1 (G(s))
=
P+ (G ; G^ ; F )f
2
P+ (G ; G^ )f
2
(8.68)
187
Proof. (1) is proved in Theorem 8.4, part (2). (2) is obtained from the forms of Pe and
Qe in Theorem 8.4, part (1). F (;s) is used since it will be stable and have well-dened
Hankel singular values.
2
The optimal Hankel norm approximation for non-square case can be obtained by rst
augmenting the function to form a square function. Forexample,
consider a stable, ra
G
tional, p m (p < m), transfer function G(s). Let Ga = 0 be an augmented square
^
G
^
transfer function and let Ga = ^ be the optimal Hankel norm approximation of
G2
Ga such that
^ a
= k+1 (Ga ):
Ga ; G
H
Then
k+1 (G)
G ; G^
H
Ga ; G^ a
H = k+1 (Ga ) = k+1 (G)
i.e., G^ is an optimal Hankel norm approximation of G(s).
188
and all equations of Theorem 8.3 and let A have dimension n1 + n2 with n1 eigenvalues
strictly in the left half plane and n2 < n1 eigenvalues strictly in the right half plane. If
E = Gs + F with Gs 2 RH1 and F 2 RH;1 then,
i = 1; 2; : : :; n1 ; n2
i (Gs ) =
i = n1 ; n2 + 1; : : : ; n1
i;n1 +n2 (F (;s))
A1 0 B1
B
A
4
5
E = 0 A2 B2 = C D ; Rei (A1 ) < 0; Rei (A2 ) > 0;
C1 C2 D
B
B
A
A
1
2
2
1
in which case G = C D , F = C 0 . The equations of Theorem 8.3
1
2
(i)-(iv) are then satised by a transformed P and Q, partitioned as,
21
P
P
Q
Q
11
12
11
P = P P ; Q = Q Q
22
21
22
12
PQ = 2 I implies that,
det(I ; P11 Q11 ) = det(I ; (2 I ; P12 Q21 ))
= det(( ; 2 )I + P12 Q21 )
= ( ; 2 )n1 ;n2 det(( ; 2 )I + Q21 P12 )
= ( ; 2 )n1 ;n2 det(I ; Q22 P22 ):
The result now follows on observing that i (G(s)) = i (P11 Q11 ) and i2 (F (;s)) =
i (Q22 P22 ):
2
Corollary 8.9 Let E (s) = G(s) ; G^(s) ;;F (s) be as dened in part (3) of Theorem 8.6
with G(s); G^ (s) 2 RH1 and F (s) 2 RH1 . Then for i = 1; 2; :::; 2k + r,
i (G ; G^ ) = k+1 (G);
and for i = 1; 2; :::; n ; k ; r,
i+3k+r (G) i (F (;s)) = i+2k+r (G ; G^ ) i+k+r (G)
189
Proof. The construction of E (s) ensures that the all-pass equations are satised and
an inertia argument easily establishes that the A-matrix has precisely n1 = n + k
eigenvalues in the open left half plane and n2 = n ; k ; r in the open right half plane.
Hence Lemma 8.8 can be applied to give the equalities. The inequalities follow from
Lemma 8.5.
2
The following lemma gives the properties of certain optimal Hankel norm approximations when the degree is reduced by the multiplicity of n . In this case some precise
statements on the error and the approximating system can be made.
Proof. Theorem 8.6 gives that A^ 2 R(n;r)(n;r) is stable and hence F (s) can be
chosen to be zero and therefore (G(s) ; G^ (s))=n (G) is all-pass. The i (G^ (s)) are
2
Applying the above reduction procedure again on G^ (s) and repeating until G^ (s) has
zero McMillan degree gives the following new representation of stable systems.
Theorem 8.11 Let G(s) be a stable, rational m m, transfer function with Hankel
G^ k (s) := D0 +
has McMillan degree r1 + r2 + : : : + rk .
k
X
i=1
i Ei (s)
(8:71)
190
Proof. Let G^k (s) be the optimal Hankel norm approximation to G^k+1 (s) (given by
Lemma 8.10) of degree r1 + r2 + : : : + rk , with G^ N (s) := G(s). Lemma 8.10 (2) applied
at each step then gives that the Hankel singular values of G^ k (s) will be 1 ; 2 ; : : : ; k
with multiplicities r1 ; r2 ; : : : ; rk , respectively. Hence Lemma 8.10 (1) gives that G^ k (s) ;
G^ k;1 (s) = k Ek (s) for some stable, all-pass Ek (s). Note also that Theorem 8.4, part
(3d), relation (i) also ensures that each G^ k (s) will have McMillan degree r1 + r2 + : : : + rk .
Finally taking D0 = G^ 0 (s) which will be a constant and combining the steps gives the
result.
2
Note that the construction
of Theorem
8.11 immediately gives an approximation
^
algorithm that will satisfy
G(s) ; G(s)
k+1 + k+2 + : : : + N . This will not be
1
an optimal Hankel norm approximation in general, but would involve less computation
since the decomposition into G^ (s) = F (s) need not be done, and at each step a balanced
realization of G^ k (s) is given by (A^k ; B^k ; C^k ) with a diagonal scaling.
An upper-bound on the L1 norm of G(s) is now obtained as an immediate consequence of Theorem 8.11.
Corollary 8.12 Let G(s) be a stable, rational p m, transfer function with Hankel
singular values 1 > 2 > : : : > N , where each i has multiplicity ri , and such that
G(1) = 0. Then
(1) kG(s)k1 2(1 + 2 + : : : + N )
(2) there exists a constant D0 such that
kG(s) ; D0 k1 1 + 2 + : : : + N :
Proof. For p = m consider the representation of G(s) given by Theorem 8.11 then
kG(s) ; D0 k1 = k1 E1 (s) + 2 E2 (s) + : : : + N EN (s)k1
1 + 2 + : : : + N
since Ek (s) are all-pass. Further setting s = 1, since G(1) = 0, gives
kD0 k 1 + 2 + : : : + N
) kG(s)k1 2(1 + 2 + : : : + N ):
For the case p =
6 m just augment G(s) by zero rows or columns to make it square, but
will have the same L1 norm, then the above argument gives upper bounds on the L1
norm of this augmented system.
Theorem 8.13 Given a stable, rational, m m, transfer function G(s) with Hankel
singular values 1 2 : : : k > k+1 = k+2 : : : = k+r > k+r+1 : : : n > 0
191
and let G^ (s) 2 RH1 of McMillan degree k be an optimal Hankel norm approximation
to G(s) obtained in Theorem 8.6, then there exists a D0 such that
n;X
k;r
i=1
i+k+r (G):
Proof. The theorem follows from Corollary 8.9 and Corollary 8.12.
kF ; D0 k1
since by Corollary 8.9,
n;X
k ;r
i=1
i (F (;s))
n;X
k;r
i=1
i+k+r (G):
for i = 1; 2; :::; n ; k ; r.
D0 can of course be obtained using any standard convex optimization algorithm:
D0 = arg min
G ; G^ ; D0
:
D0
Note that Theorem 8.13 can also be applied to non-square systems. For example,
G
consider a stable, rational, p m (p < m), transfer function G(s). Let Ga = 0
^
be an augmented square transfer function and let G^ a = G
G^ 2 be the optimal Hankel
norm approximation of Ga such that
n;X
k ;r
i=1
i+k+r (Ga )
192
0
with Da = D
D . Then
2
n;X
k;r
i=1
i+k+r (G)
n;X
k ;r
i=1
i+k+r (G))
11 A12 ; B = B1 ; C = C C
A= A
1
2
A21 A22
B2
0
= 01 I
^ B;
^ C;
^ D^ ) be dened by equations (8.27) to (8.32) (where 2 = 1 ) and dene
Let (A;
Gb :=
Gh :=
then
A11 B1
C1 0
A^ B^
C^ D^
193
Proof. (1) In order to prove that (Gb (s) ; Gh(s))= is all-pass we note that
"
A~
Gb (s) ; Gh (s) = ~
C
where
B~
D~
^
A~ = A011 A0^ ; B~ = BB^1 ; C~ = C1 ;C^ ; D~ = ;D:
Now Theorem 8.4, part (1) gives that the solutions to Lyapunov equations
A~P~ + P~ A~ + B~ B~ = 0
A~ Q~ + Q~ A~ + C~ C~ = 0
are
I
;
;
1
1
~
~
P = I ;;1 ; Q = ;; ;
(8.72)
(8.73)
(8:74)
(This is veried by noting that the blocks of equations (8.72) and (8.73) are also blocks
of equations (8.34) and (8.35) for Pe and Qe .) Hence P~ Q~ = 2 I and by Theorem 8.3
there exists D~ such that (Gb (s) ; Gh (s))= is all-pass. That D~ = ;D^ is an appropriate
choice is veried from equations (8.54) to (8.56) and Theorem 8.3, part (2).
(2) (G(s) ; Gb (s))= = (G(s) ; Gh (s))= + (Gb (s) ; Gh (s))= but the rst term on
the right hand side is all-pass by Theorem 8.4, part (2) and the second term is all-pass
by part (1) above. Hence kG(s) ; Gb (s)k1 2.
(3) Similarly using the fact that all-pass functions have unity Hankel norms gives
that
kG(s) ; Gb (s)kH kG(s) ; Gh (s)kH + kGb (s) ; Gh (s)kH = 2
(Note that Gh (s) is stable if 1 > I .)
2
Given the results of Lemma 8.15 bounds on the error in a truncated balanced realization are easily proved as follows.
Theorem 8.16 Let G(s) be a stable, rational, p m, transfer function with Hankel
singular values 1 > 2 : : : > N , where each i has multiplicity ri and let G~ k (s) be
obtained by truncating the balanced realization of G(s) to the rst (r1 + r2 + : : : + rk )
states.
Then
(1)
G(s) ; G~ k (s)
2(k+1 + k+2 + : : : + N ).
(2) G(s) ; G~ k (s
1
H
2(k+1 + k+2 + : : : + N ).
194
still give a balanced realization and the same argument is valid. Hence assume p = m.
Notice that since truncation of balanced realization are also balanced, satisfying the
truncated Lyapunov equations, the Hankel singular values of G~ i (s) will be 1 ; 2 ; : : : ; i
~
with multiplicities r1 ; r2 ; : : : ; ri , respectively. Also
obtained by truncating
Gi (s) can be
~
~
~
the balanced realization of Gi+1 (s) and hence
Gi+1 (s) ; Gi (s)
2i+1 for both L1
and Hankel norms. Hence (GN (s) := G(s))
G(s) ; G~ k (s
;1
N
X
=
( ~ i+1 (
i=k
))
s) ; G~ i (s
2(k+1 + k+2 + : : : + N )
2
In this section, we consider another operator. Again let G(s) 2 L1 . Then a Toeplitz
operator associated with G is denoted by TG and is dened as
TG : H2 7;! H2
TGf := (P+ MG)f = P+ (Gf ); for f 2 H2
i.e., TG = P+ MG jH2 . In particular if G(s) 2 H1 then TG = MGjH2 .
MG
H2
TG
>
L2
P+
?
- H2
Analogous to the Hankel operator, there is also a corresponding time domain description for the Toeplitz operator:
g(t ; )f ( )d; t 0
195
It is seen that the multiplication operator (in frequency domain) or the convolution
operator (in time domain) plays an important role in the development of the Hankel
and Toeplitz operators. In fact, a multiplication operator can be decomposed as several
Hankel and Toeplitz operators: let L2 = H2? H2 and G 2 L1 . Then the multiplication
operator associated with G can be written as
:= f 2 C : jj < 1g ;
:= f 2 C : jj 1g
@ D := f 2 C : jj = 1g :
Let L2 (@ D ) denote the Hilbert space of matrix valued functions dened on the unit
circle @ D as
Z 2
L2 (@ D ) = F () : 21
Trace F (ej )F (ej ) d < 1
0
with inner product
Z 2
1
Trace F (ej )G(ej ) d:
hF; Gi := 2
0
196
H2 (@ D ) = F () 2 L2 (@ D ) : 21
2
and let H2? (@ D ) be the (closed) subspace of L2 (@ D ) with matrix functions F () analytic
in C nD .
It can be shown that the Fourier transform (or bilateral Z transform) gives the
following isometric isomorphism:
= l2 (;1; 1)
= l2 [0; 1)
= l2 (;1; 0):
L2 (@ D )
H2 (@ D )
H2? (@ D )
Remark 8.3 It should be kept in mind that, in contrast to the variable z in the standard Z -transform, here denotes = z ;1 .
~
Analogous to the space in the half plane, L1 (@ D ) is used to denote the Banach
space of essentially bounded matrix functions with norm
H1 (@ D ) = F () 2 L1 (@ D ) :
The H1 norm is dened as
2
2[0;2]
Example 8.4 Let F () 2 L2 (@ D ), and let F () have a power series representation as
follows:
F () =
1
X
i=;1
Fi i :
Then F () 2 H2 (@ D ) if and only if Fi = 0 for all i < 0 and F () 2 H2? (@ D ) if and only
if Fi = 0 for all i 0.
3
197
y() = Gd ()u() =
1
X
1
X
i=;1 j =;1
Gi uj i+j =:
1
X
i=;1
yi i
y2
y1
y0
y;1
y;2
y;3
7 6
7 6
7 6
7 6
7 6
7 6
7 6
7=6
7 6
7 6
7 6
7 6
7 6
7 6
5 4
G0
G;1
G;2
G;3
G;4
G;5
G1
G0
G;1
G;2
G;3
G;4
G2
G1
G0
G;1
G;2
G;3
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
y0
y1
y2
y;1
y;2
y;3
7
7
7
7
7
7
7
7
7
7
7
7
7
7
5
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
G3
G2
G1
G0
G;1
G;2
G4
G3
G2
G1
G0
G;1
G0 G;1 G;2 G1
G1 G0 G;1 G2
G2 G1 G0 G3
G5
G4
G3
G2
G1
G0
32
76
76
76
76
76
76
76
76
76
76
76
76
76
76
54
G2 G3
G3 G4
G4 G5
G0 G1
G;1 G0
G;2 G;1
G2
G1
G0
7
7
u2 77
u1 77
u0 77
u;1 77 : (8:75)
u;2 77
u;3 77
32
76
76
76
76
76
76
76
76
76
76
76
76
76
76
54
u0
u1
u2
u;1
u;2
u;3
3
7
7
7
7
7
7
7
7
7
7
7
7
7
7
5
198
2
=:
T1 H1
H2 T2
u0
u1
u2
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
3
7
7
7
7
7
7
7
7
7
7
7
7
7
7
5
u;1 :
u;2
u;3
Matrices like T1 and T2 are called (block) Toeplitz matrices and matrices like H1 and
H2 are called (block) Hankel matrices. In fact, H1 is the matrix representation of the
Hankel operator ;Gd and T1 is the matrix representation of the Toeplitz operator TGd ,
and so on. Thus these operator norms can be computed from the matrix norms of their
corresponding matrix representations.
Lemma 8.17 kGd()k1 =
HT12 HT21
,
Example 8.5 Let Gd() 2 RH1(@ D ) and Gd() = C (;1 I ; A);1 B + D be a state
space realization of Gd () and let A have all the eigenvalues in D . Then
Gd () = D +
1
X
i=0
i+1 CAi B =
1
X
i=0
Gi i
P
i;1
Then for u 2 l2 (;1; 0), we have x0 = 1
i=1 A Bu;i , which denes the controllability
operator
2
3
x0 = cu = B AB A2 B
6
6
6
4
u;1
u;2
u;3
7
7
7
5
2 Rn :
..
.
n
On the other hand, given x0 and uk = 0; i 0, x 2 R , the output can be computed as
2
y0
6 y1
ox0 = 664 y2
..
.
C
7 6 CA 7
7 6
7
7 = 6 CA2 7 x0 2 l2 [0; 1)
5 4
5
..
.
199
which denes the observability operator. Of course, the adjoint operators of c and o
can be computed easily as
2
B
6 B A 7
c x0 = 664 B (A )2 775 x0 2 l2 (;1; 0)
..
.
and
2
o y = C A C (A )2 C
6
6
6
4
y0
y1
y2
3
7
7
7
5
2 Rn :
..
.
Hence, the Hankel operator has the following matrix representation:
2
C
6 CA
H = 664 CA2
..
.
7
7
7
5
G1 G2 G3
6 G2 G3 G4 7
B AB A2 B = 664 G3 G4 G5 775 :
..
.
..
.
..
.
..
.
(s) = s +21 :
The relation between a point j! on the imaginary axis and the corresponding point ej
on the unit circle is, from (8.76),
; 1:
ej = j!
j! + 1
200
This yields
d = ; !2 2+ 1 d! = ;j(j!)j2 d!;
?
H2? (j R)
;G(s)
?
- H2 (j R )
2
The above isomorphism between H2 (@ D ) and H2 (j R) has also the implication that
every discrete time H2 control problem can be converted to an equivalent continuous
time H2 control problem. It should be emphasized that the bilinear transformation is
not an isomorphism between H2 (@ D ) and H2 (j R).
In the following, we shall establish that the distance in L1 from G to the nearest
; equals k;Gk. This is the classical Nehari Theorem. Some applications
matrix in H1
201
and explicit construction of the approximation for the matrix case will be considered in
the later chapters.
Q2H;1
Remark 8.4 Note that from the mapping (8.76) and Lemma 8.18, the above theorem
is the same whether the problem is on the unit disk or on the half plane. Hence, we will
only give the proof on the unit disk.
~
Proof. We shall give the proof on the unit disk. First, it is easy to establish that the
; (@ D ), we have
Hankel norm is the lower bound: for xed Q 2 H1
k(G ; Q)f k2
kG ; Qk1 = sup
kf k2
?
f 2H2 (@ D )
kP+ (G ; Q)f k2
sup
kf k2
?
f 2H2 (@ D )
kP+ Gf k2
=
sup
?
f 2H2 (@ D ) kf k2
= k;G k :
; (@ D ) such that
Next we shall construct a Q() 2 H1
kG ; Qk1 = k;G k :
Let the power series expansion of G() and Q() be
G() =
Q() =
1
X
i=;1
0
X
i=;1
i Gi
i Qi :
(8:78)
202
From the discussion of the last section, there is a matrix representation of this operator:
2
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
G3
G2
G1
G4
G3
G2
G1
G0 ; Q0
G;1 ; Q;1 G0 ; Q0
G;2 ; Q;2 G;1 ; Q;1
G5
G4
G3
G2
G1
G0 ; Q0
77
77
77
77 :
77
77
77
5
(8:79)
G3
G2
G1
G4
G3
G2
G1
G0 ; Q0
G5
G4
G3
G2
3
7
7
7
7
7
7
5
By Parrott's Theorem (i.e., matrix dilation theory presented in Chapter 2), the minimum equals the norm of the following matrix
2
6
6
6
6
4
G3 G4 G5
G2 G3 G4
G1 G2 G3
3
7
7
7
7
5
which is the rotated Hankel matrix H1 . Hence, the minimum equals the Hankel norm
k;Gk. Next, choose Q;1 to minimize
2
6
6
6
6
6
6
6
6
4
G3
G2
G1
G0 ; Q0
G;1 ; Q;1
G4
G5
G3
G4
G2
G3
G1
G2
G0 ; Q0 G1
3
7
7
7
7
7
7
7
7
5
Again, the minimum equals k;G k. Continuing in this way gives a suitable Q.
As we have mentioned earlier we might be interested in approximating an L1 function by an H1 function. Then we have the following corollary.
203
Let G 2 RH1 and g be the impulse response of G and let 1 be the largest Hankel
singular value of G, i.e., k;G k = 1 . Suppose u 2 L2 (;1; 0] (or l2 (;1; 0)) and
v 2 L2 [0; 1) (or l2 [0; 1)) are the corresponding Schmidt pairs:
;g u = 1 v
;g v = 1 u:
Now denote the Laplace transform (or Z -transform) of u and v as U 2 RH?2 and
V 2 RH2 .
Lemma 8.21 Let G 2 RH1 and 1 be the largest Hankel singular value of G. Then
inf; kG ; Qk1 = 1
Q2H1
and
(G ; Q)U = 1 V:
Moreover, if G is a scalar function, then Q = G ; 1 V=U is the unique solution and
G ; Q is all-pass.
Proof. Let H := (G ; Q)U and note that ;GU 2 RH2 and P+H = P+(GU ) = ;GU .
Then
kH ; ;GU k22
kH k22 + k;G U k22 ; hH; ;G U i ; h;G U; H i
kH k22 + k;G U k22 ; hP+ H; ;G U i ; h;G U; P+ H i
kH k22 ; h;G U; ;GU i
kH k22 ; hU; ;G ;GU i
kH k22 ; 12 hU; U i
kH k22 ; 12 kU k22
kG ; Qk21 kU k22 ; 12 kU k22
0
=
=
=
=
=
=
= 0:
204
G(s) = CA B0 :
ALc + LcA + BB = 0
A Lo + Lo A + C C = 0:
And let 12 , be the largest eigenvalue and the corresponding eigenvector of Lc Lo:
Lc Lo = 12 :
Dene
Then
:= 1 Lo:
1
u = c = B e;A 2 L2 (;1; 0]
v = o = CeAt 2 L2 [0; 1)
U (s) =
V (s) =
;A 2 RH?
2
;B 0
A
C 0 2 RH2 :
It is easy to show that if G is a scalar, then U; V are scalar transfer functions and
U U = V V . Hence, V=U is an all-pass. The details are left as an exercise for the
reader.
A B
C 0 :
205
B
6 B A 7
u = c = 664 B (A )2 775 2 l2 (;1; 0)
..
.
C
6 CA 7
v = o = 664 CA2 775 2 l2 [0; 1)
..
.
1
X
A B 2 RH?
U () =
2
0
i=1
1
X
A
i
i
V () =
CA = CA C 2 RH2
i=0
B (A )i;1 ;i =
G() =
1
X
i=1
Gi i ; Gi 2 Rpq
G1 G2 G3
6 G2 G3 G4 7
H = 664 G3 G4 G5 775 :
..
.
..
.
..
.
..
.
206
H = U V
= diag
f1 ; 2 ; g
U = u1 u2
V = v1 v2
with U U = UU = I and V V = V V = I . Then
2
3
2
u11
v11
6 u12 7
6 v12
u = u1 = 4 . 5 ; v = v1 = 4 .
..
..
3
7
5
where u and v are partitioned such that u1i 2 Rp and v1i 2 Rq . Finally, U () and V ()
can be obtained as
U () =
V () =
1
X
i=1
1
X
i=1
In particular, if G() is an n-th order matrix polynomial, then matrix H has only a
nite number of nonzero elements and
H
0
n
H= 0 0
with
Hn
6
6
= 666
4
G1 G2 Gn;1 Gn
G2 G3 Gn 0 77
G3 G4 0
0 77 :
. .
.
. 7
..
..
..
.. 5
Gn 0 0
0
Hence, u and v can be obtained from a standard SVD of Hn .
208
modeling problem is much deeper { the universe of mathematical models from which a
model set is chosen is distinct from the universe of physical systems. Therefore, a model
set which includes the true physical plant can never be constructed. It is necessary for
the engineer to make a leap of faith regarding the applicability of a particular design
based on a mathematical model. To be practical, a design technique must help make
this leap small by accounting for the inevitable inadequacy of models. A good model
should be simple enough to facilitate design, yet complex enough to give the engineer
condence that designs based on the model will work on the true plant.
The term uncertainty refers to the dierences or errors between models and reality,
and whatever mechanism is used to express these errors will be called a representation
of uncertainty. Representations of uncertainty vary primarily in terms of the amount
of structure they contain. This re
ects both our knowledge of the physical mechanisms
which cause dierences between the model and the plant and our ability to represent
these mechanisms in a way that facilitates convenient manipulation. For example, consider the problem of bounding the magnitude of the eect of some uncertainty on the
output of a nominally xed linear system. A useful measure of uncertainty in this context is to provide a bound on the spectrum of the output's deviation from its nominal
response. In the simplest case, this spectrum is assumed to be independent of the input.
This is equivalent to assuming that the uncertainty is generated by an additive noise
signal with a bounded spectrum; the uncertainty is represented as additive noise. Of
course, no physical system is linear with additive noise, but some aspects of physical
behavior are approximated quite well using this model. This type of uncertainty received a great deal of attention in the literature during the 1960's and 1970's, and the
attention is probably due more to the elegant theoretical solutions that are yielded (e.g.,
white noise propagation in linear systems, Wiener and Kalman ltering, LQG) than to
the great practical signicance oered.
Generally, the deviation's spectrum of the true output from the nominal will depend
signicantly on the input. For example, an additive noise model is entirely inappropriate for capturing uncertainty arising from variations in the material properties of
physical plants. The actual construction of model sets for more general uncertainty can
be quite dicult. For example, a set membership statement for the parameters of an
otherwise known FDLTI model is a highly-structured representation of uncertainty. It
typically arises from the use of linear incremental models at various operating points,
e.g., aerodynamic coecients in
ight control vary with
ight environment and aircraft
congurations, and equation coecients in power plant control vary with aging, slag
buildup, coal composition, etc. In each case, the amounts of variation and any known
relationships between parameters can be expressed by conning the parameters to appropriately dened subsets of parameter space. However, for certain classes of signals
(e.g., high frequency), the parameterized FDLTI model fails to describe the plant because the plant will always have dynamics which are not represented in the xed order
model.
In general, we are forced to use not just a single parameterized model but model sets
that allow for plant dynamics which are not explicitly represented in the model structure.
209
P(j )
w(j )
210
log
actual model
211
modeled, but because our design techniques cannot eectively deal with the nonlinearity, it is treated as a conic linearity1 . As another example, we may deliberately choose
to ignore various known dynamic characteristics in order to achieve a simple nominal
design model. This is the model reduction process discussed in the previous chapters.
The following terminologies are used in this book.
Denition 9.1 Given the description of an uncertainty model set and a set of performance objectives, suppose P 2 is the nominal design model and K is the resulting
controller. Then the closed-loop feedback system is said to have
Robust Performance (RP): if the performance objectives are satised for every
plant belong to .
The nominal stability and performance can be easily checked using various standard
techniques. The conditions for which the robust stability and robust performance are
satised under various assumptions on the uncertainty set will be considered in the
following sections.
w1
- e e1
+ 6
+
-
M
++ w2
e2 e?
212
Theorem 9.1 (Small Gain Theorem) Suppose M 2 RH1 . Then the interconnected system shown in Figure 9.3 is well-posed and internally stable for all (s) 2
RH1 with
(a) kk1 1=
if and only if kM (s)k1 <
;
(b) kk1 < 1=
if and only if kM (s)k1
.
Proof. We shall only prove part (a). The proof for part (b) is similar. Without loss
of generality, assume
= 1.
(Suciency) It is clear that M (s)(s) is stable since both M (s) and (s) are stable.
Thus by Theorem 5.7 (or Corollary 5.6) the closed-loop system is stable if det(I ; M )
has no zero in the closed right-half plane for all 2 RH1 and kk1 1. Equivalently,
the closed-loop system is stable if
inf (I ; M (s)(s)) 6= 0
s2C +
s2C +
(j!0 ) = 64
1
2
...
7
5
1
1 1
213
u1 = 664
u11 ej1
u12 ej2
..
.
u1p ejp
3
7
7
7
5
v1 = 664
v11 ej1
v12 ej2
..
.
3
7
7
7
5
v1q ejq
where u1i 2 R and v1j 2 R are chosen so that i ; j 2 [;; 0) for all i; j .
Choose i 0 and j 0 so that
j!0 = ; \ j ; j!0 =
\ i ;
i
j
j + j!0
i + j!0
for i = 1; 2; : : :; p and j = 1; 2; : : :; q. Let
2
v11 11 ;+ss 3 h
i
..
7
u11 11 ;+ss u1p pp ;+ss 2 RH1 :
(s) = 1 64
5
.
1
v1q qq ;+ss
Then kk1 = 1=1 1 and (j!0 ) = 11 v1 u1 .
The theorem still holds even if and M are innite dimensional. This is summarized
as the following corollary.
Corollary 9.2 The following statements are equivalent:
(i) The system is well-posed and internally stable for all 2 H1 with kk1 < 1=
;
(ii) The system is well-posed and internally stable for all 2 RH1 with kk1 < 1=
;
(iii) The system is well-posed and internally stable for all 2 C qp with kk < 1=
;
(iv) kM k1
.
Remark 9.1 It can be shown that the small gain condition is sucient to guarantee
internal stability even if is a nonlinear and time varying \stable" operator with an
appropriately dened stability notion, see Desoer and Vidyasagar [1975].
~
The following lemma shows that if kM k1 >
, there exists a destabilizing with
kk1 < 1=
such that the closed-loop system has poles in the open right half plane.
(This is stronger than what is given in the proof of Theorem 9.1.)
Lemma 9.3 Suppose M 2 RH1 and kM k1 >
. Then there exists a 0 > 0 such
that for any given 2 [0; 0 ] there exists a 2 RH1 with kk1 < 1=
such that
det(I ; M (s)(s)) has a zero on the axis Re(s) = .
214
V = v1 v2 vq
3
2
1
7
2
= 64
5:
Write u1 and v1 in the following form:
2
6
u1 = 664
u11 ej1
u12 ej2
..
.
...
3
7
7
7
5
v1 = 664
v11 ej1
v12 ej2
..
.
3
7
7
7
5
v1q ejq
where u1i 2 R and v1j 2 R are chosen so that i ; j 2 [;; 0) for all i; j .
Choose i 0 and j 0 so that
;
;
j!
;
;
j!
j
0
i
0
\ + + j! = i ; \ + + j! = j
i
0
j
0
for i = 1; 2; : : : ; p and j = 1; 2; : : : ; q. Let
2
v11 11 +;ss 3 h
i
..
7
u11 11 +;ss u1p pp;+ss 2 RH1 :
(s) = 1 64
5
.
1
v1q qq ;+ss
u1p ejp
Then kk1 = 1=1 < 1 and det (I ; M ( + j!0 )( + j!0 )) = 0. Hence s = + j!0
is a zero for the transfer function det (I ; M (s)(s)).
2
The above lemma plays a key role in the necessity proofs of many robust stability
tests in the sequel.
215
be assumed to be stable. (Most of the robust stability tests discussed in the sequel can
be easily generalized to unstable case with some mild assumptions on the number
of unstable poles of the uncertain model, we encourage readers to ll in the details.)
In addition, we assume that the modeling error is suitably scaled with weighting
functions W1 and W2 , i.e., the uncertainty can be represented as W1 W2 .
We shall consider the standard setup shown in Figure 9.4, where is the set of uncertain plants with P 2 as the nominal plant and with K as the internally stabilizing
controller for P . The sensitivity and complementary sensitivity matrix functions are
dened as usual as
So = (I + PK );1 ; To = I ; So
and
Si = (I + KP );1 ; Ti = I ; Si :
Recall that the closed-loop system is well-posed and internally stable if and only if
I K ;1 = (I + K );1 ;K (I + K );1 2 RH :
1
; I
(I + K );1 (I + K );1
-f
6
;
- K
-
K ;1 =
(I + KSo W1 W2 );1 Si
;KSo(I + W1 W2 KSo );1
;
1
; I
(I + So W1 W2 K ) So (P + W1 W2 ) So (I + W1 W2 KSo);1
is well-posed and internally stable if (I + W2 KSo W1 );1 2 RH1 since
det(I + KSo W1 W2 ) = det(I + W1 W2 KSo) = (I + SoW1 W2 K )
= det(I + W2 KSoW1 ):
216
But (I + W2 KSo W1 );1 2 RH1 is guaranteed if kW2 KSoW1 k1 < 1. Hence
kW2 KSoW1 k1 1 is sucient for robust stability.
To show the necessity, note that robust stability implies that
K (I + K );1 = KSo(I + W1 W2 KSo );1 2 RH1
for all admissible . This in turn implies that
W2 K (I + K );1 W1 = I ; (I + W2 KSoW1 );1 2 RH1
for all admissible . By small gain theorem, this is true for all 2 RH1 with
Similarly, it is easy to show that the closed-loop system is stable for all 2 RH1
with kk1 1 if and only if kW2 KSoW1 k1 < 1.
In this section, we assume that the system model is described by the following set of
multiplicative perturbation
P = (I + W1 W2 )P
with W1 ; W2 ; 2 RH1 . Consider the feedback system shown in the Figure 9.5.
dm
z
w
- W2 - - e?
- W1
r - e- K
;6
- P
d~
?
Wd
d
?
?
- e - e y- We
e-
Theorem 9.5 Let = f(I + W1 W2 )P : 2 RH1g and let K be a stabilizing controller for the nominal plant P . Then
(i) the closed-loop system is well-posed and internally stable for all 2 RH1 with
kk1 < 1 if and only if kW2 ToW1 k1 1.
(ii) the closed-loop system is well-posed and internally stable for all 2 RH1 with
kk1 1 if kW2 ToW1 k1 < 1.
217
(iii) the robust stability of the closed-loop system for all 2 RH1 with kk1 1
does not necessarily imply kW2 To W1 k1 < 1.
(iv) the closed-loop system is well-posed and internally stable for all 2 RH1 with
kk1 1 only if kW2 To W1 k1 1.
(v) In addition, assume that neither P nor K has poles on the imaginary axis. Then
the closed-loop system is well-posed and internally stable for all 2 RH1 with
kk1 1 if and only if kW2 ToW1 k1 < 1.
Proof. We shall rst prove that the condition in (i) is necessary for robust stability.
Suppose kW2 ToW1 k1 > 1. Then by Lemma 9.3, for any given suciently small > 0,
there is a 2 RH1 with kk1 < 1 such that (I + W2 ToW1 );1 has poles on the axis
Re(s) = . This implies
stabilizes P . We have
and
To = s +1 1 ; kTo k1 = 1
1
1
1 + s+1
218
(s) = ;1 +
1
X
i=1
an sn ; ai 2 R:
s)
We now claim that a1 0. Otherwise, d(
ds s=0 = a1 < 0, along with (0) = ;1,
implies kk1 > 1. Hence (s) = ;1 + a1 s + s2 g(s) for some g(s) and a1 0 and
(I + K );1 = (I + K );1 = K (I + K );1 = s +s 1
1+
1
s+1
= 1 + a1 + sg
1
As another example, consider a left coprime factor perturbed plant described in Figure 9.6.
; ~
- e
M
w
-?
-M~ ;1
e
- N~
219
z2
y
-
K (I + PK );1 M~ ;1
1:
I
1
(N~ + ~ N )U + (M~ + ~ M )V
;1
2 RH1 :
(9:3)
~ + MV
~ );1 2 RH1 . Hence (9.3) holds if and only if
Since K stabilizes P , (NU
~ + MV
~ );1
I + (~ N U + ~ M V )(NU
;1
2 RH1 :
By the small gain theorem, the above is true for all kk1 < 1 if and only if
K
U ~
;1 ~ ;1
~ ;1
V (NU + MV )
1 =
I (I + PK ) M
1 1:
Similarly, one can show that the closed-loop system is well-posed and internally
stable for all kk1 1 if and only if
K (I + PK );1 M~ ;1
< 1:
I
1
Table 9.1 summaries robust stability tests on the plant uncertainties under various
assumptions. All of the tests pertain to the standard setup shown in Figure 9.4, where
is the set of uncertain plants with P 2 as the nominal plant and with K as the
internally stabilizing controller of P .
220
Example 9.2 Suppose an uncertain system with changing numbers of right-half plane
poles is described by
P = s ;1 : 2 R; jj 1 :
1 2 P has no right-half
Then P1 = s;1 1 2 P has one right-half plane pole and P2 = s+1
plane pole. Nevertheless, the set of P can be covered by a set of feedback uncertain
plants:
with P = 1s .
221
(I + W1 W2 );1 P
Representative Types of
Uncertainty Characterized
output (sensor) errors
neglected HF dynamics
changing # of rhp zeros
input (actuators) errors
neglected HF dynamics
changing # of rhp zeros
LF parameter errors
changing # of rhp poles
P (I + W1 W2 );1
LF parameter errors
changing # of rhp poles
(I + W1 W2 )P
P (I + W1 W2 )
P + W1 W2
P (I + W1 W2 P );1
(M~ + ~ M );1 (N~ + ~ N )
P = M~ ;1 N~
= ~ N ~ M
(N + N )(M + M );1
;1
P = NM
N
=
M
kW2 ToW1 k1 1
kW2 Ti W1 k1 1
kW2 So W1 k1 1
kW2 Si W1 k1 1
kW2 KSoW1 k1 1
kW2 So PW1 k1 1
K
;
1
~
S
M
o
I
;1
M Si [K I ]
1
1
1
Table 9.1: Unstructured Robust Stability Tests (# stands for `the number'.)
222
This set of plants have changing numbers of right-half plane zeros since the plant has
no right-half plane zero when = 0 and has one right-half plane zero when = ;2.
The uncertain plant can be covered by a set of multiplicative perturbed plants:
A robust stability problem can be viewed as another nominal performance problem. For
example, the output multiplicative perturbed robust stability problem can be treated as
a sensor noise rejection problem shown in Figure 9.7 and vise versa. It is clear that the
system with output multiplicative uncertainty as shown in Figure 9.5 is robustly stable
for kk1 < 1 i the H1 norm of the transfer function from w to z , Tzw , is no greater
than 1. Since Tzw = Ten~ , hence kTzw k1 1 i supkn~ k2 1 kek2 = kW2 ToW1 k1 1.
e
;6
- K
- P
- W2
e
-
? W1 n~
Figure 9.7: Equivalence Between Robust Stability With Output Multiplicative Uncertainty and Nominal Performance With Sensor Noise Rejection
There is, in fact, a much more general result along this line of equivalence: any
robust stability problem (with open ball of uncertainty) can be regarded as an equivalent
performance problem. This will be considered in Chapter 11.
r-e - K
;6
223
d~
?
Wd
d
?
- P 2 - e y- We
e-
L1 -optimal control framework and the latter leads to H2 and H1 design frameworks,
respectively. In this section, we will focus primarily on the H2 and H1 performance
objectives with unstructured model uncertainty descriptions. The performance under
structured uncertainty will be considered in Chapter 11.
P 2
Ted~
2 =
sup
kk1 <1
1 +
ws "
=
ws "
wt
2
1 ; jwt j
2
224
Suppose the performance criterion is to keep the energy of error e as small as possible,
i.e.,
sup kek2
kd~k2 1
for some small . By scaling the error e (i.e., by properly selecting We ) we can assume
without loss of generality that = 1. Then the robust performance criterion in this case
can be described as requiring that the closed-loop system be robustly stable and that
Ted~ 1 1; 8P 2 :
(9:5)
More specically, an output multiplicatively perturbed system will be analyzed rst.
The analysis for other classes of models can be done analogously. The perturbed model
can be described as
kW2 ToW1 k1 1
and
The exact analysis for this robust performance problem is not trivial and will be given
in Chapter 11. However, some sucient conditions are relatively easy to obtain by
bounding these two inequalities, and they may shed some light on the nature of these
problems. It will be assumed throughout that the controller K internally stabilizes the
nominal plant P .
225
Theorem 9.7 Suppose P 2 f(I + W1 W2)P : 2 RH1 ; kk1 < 1g and K internally stabilizes P . Then the system robust performance is guaranteed if either one of
the following conditions is satised
(i) for each frequency !
(9:7)
(9:8)
Proof. It is obvious that both condition (9.7)
and
condition (9.8) guarantee that
kW2 To W1 k1 1. So it is sucient to show that
Ted~
1 1; 8 2 RH1 ; kk1 < 1.
Now for any frequency !, it is easy to see that
2 o
Hence condition (9.7) guarantees (Ted~) 1 for all 2 RH1 with kk1 < 1 at all
frequencies.
Similarly, suppose W1 and Wd are invertible; write
Wd )(W1;1 Wd ) :
(Ted~) 1(W;eS(oW
T W )()
2 o 1
Hence by condition (9.8), (Ted~) 1 is guaranteed for all 2 RH1 with kk1 < 1
at all frequencies.
2
Remark 9.2 It is not hard to show that either one of the conditions in the theorem is
also necessary for scalar valued systems.
226
- W2 - - W1
- K
6
;
e
e
?
Wd
- e?- e? - We
- P
Figure 9.9: Robust Performance with Unstructured Uncertainty vs. Robust Stability
with Structured Uncertainty
Remark 9.5 Note that if W1 and Wd are invertible, then Ted~ can also be written as
Ted~ = We So Wd I + (W1;1 Wd );1 W2 ToW1 (W1;1 Wd ) ;1 :
227
We now consider the system with skewed specications, i.e., the uncertainty and performance are not measured at the same location. For instance, the system performance
is still measured in terms of output sensitivity, but the uncertainty model is in input
multiplicative form:
:= fP (I + W1 W2) : 2 RH1; kk1 < 1g :
For systems described by this class of models, see Figure 9.8, the robust stability condition becomes
kW2 Ti W1 k1 1;
and the nominal performance condition becomes
kWe So Wd k1 1:
To consider the robust performance, let T~ed~ denote the transfer matrix from d~ to e.
Then
T~ed~ = We So (I + PW1 W2 KSo);1 Wd
= We So Wd I + (Wd;1 PW1 )(W2 Ti W1 )(Wd;1 PW1 );1 ;1 :
The last equality follows if W1 , Wd , and P are invertible and, if W2 is invertible, can
also be written as
T~ed~ = We So Wd (W1;1 Wd );1 I + (W1;1 PW1 )(W2 P ;1 W2;1 )(W2 ToW1 ) ;1 (W1;1 Wd ):
Then the following results follow easily.
Theorem 9.8 Suppose P 2 = fP (I + W1 W2 ) : 2 RH1 ; kk1 < 1g and K
internally stabilizes P . Assume that P; W1 ; W2 , and Wd are square and invertible. Then
the system robust performance is guaranteed if either one of the following conditions is
satised
(i) for each frequency !
(We So Wd ) + (Wd;1 PW1 ) (W2 Ti W1 ) 1;
(9:9)
(ii) for each frequency !
(W1;1 Wd ) (We So Wd ) + (W1;1 PW1 ) (W2 P ;1 W2;1 ) (W2 ToW1 ) 1: (9:10)
Remark 9.6 If the appropriate invertibility conditions are not satised, then an alter-
228
Remark 9.7 It is important to note that in this case, the robust stability condition is
Remark 9.8 It is also noted that the robust performance condition is related to the
condition number of the weighted nominal model. So in general if the weighted nominal
model is ill-conditioned at the range of critical frequencies, then the robust performance
condition may be far more restrictive than the robust stability condition and the nominal
performance condition together. For simplicity, assume W1 = I , Wd = I and W2 = wt I
where wt 2 RH1 is a scalar function. Further, P is assumed to be invertible. Then
robust performance condition (9.10) can be written as
229
The fact that the condition number appeared in the robust performance test for
skewed problems can be given another interpretation by considering two sets of plants
1 and 2 as shown below.
- e? - P
- w~t -
-
- P
- e? -
;0:2 0:1 1 0 1 3
6 ;0:05
0 0 0 0:7 77
6
1
s
(s + 1)(s + 0:07)
6
7
0 ;1 1 0 7 =
P (s) = 6 0
;
0
:
05
0
:
7(s + 1)(s + 0:13)
a
(
s
)
4
1
0 0 0 0 5
0
0 0 0
230
10
condition number
10
10
10 -3
10
-2
10
10
-1
10
10
10
frequency
-e
;6
- L(s)
Suppose that the above closed-loop feedback system with L(s) = L0 (s) is stable.
Then the system is said to have
Gain Margin kmin and kmax if the closed-loop system is stable for all L(s) = kL0(s) with
kmin < k < kmax but unstable for L(s) = kmax L0 (s) and for L(s) = kmin L0 (s)
where 0 kmin 1 and kmax 1.
Phase Margin min and max if the closed-loop system is stable for all L(s) = e;j L0 (s)
with min < < max but unstable for L(s) = e;jmax L0 (s) and for L(s) =
e;jmin L0(s) where ; min 0 and 0 max .
These margins can be easily read from the open-loop system Nyquist diagram as shown
in Figure 9.12 where kmax and kmin represent how much the loop gain can be increased
231
and decreased, respectively, without causing instability. Similarly max and min represent how much loop phase lag and lead can be tolerated without causing instability.
1/K min
1/K max
min
-1
-1
max
L(j)
L(j)
a;s ; a>1
P = as
;1
ab
a2
max
min
max
i.e., very large gain margin but arbitrarily small phase margin.
232
i.e., very large phase margin but arbitrarily small gain margin.
The open-loop frequency response of the system is shown in Figure 9.13 for a = 2 and
b = 1; b = 1:9 and b = 0:55, respectively.
Sometimes, gain margin and phase margin together may still not be enough to
indicate the true robustness of a system. For example, it is possible to construct a
(complicated) controller such that
k < 1 ; k > a; = ;; >
min
max
min
max
but the Nyquist plot is arbitrarily close to (;1; 0). Such a controller is complicated to
construct; however, the following controller should give the reader a good idea of its
construction:
+ 0:55 1:7s2 + 1:5s + 1 :
Kbad = 3s:3+s 3+:31 0s:55
s + 1 s2 + 1:5s + 1:7
The open-loop frequency response of the system with this controller is also shown in
Figure 9.13 by the dotted line. It is easy to see that the system has at least the same
gain margin and phase margin as the system with controller K = 1, but the Nyquist plot
is closer to (;1; 0). Therefore this system is less robust with respect to the simultaneous
gain and phase perturbations. The problem is that the gain margin and phase margin
do not give the correct indication of the system's robustness when the gain and phase
are varying simultaneously.
I1 !_ 1 ; !2
(I1 ; I3 ) = T1
I1 !_ 2 ; !1
(I3 ; I1 ) = T2 :
233
0.5
-0.5
-1
-1.5
-2
-2.5
-2
-1.5
-1
-0.5
0.5
u1 := T1 =I1 ; a := (1 ; I =I )
:
3 1
u2
T2 =I1
234
u = K1 r ; y
where
K1 = 1 +1 a2 a1 ;1a :
y
P
r~
6;
f
K1
235
and the sensitivity function and the complementary sensitivity function are given by
1
1
s
;
a
1
a
;
1
;
1
S = (I + P ) = s + 1 a s ; T = P (I + P ) = s + 1 ;a 1 :
It is noted that this controller design has the property of decoupling the loops. Furthermore, each single loop has the open-loop transfer function as
1
so each loop has phase margin max = ;min = 90o and gain margin kmin = 0; kmax =
1.
Suppose one loop transfer function is perturbed, as shown in Figure 9.16.
w
y1
6z
?
e
y2
u2
u1
e
6
;
6
;
1
z (s)
w(s) = ;T11 = ; s + 1 :
kT11 k1
22
236
y1
y2
6SoS
f
?
/
f
6
f
11
g11
12
g12
21
g21
22
?
f
g22
;
r~1
?
f
6
;
f
r~2
237
22
0.707
0.707
11
238
10
Linear Fractional
Transformation
This chapter introduces a new matrix function: linear fractional transformation (LFT).
We show that many interesting control problems can be formulated in an LFT framework and thus can be treated using the same technique.
11 M12
(p1 +p2 )(q1 +q2 ) ;
M= M
M21 M22 2 C
239
240
and let ` 2 C q2 p2 and u 2 C q1 p1 be two other complex matrices. Then we can
formally dene a lower LFT with respect to ` as the map
F`(M; ) :
C q2 p2
7! C p1 q1
with
Fu (M; ) :
with
C q1 p1
7! C p2 q2
z
1
M
y1
- `
w1
u1
z2
y2
- u
M
z1
y1
y2
z2
w1 ;
u1
u2
w2 ;
11 M12
= M wu 1 = M
M21 M22
1
u1 = ` y1
11 M12
= M wu2 = M
M21 M22
2
u = u y2 :
u2
w2
It is easy to verify that the mapping dened on the left diagram is equal to F`(M; ` )
and the mapping dened on the right diagram is equal to Fu (M; u ). So from the
above diagrams, F`(M; ` ) is a transformation obtained from closing the lower loop
on the left diagram; similarly Fu (M; u ) is a transformation obtained from closing the
upper loop on the right diagram. In most cases, we will use the general term LFT in
referring to both upper and lower LFTs and assume that the contents will distinguish
the situations. The reason for this is that one can use either of these notations to express
241
22 M21
a given object. Indeed, it is clear that Fu (N; ) = F` (M; ) with N = M
M12 M11 .
It is usually not crucial which expression is used; however, it is often the case that one
expression is more convenient than the other for a given problem. It should also be
clear to the reader that in writing F` (M; ) (or Fu (M; )) it is implied that has
compatible dimensions.
A useful interpretation of an LFT, e.g., F` (M; ), is that F` (M; ) has a nominal
mapping, M11 , and is perturbed by , while M12; M21 , and M22 re
ect a prior knowledge as to how the perturbation aects the nominal map, M11 . A similar interpretation
can be applied to Fu (M; ). This is why LFT is particularly useful in the study of
perturbations, which is the focus of the next chapter.
The physical meaning of an LFT in control science is obvious if we take M as a proper
transfer matrix. In that case, the LFTs dened above are simply the closed-loop transfer
matrices from w1 7! z1 and w2 7! z2 , respectively, i.e.,
Tzw1 = F` (M; ` );
Tzw2 = Fu (M; u )
where M may be the controlled plant and may be either the system model uncertainties or the controllers.
Denition 10.2 An LFT, F`(M; ), is said to be well dened (or well-posed) if (I ;
M22 ) is invertible.
Note that this denition is consistent with the well-posedness denition of the feedback system, which requires the corresponding transfer matrix be invertible in Rp (s). It
is clear that the study of an LFT that is not well-dened is meaningless, hence throughout the book, whenever an LFT is invoked, it will be assumed implicitly to be well
dened. It is also clear from the denition that, for any M , F` (M; 0) is well dened;
hence any function that is not well dened at the origin cannot be expressed as an LFT
in terms of its variables. For example, f () = 1= is not an LFT of .
In some literature, LFT is used to refer to the following matrix functions:
(A + BQ)(C + DQ);1
(C + QD);1 (A + QB )
or
with
;1 B ; AC ;1 D
C ;1 A
C ;1 :
M = AC
;
N
=
;
1
;
1
;
1
C
;C D
B ; DC A ;DC ;1
242
11 M12
Lemma 10.2 Let F`(M; Q) be a given LFT with M = M
M21 M22 , then
(a) if M21 is invertible,
F` (M; Q) = (C + QD);1 (A + QB )
with
A I A
N = 4 ;B 0 ;B 5 ;
D 0 D
= Q Q :
Q11 Q12
11 M12
M= M
M21 M22 ; Q = Q21 Q22
A B1 B2
; G = 4 C1 D11 D12 5 :
C2 D21 D22
243
22 M21
N = I0 I0 M I0 I0 = M
M12 M11
where the dimensions of identity matrices are compatible with the partitions of M
and N .
(ii) Suppose Fu (M; ) is square and well-dened and M22 is nonsingular. Then the
inverse of Fu (M; ) exists and is also an LFT with respect to :
;1
;1
N = M11 ;MM;121 MM22 M21 ;MM12;M1 22 :
22 21
22
3
5
= 01 0 :
2
= 01 0 :
2
(v) Consider the following interconnection structure where the dimensions of 1 are
compatible with A:
z
Fu (G; 1 )
- Fu (Q; 2 )
Then the mapping from w to z is given by
244
2
A + B2 Q22 L1 C2
B2 L2 Q21
B1 + B2 Q22 L1 D21
Q12 L1 C2
Q11 + Q12 L1 D22 Q21
Q12 L1 D21
N =4
C1 + D12 L2 Q22 C2
D12 L2Q21
D11 + D12 Q22 L1 D21
where L1 := (I ; D22 Q22 );1 , L2 := (I ; Q22 D22 );1 , and =
3
5
1 0 .
0 2
Property (v) shows that if the open-loop system parameters are LFTs of some variables, then the closed-loop system parameters are also LFTs of the same variables. This
property is particularly useful in perturbation analysis and in building the interconnection structure. Similar results can be stated for lower LFT. It is usually convenient to
interpret an LFT Fu (M; ) as a state space realization of a generalized system with
frequency structure . In fact, all the above properties can be reduced to the standard
transfer matrix operations if = 1s I .
The following proposition is concerned with the algebraic properties of LFT s in the
general control setup.
Lemma 10.4 Let P = PP1121 PP1222 and let K be rational transfer function matrices
and let G = F` (P; K ). Then
(a) G is proper if P and K are proper with det(I ; P22 K )(1) =
6 0.
(b) F`(P; K1 ) = F`(P; K2 ) implies that K1 = K2 if P12 and P21 have normal full
column and row rank in Rp (s), respectively.
(c) If P and G are proper, det P (1) 6= 0, det P ; G0 00 (1) 6= 0 and P12 and
P21 are square and invertible for almost all s, then K is proper and
K = Fu (P ;1 ; G):
Proof.
F`(P; K1 ) ; F`(P; K2 ) = P12 (I ; K2 P22 );1 (K1 ; K2)(I ; P22 K1 );1 P21 :
(c) it is sucient to show that the formula for K is well-posed and K thus obtained
is proper. Let Q = P ;1 , which will be proper since det P (1) 6= 0, and dene
245
I 0 P ;1 I0 G
= det P ;1 P ; G0 00
6= 0:
Remark 10.1 This lemma shows that under certain conditions, an LFT of transfer
matrices is a bijective map between two sets of proper and real rational matrices. When
given proper transfer matrices P and G with compatible dimensions which satisfy conditions in (c), there exists a unique proper K such that G = Fl (P; K ). On the other
hand, the conditions of part (c) show that the feedback systems are well-posed.
~
Remark 10.2 A simple interpretation of the result (c) is given by considering the
signals in the feedback systems,
z
w
P
y
u
- K
hence
w
u
z
y
= P wu ;
u = Ky
) z = F` (P; K )w = Gw;
= P ;1 yz ; z = Gw
) u = Fu (P ;1 ; G)y; or K = Fu (P ;1 ; G):
246
Polynomials
A very commonly encountered object in control and mathematics is a polynomial function. For example,
p() = a0 + a1 + + an n
with indeterminate . It is easy to verify that p() can be written in the following LFT
form:
p() = F`(M; In )
with
3
2
a0 a1 an;1 an
1 0 0 0 77
M = 0 1. 0. 0. 777 :
.
.. 5
0 .. . . ..
0 0 1 0
Hence every polynomial is a linear fraction of their indeterminates. More generally, any
multivariate (matrix) polynomials are also LFTs in their indeterminates; for example,
6
6
6
6
6
4
p(1 ; 2 ) = F` (N; )
2
6
6
6
6
4
a6 1 0 1 0
a4 0 a1 0 a3
3
7
7
N = 1 0 0 0 0 77 ;
= 1 I2 I :
2 2
a5 0 0 0 a2 5
1 0 0 0 0
It should be noted that these representations or realizations of polynomials are neither
unique nor necessarily minimal. Here a minimal realization refers to a realization with
the smallest possible dimension of . As commonly known, in multidimensional systems
and lter theory, it is usually very hard, if not impossible, to nd a minimal realization
for even a two variable polynomial. In fact, the minimal dimension of depends also on
the eld (real, complex, etc.) of the realization. More detailed discussion of this issue
is beyond the scope of this book, the interested readers should consult the references in
2-d or n-d systems or lter theory.
247
Rational Functions
M = 1 ;
;
:
;
N = 4 1 0 ;
5 :
1 0 ;
Although the latter can be reduced to the former, it is not easy to see how to carry out
such reduction for a complicated problem, even if it is possible.
We can use the LFT formulae to establish the relationship between transfer matrices
and their state space realizations. A system with a state space realization as
x_ = Ax + Bu
y = Cx + Du
has a transfer matrix of
A
B
;
1
G(s) = D + C (sI ; A) B = F (
; 1 I ):
u
C D
248
B ; ):
G(s) = Fu ( CA D
More generally, consider a discrete time 2-D (or MD) system realized by the rst-order
state space equation
;1
= z10 I z ;01I =: 10I 0I
2
2
B1
11 A12
A= A
A21 A22 ; B = B2 ; C = C1 C2
G(z1 ; z2 ) = D + C (
A
=: Fu ( C
z1 I
0
;1
;1
z2 I ; A) B = D + C (I ; A) B
B
D ; ):
-
A B
C D
The following notation for a transfer matrix has already been adopted in the previous
chapters:
A B := F ( A B ; ):
u C D
C D
249
It is easy to see that this notation can be adopted for general dynamical systems,
e.g., multidimensional systems, as far as the structure is specied. This notation
means that
the transfer
matrix can be expressed as an LFT of with the coecient
A
B
matrix C D . In this special case, we say the parameter matrix is the frequency
structure of the system state space realization. This notation is deliberately somewhat
ambiguous and can be viewed as both a transfer matrix and one of its realizations. The
ambiguity is benign and convenient and can always be resolved from the context.
Frequency Transformation
The bilinear transformation between the z -domain and s-domain
+1
s = zz ;
1
transforms the unit disk to the left-half plane and is the simplest example of an LFT.
We may rewrite it in our standard form as
1 I = I ; p2I z ;1I (I + z ;1I );1 p2I = F (N; z ;1I )
u
s
where
p
I
p
N = ; 2I ;2II :
B = F (M; 1 I )
G(s) = CA D
u
s
where
B ;
M = CA D
;1
A);1 B :
M~ = ;(pI 2;CA(I) ; (AI)+;1A) C;(I ;2(IA;
;
) 1B + D
The transformation from the z -domain to the s-domain can be obtained similarly.
250
u
f
i
;6
W1
y K
d
?
-i
- P
v-
- W2
? n
F
z
w
G
y
u
- K
with
w=
and
d
n
z=
W2 P
W1
G=4 0
;FP ;F ;FP
W2 P
v
uf
0
0
Using the properties of LFTs, we can show that constrained structure control synthesis
problems can be converted to constrained structure constant output feedback problems.
Consider the synthesis structure in the last example and assume
2
A B1 B2
G = 4 C1 D11 D12
C2 D21 D22
3
5
BK :
K = ACK D
K
K
F`(G; K ) = F` (M (s); F )
and
251
2
A 0 B1 0 B2
6 0 0
0 I 0
6
M (s) = 66 C1 0 D11 0 D12
4 0 I
0 0 0
C2 0 D21 0 D22
3
7
7
7
7
5
BK :
F = ACK D
K
K
Note that F is a constant matrix, not a system matrix. Hence if the controller structure
is xed (or constrained), then the corresponding control problem becomes a constant
(constrained) output feedback problem.
One natural type of uncertainty is unknown coecients in a state space model. To motivate this type of uncertainty description, we will begin with a familiar mass/spring/damper
system, shown below in Figure 10.1.
6
F
m
XX
k
XX
X
XX
F:
x + mc x_ + mk x = m
Suppose that the 3 physical parameters m; c, and k are not known exactly, but are
believed to lie in known intervals. In particular, the actual mass m is within 10% of
a nominal mass, m , the actual damping value c is within 20% of a nominal value of
c, and the spring stiness is within 30% of its nominal value of k. Now introducing
perturbations m , c, and k , which are assumed to be unknown but lie in the interval
[;1; 1], the block diagram for the dynamical system can be shown in Figure 10.2.
252
x_
x
1
m (1+0:1m )
6;
-d
6+
c(1 + 0:2c )
(1 + 0:3k )
k
; 0m:1 . Suppose that the input signals of the dynamical system are
1 ;0:1
1
m
with M1 =
selected as x1 = x; x2 = x;_ F , and the output signals are selected as x_ 1 and x_ 2 . To
represent the system model as an LFT of the natural uncertainty parameters m ; c and
k , we shall rst isolate the uncertainty parameters and denote the inputs and outputs
of k ; c and m as yk ; yc ; ym and uk ; uc ; um, respectively, as shown in Figure 10.3. Then
2
6
6
6
6
6
6
4
i.e.,
x_ 1
x_ 2
yk
yc
ym
7
7
7
7
7
7
5
6
6
6
6
6
6
4
; mk
; mc
1
m
0:3k 0 0
0 0:2c 0
;k ;c 1
"
; m1
0
0
;1
x_ 1
x_ 2
; m1
2
6
6
; 0m:1 777 66
0
0
0
0
;1 ;0:1
76
76
76
56
6
4
x1
6
= F` (M; ) 4 x2
F
x1
7
x2 77
F 77 ;
7
uk 77
uc 75
um
3
7
5
2
6
4
uk
uc
um
yk
7
6
=
5
4 yc
ym
3
7
5
2
6
6
6
6
6
6
4
; mk ; mc
M = 0:3k
253
0
0
0
;1
0
0
0
0
;1 ;0:1
m
0 0
0 0:2c 0
;k ;c 1
x1
x2
ym
7
7
7
7
7
7
5
F
6;
M1
um
- m
- c
- k
k 0 0
6
; = 4 0 c 0 75 :
0 0 m
- 0:2 - c
yc
-e -e
6 6
uc
- 0:3
-e
6
uk
yk k
We will consider a special class of state space models with unknown coecients and show
how this type of uncertainty can be represented via the LFT formulae with respect to
an uncertain parameter matrix so that the perturbations enter the system in a feedback
form. This type of modeling will form the basic building blocks for components with
parametric uncertainty.
Consider a linear system G (s) that is parameterized by k uncertain parameters,
1 ; : : : ; k , and has the realization
2
G (s)
6
6
= 66
4
A+
C+
k
X
i A^i B +
k
X
i=1
k
X
i=1
k
X
i=1
i=1
i C^i D +
i B^i
i D^i
3
7
7
7
7
5
254
Here A; A^i 2 Rnn ; B; B^i 2 Rnnu ; C; C^i 2 Rny n , and D; D^i 2 Rny nu .
The various terms in these state equations are interpreted as follows: the nominal
system description G(s), given by known matrices A; B; C; and D, is (A; B; C; D) and
the parametric uncertainty in the nominal system is re
ected by the k scalar uncertain
parameters 1 ; : : : ; k , and we can specify them, say by i 2 [;1; 1]. The structural
knowledge about the uncertainty is contained in the matrices A^i ; B^i ; C^i , and D^i . They
re
ect how the i'th uncertainty, i , aects the state space model.
Now, we consider the problem of describing the perturbed system via the LFT
formulae so that all the uncertainty can be represented as a nominal system with the
unknown parameters entering it as the feedback gains. This is shown in Figure 10.4.
Since G (s) = Fu (M ; 1s I ) where
2
6
6
6
M = 6
4
A+
C+
k
X
k
X
i=1
k
X
i=1
k
X
i A^i B +
i=1
i C^i D +
i=1
i B^i
i D^i
3
7
7
7
7
5
Ci Di
"
#"
Pi = Li
Wi
#
Ri
Zi
where Li 2 Rnqi , Wi 2 Rny qi , Ri 2 Rnqi and Zi 2 Rnu qi . Hence, we have
"
i Pi = Li
Wi
"
Ri
[i Iqi ]
Zi
#
M =
M11
}|
A B
C D
#{
z"
M12
}|
L1 Lk
W1 Wk
z
#{ 2
6
6
4
p
1 Iq1
}|
{ z2
3
...
76
76
54
k Iqk
M21
}|
R1 Z1
..
.
..
.
Rk Zk
{
3
7
7
5
255
"
A B B2
C D D12
C2 D21 D22
x_
y
1I
y2
1 I
...
u2
k I
B2 =
D12 =
C2 =
and
h
h
h
h
L1 L2 Lk
W1 W2 Wk
R1 R2 Rk
D21 = Z1 Z2 Zk
D22 = 0
"
i
i
i
256
where a; b; c; d and e are given constants or transfer functions. We would like to write
G as an LFT in terms of 1 and 2 . We shall do this in three steps:
1. Draw a block diagram for the input/output relation with each separated as
shown in Figure 10.7.
2. Mark the inputs and outputs of the 's as y's and u's, respectively. (This is
essentially pulling out the s).
3. Write z and y's in terms of w and u's with all 's taken out. (This step is equivalent
to computing the transformation in the shadowed box in Figure 10.6.)
2
6
6
6
6
6
6
4
y1
y2
y3
y4
z
u1
7
6
7
6 u2
7
6
7 = M 6 u3
7
6
7
6
5
4 u4
w
3
7
7
7
7
7
7
5
257
- 1
- 2
- 3
. . . . . . . . . . . . . . . . . . . .
.......... . . . . ....................................................................
.... .
....................... .. .. .. .. .... . .
.. . . . . .
.....
. .
. . . . . . . . . .................................... ? .........
. .
. . . .... ...... ............ ..... ... .......
.... ...... ............ ..... ... ... ....... . . . . .........
..... . .
.................................................................
.... .
........
...............................
.....
........
. .
. . . . ...
......... .. ... .. .......................................
.......
........................................ . . . . . .......
. . . . . . . . . . . . . . . . . . . .
- K
Figure 10.6: Pulling out the s
where
M=
Then
6
6
6
6
6
6
4
0
1
1
0
0
0
0
be
ae
0
0 0
0 0
bd + c 0
ad 1
"
z = Fu (M; )w; = 1 I2
0
1
0 777
0 77 :
b
a
7
5
2 I2
All LFT examples in the last section can be obtained following the above steps.
258
a
u4
y
z ?
e
2 4 e
6
c
6
b
u3
y u
y
2 3 1 1 1
?
;d
y2
1
w
6
u2
;e
-e
6
"
"
Fl (P; K11 )
P12 (I ; K11 P22 );1 K12 :
S (P; K ) :=
K21 (I ; P22 K11 );1 P21
Fu (K; P22 )
Note that this denition is dependent on the partitioning of the matrices P and K
above. In fact, this star product may be well dened for one partition and not well
dened for another; however, we will not explicitly show this dependence because it is
always clear from the context. In a block diagram, this dependence appears, as shown
in Figure 10.8.
Now suppose that P and K are transfer matrices with state space representations:
2
A B1 B2
P = 64 C1 D11 D12
C2 D21 D22
Then the transfer matrix
AK BK 1 BK 2
K = 64 CK 1 DK 11 DK 12 75 :
CK 2 DK 21 DK 22
7
5
"
"
z
S (P; K ) : w !
7
w^
z^
P
aa !!!
a
!
!! aaa
K
z^
259
z
z^
w
S (P; K ) w^
w^
A B1 B2
6
6
S (P; K ) = 4 C1 D 11 D 12
C2 D 21 D 22
where
A =
B =
C =
D =
"
"
"
"
3
7
7
5
"
A B
=
C D
A + B2 R~ ;1 DK 11 C2
B2 R~ ;1 CK 1
;
1
BK 1 R C2
AK + BK 1 R;1 D22 CK 1
#
B1 + B2 R~ ;1 DK 11 D21
B2 R~ ;1 DK 12
BK 1 R;1 D21
BK 2 + BK 1 R;1 D22 DK 12
#
C1 + D12 DK 11 R;1 C2
D12 R~ ;1 CK 1
DK 21 R;1C2
CK 2 + DK 21 R;1 D22 CK 1
D11 + D12 DK 11 R;1D21
D12 R~ ;1 DK 12
DK 21 R;1 D21
DK 22 + DK 21 R;1 D22 DK 12
R = I ; D22 DK 11 ; R~ = I ; DK 11 D22 :
A = S
B = S
"
"
A
C2
B1
D21
# "
B2 ; DK 11
D22
BK 1
# "
B2 ; DK 11
D22
BK 1
#!
CK 1 ;
AK
#!
DK 12 ;
BK 2
260
C = S
D = S
"
"
C1
C2
D11
D21
# "
D12 ; DK 11
D22
DK 21
# "
D12 ; DK 11
D22
DK 21
#!
CK 1 ;
CK 2
#!
DK 12 :
DK 22
11
262
number of the transfer matrix between its original location and the re
ected location.
Thus it is highly desirable to treat the uncertainties as they are and where they are.
The structured singular value is dened exactly for that purpose.
z = Fu (F` (P; K ); ) w
= F` (Fu (P; ); K ) w:
We will focus our discussion in this section on analysis methods; therefore, the
controller may be viewed as just another system component and absorbed into the
interconnection structure. Denote
"
1
2
263
-
z
- K
Figure 11.1: General Framework
and then the general framework reduces to Figure 11.2, where
-
M
z
"
w := d1 ; z := e1
d2
e2
then the system is robustly stable for all (s) 2 RH1 with kk1 < 1 if and only if
Fu (M; ) 2 RH1 for all admissible , which is guaranteed by kM11 k1 1. (Note
that this is not necessarily equivalent to (I ; M11);1 2 RH1 if belongs to a closed
ball as shown in Theorem 9.5.)
The analysis results presented in the previous chapters together with the associated synthesis tools are summarized in Table 11.1 with various uncertainty modeling
assumptions.
However, the analysis is not so simple for systems with multiple sources of model
uncertainties, including the robust performance problem for systems with unstructured
264
d2
e
d1 e1
- e - K - e? 2- P
;6
e
3
- W2 3- dW1
- e?
- F (s)
..
.
- 1 (s)I
..
.
M11 (s)
..
.
Performance
Specications
Perturbation
Assumptions
265
Analysis
Tests
Synthesis
Methods
LQG
=0
w = U0 (t)
E (U0 U0 ) = I
E (kz k22 ) 1
kwk2 1
kz k2 1
=0
kwk2 1
Internal Stability
kk1 < 1
kM22 k2 1
Wiener-Hopf
H2
kM22 k1 1 Singular Value
Loop Shaping
kM11 k1 1
H1
266
Conceptually, the structured singular value is nothing but a straightforward generalization of the singular values for constant matrices. To be more specic, it is instructive
at this point to consider again the robust stability problem of the following standard
feedback interconnection with stable M (s) and (s).
w1
- e e1
+ 6
+
-
++ w2
e2 e?
M
One important question one might ask is how large (in the sense of kk1 ) can
be without destabilizing the feedback system. Since the closed-loop poles are given by
det(I ; M ) = 0, the feedback system becomes unstable if det(I ; M (s)(s)) = 0 for
some s 2 C + . Now let > 0 be a suciently small number such that the closed-loop
system is stable for all stable kk1 < . Next increase until max so that the closedloop system becomes unstable. So max is the robust stability margin. By small gain
theorem,
1 = kM k := sup (M (s)) = sup (M (j!))
1
max
s2C +
s2C +
267
s2C +
s2C +
s2C +
s2C +
Now suppose sups2C + (M (s)) > 1=, then by the denition of , there is an so 2
[ f1g and a complex structured such that () < and det(I ; M (so )) = 0.
This implies that there is a 0 !^ 1 and 0 < 1 such that det(I ; M (j !^ ) ) =
0. This in turn implies that (M (j !^ )) > 1= since ( ) < . In other words,
sups2C + (M (s)) sup! (M (j!)). The proof is complete.
2
C+
11.2.2 Denitions of
This section is devoted to dening the structured singular value, a matrix function
denoted by (). We consider matrices M 2 C nn . In the denition of (M ), there is an
underlying structure (a prescribed set of block diagonal matrices) on which everything
in the sequel depends. For each problem, this structure is, in general, dierent; it
depends on the uncertainty and performance objectives of the problem. Dening the
structure involves specifying three things: the type of each block, the total number of
blocks, and their dimensions.
There are two types of blocks: repeated scalar and full blocks. Two nonnegative
integers, S and F , represent the number of repeated scalar blocks and the number of
full blocks, respectively. To bookkeep their dimensions, we introduce positive integers
r1 ; : : : ; rS ; m1 ; : : : ; mF . The i'th repeated scalar block is ri ri , while the j 'th full block
is mj mj . With those integers given, we dene C nn as
= diag [1Ir1 ; : : : ; sIrS ; 1; : : : ; F ] : i 2 C ; j 2 C mj mj :
(11:3)
For consistency among all the dimensions, we must have
S
X
i=1
ri +
F
X
j =1
mj = n:
Often, we will need norm bounded subsets of , and we introduce the following notation:
B = f 2 : () 1g
(11:4)
Bo = f 2 : () < 1g
(11:5)
268
where the superscript \o" symbolizes the open ball. To keep the notation as simple as
possible in (11.3), we place all of the repeated scalar blocks rst; in actuality, they can
come in any order. Also, the full blocks do not have to be square, but restricting them
as such saves a great deal in terms of notation.
(11:6)
Remark 11.1 Without a loss in generality, the full blocks in the minimal norm can
each be chosen to be dyads (rank = 1). To see this, assume S = 0, i.e., all blocks are
full blocks. Suppose that I ; M is singular for some 2 . Then there is an x 2 C n
x1
x2
x = 66 ..
xF
3
7
7
7
7
5
; xi 2 C mi ; i = 1; : : : ; F
and dene
x~1
x~2
x~ = 66 ..
and
.
x~F
y1
y2
7
7
7
7
5
:= Dx = 66
6
6
4
d1 x1
d2 x2
..
.
dm xF
3
7
7
7
7
5
1 x~1
6
7
6
~2 77
6
7
6 2 x
y = 66 .. 77 := ~x = 66 .. 77 :
4 . 5
4
. 5
yF
~xF
It follows that kx~i k = 1 if xi 6= 0, kx~i k = 0 if xi = 0 , and y 6= 0. Hence, dene a new
perturbation ~ 2 C nn as
~ := diag [y1 x~1 ; : : : ; yF x~F ] :
269
~ () and y = ~
~ x. Note that D = D and D~ = ~ D, we have
Obviously, ()
M x = x =) MD;1 ~x = x =) MD;1 y = x
~ x = x =) M ~ x = x
=) MD;1 ~
i.e., I ; M ~ is also singular. Hence we have replaced a general perturbation which
satises the singularity condition with a perturbation ~ that is no larger (in the ()
sense) and has rank 1 for each blocks but still satises the singularity condition. ~
An alternative expression for (M ) follows from the denition.
Lemma 11.2 (M ) = max (M )
2B
Proof. The only 's in which satisfy the det (I ; M ) = 0 constraint are
reciprocals of nonzero eigenvalues of M . The smallest one of these is associated
with the largest (magnitude) eigenvalue, so, (M ) = (M ).
2
270
"
0
(1) M =
for any > 0. Then (M ) = 0 and (M ) = . But det(I ;M ) =
0 0
1 so (M ) = 0.
"
Thus neither nor provide useful bounds even in simple cases. The only time they
do provide reliable bounds is when .
However, the bounds can be rened by considering transformations on M that do
not aect (M ), but do aect and . To do this, dene the following two subsets
of C nn :
(
U = fU 2 : UU = In g
D =
(11:9)
(11:10)
(11:11)
(11:12)
(11:13)
same.
(11:14)
where the equality comes from Lemma 11.2. Note that the last element in the D matrix
is normalized to 1 since for any nonzero scalar
, DMD;1 = (
D) M (
D);1 .
271
Remark 11.2 Note that the scaling set D in Theorem 11.3 and in the inequality (11.14)
does not necessarily be restricted to Hermitian. In fact, they can be replaced by any
set of nonsingular matrices that satisfy (11.12). However, enlarging the set of scaling
matrices does not improve the upper bound in inequality (11.14). This can be shown
as follows: Let D be any nonsingular matrix such that D = D. Then there exist a
Hermitian matrix 0 < R = R 2 D and a unitary matrix U such that D = UR and
;
inf
DMD;1 = inf
URMR;1U = Rinf
RMR;1 :
D
D
2D
Therefore, there is no loss of generality in assuming D to be Hermitian.
11.2.3 Bounds
max
(UM ) (M ) Dinf
DMD;1 :
U 2U
2D
The lower bound is always an equality [Doyle, 1982].
F=
0
yes
no
272
S = 0, F = 1 : (M ) = (M ).
;
S = 1, F = 0 : (M ) = (M ) = Dinf2D DMD;1 . This is a standard result in
linear algebra. In fact, without a loss in generality, the matrix M can be assumed
in Jordan Canonical form. Now let
2
J1 =
6
6
6
6
6
6
6
4
1
1
... ...
1
7
7
7
7
7
7
7
5
6
6
6
= 666
6
4
; D1
...
kn1 ;2
kn1 ;1
7
7
7
7
7
7
7
5
2 C n1 n1 :
M PM ; P < 0
(Lyapunov asymptotic stability).
S = 0, F 4 : For this case, the upper bound is not always equal to . This
is important, as these are the cases that arise most frequently in applications.
Fortunately, the bound seems to be close to . The worst known example has a
ratio of over the bound of about :85, and most systems are close to 1.
273
The above bounds are much more than just computational schemes. They are also
theoretically rich and can unify a number of apparently quite dierent results in linear
systems theory. There are several connections with Lyapunov asymptotic stability, two
of which were hinted at above, but there are further connections between the upper
bound scalings and solutions to Lyapunov and Riccati equations. Indeed, many major
theorems in linear systems theory follow from the upper bounds and from some results
of Linear Fractional Transformations. The lower bound can be viewed as a natural
generalization of the maximum modulus theorem.
Of course one of the most important uses of the upper bound is as a computational
scheme when combined with the lower bound. For reliable use of the theory, it is
essential to have upper and lower bounds. Another important feature of the upper bound
is that it can be combined with H1 controller synthesis methods to yield an ad-hoc
-synthesis method. Note that the upper bound when applied to transfer functions is
simply a scaled H1 norm. This is exploited in the D ; K iteration procedure to perform
approximate -synthesis (Doyle[1982]), which will be brie
y introduced in section 11.4.
M = M11 M12
M21 M22
(11:15)
and suppose there are two dened block structures, 1 and 2 , which are compatible
in size with M11 and M22 , respectively. Dene a third structure as
=
("
1 0
0 2
: 1 2 1 ; 2 2 2 :
(11:16)
Now, we may compute with respect to three structures. The notations we use to keep
track of these computations are as follows: 1 () is with respect to 1 , 2 () is with
respect to 2 , and () is with respect to . In view of these notations, 1 (M11 ),
2 (M22 ) and (M ) all make sense, though, for instance, 1 (M ) does not.
This section is interested in following constant matrix problems:
determine whether the LFT Fl (M; 2 ) is well posed for all 2 2 2 with (2 )
(< ), and,
if so, then determine how \large" Fl (M; 2 ) can get for this norm-bounded set
of perturbations.
Let 2 2 2 . Recall that Fl (M; 2 ) is well posed if I ; M22 2 is invertible. The
rst theorem is nothing more than a restatement of the denition of .
Theorem 11.6 The linear fractional transformation Fl (M; 2) is well posed
274
(M ) < 1
(M ) 1
()
()
8
>
>
<
>
>
:
2 2B2
8
>
>
<
2 (M22 ) 1; and
>
>
:
Proof. We shall only prove the rst part of the equivalence. The proof for the second
part is similar.
( Let i 2 i be given, with (i ) 1, and dene = diag [1 ; 2 ]. Obviously
2 . Now
"
#
I
;
M
11 1 ;M12 2
det (I ; M ) = det
:
(11:17)
;M I ; M
21 1
22 2
275
Remark 11.3 This theorem forms the basis for all uses of in linear system robustness
analysis, whether from a state-space, frequency domain, or Lyapunov approach.
~
The role of the block structure 2 in the MAIN LOOP theorem is clear - it is
the structure that the perturbations come from; however, the role of the perturbation
structure 1 is often misunderstood. Note that 1 () appears on the right hand side
of the theorem, so that the set 1 denes what particular property of Fl (M; 2 ) is
considered. As an example, consider the theorem applied with the two simple block
structures considered right after Lemma 11.2. Dene 1 := f1 In : 1 2 C g . Hence, for
A 2 C nn ; 1 (A) = (A). Likewise, dene 2 = C mm ; then for D 2 C mm ; 2 (D) =
(D). Now, let be the diagonal augmentation of these two sets, namely
:=
("
0mn
2
Let A 2 C nn ; B 2 C nm ; C 2 C mn , and D 2 C mm be given, and interpret them as
the state space model of a discrete time system
M= A B :
C D
Applying the theorem with this data gives that the following are equivalent:
The spectral radius of A satises (A) < 1, and
1 2C
j1 j1
(11:19)
(11:20)
(11:21)
2 2C
(2 )1
276
The rst condition is recognized by two things: the system is stable, and the jj jj1
norm on the transfer function from u to y is less than 1 (by replacing 1 with 1z )
jzj1
j1 j1
The second condition implies that (I ; D2 );1 is well dened for all (2 ) 1 and
that a robust stability result holds for the uncertain dierence equation
The most well-known use of as a robustness analysis tool is in the frequency domain.
Suppose G(s) is a stable, multi-input, multi-output transfer function of a linear system.
For clarity, assume G has q1 inputs and p1 outputs. Let be a block structure, as in
equation (11.3), and assume that the dimensions are such that C q1 p1 . We want
to consider feedback perturbations to G which are themselves dynamical systems with
the block-diagonal structure of the set .
Let M () denote the set of all block diagonal and stable rational transfer functions
that have block structures such as .
Theorem 11.8 Let > 0. The loop1 shown below is well-posed and internally stable
for all () 2 M () with kk1 < if and only if
sup (G(j!))
!2R
- e e1
+ 6
+
277
-
G(s)
++ w2
e2 ?
e
the system.
2
Hence, the peak value on the plot of the frequency response determines the size
of perturbations that the loop is robustly stable against.
Remark 11.4 The internal stability with closed ball of uncertainties is more complicated. The following example is shown in Tits and Fan [1994]. Consider
"
#
0 ;1
1
G(s) = s + 1
1 0
and = (s)I2 . Then
sup (G(j!)) = sup jj! 1+ 1j = (G(j 0)) = 1:
! 2R
! 2R
On the other hand, (G(s)) < 1 for all s 6= 0; s 2 C + , and the only matrices in the
form of ; =
I2 with j
j 1 for which
det(I ; G(0);) = 0
are the complex matrices jI2 . Thus, clearly, (I ; G(s)(s));1 2 RH1 for all real
rational (s) = (s)I2 with kk1 1 since (0) must be real. This shows that
sup!2R (G(j!)) < 1 is not necessary for (I ; G(s)(s));1 2 RH1 with the closed
ball of structured uncertainty kk1 1. Similar examples with no repeated blocks are
1 M where M is any real matrix with (M ) = 1 for
generated by setting G(s) = s+1
which there is no real 2 with () = 1 such that det(I ; M ) = 0. For example,
let
82
9
2
3
3
#
0 "
>
>
< 1
=
6
7 ;
6
7
M =4
5
; = >4
2
5 ; i 2 C
>
0 ;
:
;
3
278
s+1
G(s) = 64 0
1
"
1
10
s+1
1
0
0
3
7
5
1
with kk1 < 1. Then Fu (G; ) =
1 2 RH1 for all
1 ; 1 s+1
2
admissible (kk1 < 1) but (I ; G11 );1 2 RH1 is true only for kk1 < 0:1. ~
and =
Often, stability is not the only property of a closed-loop system that must be robust to
perturbations. Typically, there are exogenous disturbances acting on the system (wind
gusts, sensor noise) which result in tracking and regulation errors. Under perturbation,
the eect that these disturbances have on error signals can greatly increase. In most
cases, long before the onset of instability, the closed-loop performance will degrade to
the point of unacceptability, hence the need for a \robust performance" test. Such a
test will indicate the worst-case level of performance degradation associated with a given
level of perturbations.
Assume Gp is a stable, real-rational, proper transfer function with q1 + q2 inputs
and p1 + p2 outputs. Partition Gp in the obvious manner
"
P := 0 0 : 2 ; f 2 C q2 p2 :
f
The setup is to theoretically address the robust performance questions about the
loop shown below
279
- (s)
Gp (s)
z
Theorem 11.9 Let > 0. For all (s) 2 M () with kk1 < 1 , the loop shown
above is well-posed, internally stable, and kFu (Gp ; )k1 if and only if
sup P (Gp (j!)) :
!2R
Note that by internal stability, sup!2R (G11 (j!)) , then the proof of this
theorem is exactly along the lines of the earlier proof for Theorem 11.8, but also appeals
to Theorem 11.7. This is a remarkably useful theorem. It says that a robust performance
problem is equivalent to a robust stability problem with augmented uncertainty as
shown in Figure 11.5.
- f
-
Gp (s)
=
1
2
2 RH1
280
then
"
"
#!
(11:22)
Since the minimization is convex in log d! [see, Doyle, 1982], the optimal d! can be found
by a search; however, two approximations to d! can be obtained easily by approximating
the right hand side of (11.22):
(1) From Lemma 2.10, we have
(G(j!)) d inf
2R
!
=
inf
d! 2R+
"
#!
1
2
2
d2 kG21 (j!)k + kG22 (j!)k
d^! =
8 q
kG21 (j!)k
>
>
<
kG12 (j!)k
>
>
:
(11:23)
(G(j!))
"
inf
d! 2R+
=
=
inf
d 2R
kG11
281
kG11 (j!)k2F + d2! kG12 (j!)k2F + d12 kG21 (j!)k2F + kG22 (j!)k2F
(j!)k2
F + kG22
(j!)k2
8 q
kG21 (j!)kF
>
>
<
kG12 (j!)k
d~! = > 0
>
:
(11:24)
It can be shown that the approximations for the scalar d! obtained above are exact
for a 2 2 matrix G. For higher dimensional G, the approximations for d! are still
reasonably good. Hence an approximation of can be obtained as
(G(j!))
"
#!
"
#!
(11:25)
or, alternatively, as
(G(j!))
(11:26)
We can now see how these approximated tests are compared with the sucient
conditions obtained in Chapter 9.
Example 11.1 Consider again the robust performance problem of a system with output multiplicative uncertainty in Chapter 9 (see Figure 9.5):
kW2 ToW1 k1 1
(11:27)
(11:28)
282
for all 2 RH1 with kk1 < 1. But (11.27) and (11.28) are satised i for each
frequency !
(G(j!)) = d inf
2R
!
"
d! We So W1
We So Wd
#!
1:
Note that, in contrast to the sucient condition obtained in Chapter 9, this condition is
an exact test for robust performance. To compare the test with the criteria obtained
in Chapter 9, some upper bounds for can be derived. Let
s
e So W1 k :
d! = kkW
W2 ToWd k
(G(j!))
where W1 is assumed to be invertible in the last two inequalities. The last term is
exactly the sucient robust performance criteria obtained in Chapter 9. It is clear that
any term preceding the last forms a tighter test since (W1;1 Wd ) 1. Yet another
alternative sucient test can be obtained from the above sequence of inequalities:
q
)2 +
(di )2
= + :
Proof. Consider a function y = x + 1=x, then y is a convex function and the maximization over a closed interval is achieved at the boundary of the interval. Hence for
any xed d
1
1
1
2
2
2
max
(di ) + (d )2 = max (d ) + (d )2 ; (d ) + (d )2 :
i
i
283
Example 11.2 As another example, consider again the skewed specication problem
from Chapter 9. Then the corresponding G matrix is given by
"
We So Wd
(G(j!)) = d inf
2R
!
"
;W2 Ti W1
d! We So PW1
;d! W2 KSoWd
We So Wd
#!
1
for all ! 0. As in the last example, an upper bound can be obtained by taking
s
We So PW1 k :
d! = kkW
2 KSo Wd k
Then
We = ws I; Wd = I; W1 = I; W2 = wt I;
and P is stable and has a stable inverse (i.e., minimum phase) (P can be strictly proper).
Furthermore, we shall assume that the controller has the form
K (s) = P ;1 (s)l(s)
where l(s) is a scalar loop transfer function which makes K (s) proper and stabilizes
the closed-loop. This compensator produces diagonal sensitivity and complementary
sensitivity functions with identical diagonal elements, namely
So = Si = 1 +1l(s) I; To = Ti = 1 +l(sl)(s) I:
284
Denote
;1
G = ;wt I ;wt P
:
ws "P
ws "I
(G(j!)) = dinf
2R
"
#!
ws "I
P (j!) = U V ; = diag(1 ; 2 ; : : : ; m )
with 1 = and m = where m is the dimension of P . Then
(G(j!)) = dinf
2R
"
#!
ws "I
since unitary operations do not change the singular values of a matrix. Note that
"
ws "I
Mi =
"
ws "di
Hence
(G(j!)) = dinf
max
2R+ i
"
ws "
#!
ws "di
ws "
#
!
h
i
;
w
t
= dinf
max
1 (di );1
2R+ i
ws "di
p
= dinf
max
(1 + jdi j;2 )(jws "di j2 + jwt j2 )
2R i
+
"
= dinf
max jws "j2 + jwt j2 + jws "di j2 + wdt :
2R+ i
i
285
Using Lemma 11.10, it is easy to show that the maximum is achieved at either or
and that optimal d is given by
j ;
d2 = jwjw"tj
s
(11:29)
Note that if jws "j and jwt j are not too large, which are guaranteed if the nominal
performance and robust stability conditions are satised, then the structured singular
value is proportional to the square root of the plant condition number:
p
(11:30)
3
This conrms our intuition that an ill-conditioned plant with skewed specications is
hard to control.
The approximations given in the last subsection can be generalized to the multiple block
problem by assuming that M is partitioned consistently with the structure of
= diag(1 ; 2 ; : : : ; F )
so that
M=
and
6
6
6
6
4
..
.
..
.
MF 1 MF 2 MFF
3
7
7
7
7
5
D = diag(d1 I; : : : ; dF ;1 I; I ):
Now
DMD;1 = Mij di ; dF := 1:
dj
And hence
(M )
286
v
u
F X
F
uX
2 d2i
t
Dinf2D kMij k ddi Dinf
k
M
k
ij
2D i=1 j=1
d2j
j
Dinf2D
v
u F F
uX X
t
i=1 j =1
F X
F
X
i=1 j =1
kMij k2 dd2i
j
F X
F
X
i=1 j =1
with dF = 1. The optimal di minimizing of the above two problems satisfy, respectively,
d4
k=
and
2 2
i6=k kMik k di ;
2 2
j 6=k kMkj k =dj
k = 1; 2; : : :; F ; 1
(11:31)
(11:32)
i6=k
k=P
d4
3
7
5
F X
F
X
i=1 j =1
kMij k2 dd2i
j
287
is Dsubopt = diag(0:3212; 0:4643; 1) which is solved from equation (11.31). Using this
Dsubopt , we obtain another upper bound for the structured singular value:
;1 ) = 12:2538:
(M ) (Dsubopt MDsubopt
One may also use this Dsubopt as an initial guess for the exact optimization.
G11 G12
where G =
G21 G22
is chosen, respectively, as
"
(11:34)
288
z
G
-K
by iteratively solving for K and D. This is the so-called D-K Iteration. The stable and
minimum phase scaling matrix D(s) is chosen such that D(s)(s) = (s)D(s) (Note
that D(s) is not necessarily belong to D since D(s) is not necessarily
Hermitian,
see
Remark 11.2). For a xed scaling transfer matrix D, minK
DF` (G; K )D;1
1 is a
standard H1 optimization problem which will be solved
in the later
part of the book.
For a given stabilizing controller K , inf D;D;1 2H1
DF`(G; K )D;1
1 is a standard
convex optimization problem and it can be solved pointwise in the frequency domain:
Indeed,
inf
D;D;1 2H1
This follows intuitively from the following arguments: the left hand side is always no
smaller than the right hand side, and, on the other hand, given the minimizing D!
from the right hand side across the frequency, there is always a rational function D(s)
uniformly approximating the magnitude frequency response D! .
Note that when S = 0, (no scalar blocks)
D! = diag(d!1 I; : : : ; d!F ;1 I; I ) 2 D;
which is a block-diagonal scaling matrix applied pointwise across frequency to the frequency response F` (G; K )(j!).
D-K Iterations proceed by performing this two-parameter minimization in sequential
fashion: rst minimizing over K with D! xed, then minimizing pointwise over D!
with K xed, then again over K , and again over D! , etc. Details of this process are
summarized in the following steps:
289
D
G
D;1
- K
Figure 11.7: -Synthesis via Scaling
(i) Fix an initial estimate of the scaling matrix D! 2 D pointwise across frequency.
(ii) Find scalar transfer functions di (s); d;i 1 (s) 2 RH1 for i = 1; 2; : : :; (F ; 1) such
that jdi (j!)j d!i . This step can be done using the interpolation theory [Youla
and Saito, 1967]; however, this will usually result in very high order transfer
functions, which explains why this process is currently done mostly by graphical
matching using lower order transfer functions.
(iii) Let
D(s) = diag(d1 (s)I; : : : ; dF ;1 (s)I; I ):
Construct a state space model for system
"
G^ (s) = D(s)
"
;1
G(s) D (s)
^ K )
F` (G;
1
over all stabilizing K 's. Note that this optimization problem uses the scaled
version of G. Let its minimizing controller be denoted by K^ .
(v) Minimize [D! F` (G; K^ )D!;1 ] over D! , pointwise across frequency.4 Note that
this evaluation uses the minimizing K^ from the last step, but that G is unscaled.
The minimization itself produces a new scaling function. Let this new function
be denoted by D^ ! .
(vi) Compare D^ ! with the previous estimate D! . Stop if they are close, but, otherwise,
replace D! with D^ ! and return to step (ii).
With either K or D xed, the global optimum in the other variable may be found using
the and H1 solutions. Although the joint optimization of D and K is not convex and
4
290
the global convergence is not guaranteed, many designs have shown that this approach
works very well [see e.g. Balas, 1990]. In fact, this is probably the most eective design
methodology available today for dealing with such complicated problems. The detailed
treatment of analysis is given in Packard and Doyle [1991]. The rest of this book will
focus on the H1 optimization which is a fundamental tool for synthesis.
12
Parameterization of
Stabilizing Controllers
The basic conguration of the feedback systems considered in this chapter is an LFT
as shown in Figure 12.1, where G is the generalized plant with two sets of inputs: the
exogenous inputs w, which include disturbances and commands, and control inputs u.
The plant G also has two sets of outputs: the measured (or sensor) outputs y and the
regulated outputs z . K is the controller to be designed. A control problem in this setup
is either to analyze some specic properties, e.g., stability or performance, of the closedloop or to design the feedback control K such that the closed-loop system is stable
in some appropriate sense and the error signal z is specied, i.e., some performance
condition is satised. In this chapter we are only concerned with the basic internal
stabilization problems. We will see again that this setup is very convenient for other
general control synthesis problems in the coming chapters.
Suppose that a given feedback system is feedback stabilizable. In this chapter, the
problem we are mostly interested in is parameterizing all controllers that stabilize the
system. The parameterization of all internally stabilizing controllers was rst introduced
by Youla et al [1976]; in their parameterization, the coprime factorization technique is
used. All of the existing results are mainly in the frequency domain although they can
also be transformed to state-space descriptions. In this chapter, we consider this issue
in the general setting and directly in state space without adopting coprime factorization
technique. The construction of the controller parameterization is done via considering
a sequence of special problems, which are so-called full information (FI) problems, disturbance feedforward (DF) problems, full control (FC) problems and output estimation
291
292
w
G
y
u
- K
A B1 B2
G
11 (s) G12 (s)
6
G(s) =
= 4 C1 D11 D12 75 :
(12:1)
G21 (s) G22 (s)
C2 D21 D22
The stabilization problem is to nd feedback mapping K such that the closed-loop
"
system is internally stable; the well-posedness is required for this interconnection. This
general synthesis question will be called the output feedback (OF) problem.
293
Lemma 12.1 There exists a proper K achieving internal stability i (A; B2) is stabilizable and (C2 ; A) is detectable. Further, let F and L be such that A + B2 F and A + LC2
are stable, then an observer-based stabilizing controller is given by
#
"
Proof. (() By the stabilizability and detectability assumptions, there exist F and
L such that A + B2 F and A + LC2 are stable. Now let K (s) be the observer-based
controller given in the lemma, then the closed-loop A-matrix is given by
A~ =
"
A
B2 F
:
;LC2 A + B2 F + LC2
A + LC2
0
:
;LC2 A + B2 F
Thus the spectrum of A~ equals the union of the spectra of A + LC2 and A + B2 F . In
particular, A~ is stable.
()) If (A; B2 ) is not stabilizable or if (C2 ; A) is not detectable, then there are some
eigenvalues of A~ which are xed in the right half-plane, no matter what the compensator
is. The details are left as an exercise.
2
The stabilizability and detectability conditions of (A; B2 ; C2 ) are assumed throughout the remainder of this chapter1. It follows that the realization for G22 is stabilizable
and detectable, and these assumptions are enough to yield the following result:
G22
u
- K
Figure 12.2: Equivalent Stabilization Diagram
1 It should be clear that the stabilizability and detectability of a realization for G do not guarantee
the stabilizability and/or detectability of the corresponding realization for G22 .
294
In other words, K (s) internally stabilizes G(s) if and only if it internally stabilizes
G22 .
Proof. The necessity follows from the denition. To show the suciency, it is sucient
to show that the system in Figure 12.1 and that in Figure 12.2 share the same A-matrix,
which is obvious.
2
From Lemma 12.2, we see that the stabilizing controller for G depends only on G22 .
Hence all stabilizing controllers for G can be obtained by using only G22 , which is how
it is usually done in the conventional Youla parameterization. However, it will be shown
that the general setup is very convenient and much more useful since any closed-loop
system information can also be considered in the same framework.
Remark 12.1 There should be no confusion between a given realization for a transfer
matrix G22 and the inherited realization from G where G22 is a submatrix. A given
realization for G22 may be stabilizable and detectable while the inherited realization
may be not. For instance,
"
G22 = s +1 1 = ;1 1
1 0
;1 0 0 1
6
G11 G12 = 66 0 1 1 0
6
G21 G22
1 0 0
4 0
1 0 0 0
is
;1 0 1
G22 = 64 0 1 0
1 0 0
3
7
5
= s +1 1
3
7
7
7
7
5
295
Before we get into the study of the algebraic structure of control systems, we now introduce the concept of algebraic duality which will play an important role. It is well known
that the concepts of controllability (stabilizability) and observability (detectability) of
a system (C; A; B ) are dual because of the duality between (C; A; B ) and (B T ; AT ; C T ).
So, to deal with the issues related to a system's controllability and/or observability, we
only need to examine the issues related to the observability and/or controllability of its
dual system, respectively. The notion of duality can be generalized to a general setting.
Consider a standard system block diagram
z
w
G
y
u
- K
where the plant G and controller K are assumed to be linear time invariant. Now
consider another system shown below
z~
w~
y~
GT
- KT
u~
whose plant and controller are obtained by transposing G and K . We can check easily
T = [F` (G; K )]T = F` (GT ; K T ) = Tz~w~ . It is not dicult to see that K internally
that Tzw
stabilizes G i K T internally stabilizes GT . And we say that these two control structures
are algebraically dual, especially, GT and K T which are dual objects of G and K ,
respectively. So as far as stabilization or other synthesis problems are concerned, we
can obtain the results for GT from the results for its dual object G if they are available.
Now, we consider some special problems which are related to the general OF problems stated in the last section and which are important in constructing the results for
OF problems. The special problems considered here all pertain to the standard block
diagram, but to dierent structures than G. The problems are labeled as
296
A
6
6 C1
GFI = 66 " #
I
4
B1
D
" 11#
B2
7
D 7
" 12# 7 :
7
0
0
A B1
6
GFC = 664 C1 D11
C2 D21
h
h
h
i 3
I 0
0 I
i 7
7
7
i 5
0 0
A B1 B2
GDF = 64 C1 D11 D12 75 :
C2 I
0
OE. Output estimation, with the corresponding plant
2
A B1 B2
6
GOE = 4 C1 D11 I 75 :
C2 D21 0
The motivations for these special problems will be given later when they are considered. There are also two additional structures which are standard and which will not
be considered in this chapter; they are
SF. State feedback
A B1 B2
6
GSF = 4 C1 D11 D12 75 :
I 0 0
2
A B1 I
6
GOI = 4 C1 D11 0 75 :
C2 D21 0
297
Here we assume that all physical variables have compatible dimensions. We say
that these special problems are special cases of OF problems in the sense that their
structures are specied in comparison to OF problems.
The structure of transfer matrices shows clearly that FC, OE (and OI) are duals of
FI, DF (and SF), respectively. These relationships are shown in the following diagram:
FI
dual
- FC
6
equivalent
equivalent
dual
- OE
DF
The precise meaning of \equivalent" in this diagram will be explained below.
"
x .
In the FI problem, the controller is provided with Full Information since y =
w
For the FI problem, we only need to assume that (A; B2 ) is stabilizable to guarantee the
solvability. It is clear that if any output feedback control problem is to be solvable then
the corresponding FI problem has to be solvable, provided FI is of the same structure
with OF except for the specied parts.
To motivate the name Disturbance Feedforward, consider the special case with C2 =
0. Then there is no feedback and the measurement is exactly w, where w is generally
regarded as disturbance to the system. Only the disturbance, w, is fed through directly
to the output. As we shall see, the feedback caused by C2 6= 0 does not aect the
transfer function from w to the output z , but it does aect internal stability. In fact,
the conditions for the solvability of the DF problem are that (A; B2 ) is stabilizable and
(C2 ; A) is detectable.
Now we examine the connection between the DF problem and the FI problem and
show the meaning of their equivalence. Suppose that we have controllers KFI and KDF
and let TFI and TDF denote the closed-loop Tzw s in
z
w
GFI
yFI
u
- KFI
z
w
GDF
yDF
u
- KDF
298
The question of interest is as follows: given either the KFI or the KDF controller,
can we construct the other in such a way that TFI = TDF ? The answer is positive.
Actually, we have the following:
Lemma 12.3 Let GFI and GDF be given as above. Then
"
#
I
0 0
(i) GDF (s) =
GFI (s).
0 C2 I
(ii) GFI = S (GDF ; PDF ) (where S (; ) denotes the star-product)
GDF
aa !!!
a
!
!! aaa
PDF
2
6
6
PDF (s) = 66
4
A ; B1 C2
"
I
; C2
B1
0
0
"
B2
7
I # 77 :
"
7
0
0
Proof. Part (i) is obvious. Part (ii) follows from applying the star product formula.
Nevertheless, we will give a direct proof to show the system structure. Let x and x^
denote the state of GDF and PDF , respectively. Take e := x ; x^ and x^ as the states of
the resulting interconnected system; then its realization is
2
6
6
6
6
6
6
4
A ; B1 C2
B1 C2
C
" 1 #
0
A
C
" 1#
I
B1
D
" 11#
0
I
0
which is exactly GFI , as claimed.
C2
7
2 7
7
7
12
"
# 7
0 75
B
D
A
6
6 C1
= 66 " #
I
4
B1
D
" 11#
B2
D
" 12#
0
0
3
7
7
7
7
5
Remark 12.2 There is an alternative way to see part (ii). The fact is that in the DF
problems, the disturbance w and the system states x can be solved in terms of y and u:
2
A ; B1 C2 B1 B2
x =6
I
0 0
4
w
;C2
I 0
Now connect V up with GDF as shown below
"
3
7
5
"
y =: V
u
"
y :
u
299
GDF
y
x
w
w
u
Then it is easy to show that the transfer function from (w; u) to (z; x; w) is GFI , and,
furthermore, that the internal stability is not changed if A ; B1 C2 is stable.
~
The following theorem follows immediately:
u
y^
- KFI
yDF
u
Remark 12.3 This theorem shows that if A ; B1C2 is stable, then problems FI and
DF are equivalent in the above sense. Note that the transfer function from w to yDF is
"
G21 (s) = A B1 :
C2 I
Hence this stability condition implies that G21 (s) has neither right half plane invariant
zeros nor hidden unstable modes.
~
300
z = C1 x + D11 w + u
where z is to be controlled by an appropriately designed control u. In general, our
control objective will be to specify z in some well-dened mathematical sense. To put
it in other words, it is desired to nd a u that will estimate C1 x + D11 w in such dened
mathematical sense. So this kind of control problem can be regarded as an estimation
problem. We are focusing on this particular estimation problem because it is the one
that arises in solving the output feedback problem. A more conventional estimation
problem would be the special case where no internal stability condition is imposed
and B2 = 0. Then the problem would be that of estimating the output z given the
measurement y. This special case motivates the term output estimation and can be
obtained immediately from the results obtained for the general case.
The following discussion will explain the meaning of equivalence between FC and
OE problems. Consider the following FC and OE diagrams:
z
yFC
GFC
- KFC
w
u
z
yOE
GOE
- KOE
I 0
6
(i) GOE (s) = GFC (s) 4 0 B2
0 I
3
7
5
w
u
A ; B2 C1 0
6
6
POE (s) = 64 C1
0
C2
I
I ;B2i
h
0 I i
h
0 0
301
i 3
7
7
7
5
"
Remark 12.4 It is seen that if A ; B2C1 is stable, then FC and OE problems are
equivalent in the above sense. This condition implies that the transfer matrix G12 (s)
from u to z has neither right half-plane invariant zeros nor hidden unstable modes,
which indicates that it has a stable inverse.
~
Consider again the standard system block diagram in Figure 12.1 with
2
"
#
A B1 B2
G
7
11 (s) G12 (s)
6
:
G(s) = 4 C1 D11 D12 5 =
G21 (s) G22 (s)
C2 D21 D22
Suppose (A; B2 ) is stabilizable and (C2 ; A) is detectable. In this section we discuss the
following problem:
302
This parameterization for all stabilizing controllers is usually called Youla parameterization. As we have mentioned early, the stabilizing controllers for G will depend only on
G22 . However, it is more convenient to consider the problem in the general framework
as will be shown. The parameterization of all stabilizing controllers is easy when the
plant itself is stable.
Theorem 12.7 Suppose G 2 RH1; then the set of all stabilizing controllers can be
described as
(12:2)
Remark 12.5 This result is very natural considering Corollary 5.5, which says that a
controller K stabilizes a stable plant G22 i K (I ; G22 K );1 is stable. Now suppose
Q = K (I ; G22 K );1 is a stable transfer matrix, then K can be solved from this equation
which gives exactly the controller parameterization in the above theorem.
~
Proof. Note that G22 (s) is stable by the assumptions on G. Now use straightforward
algebra to verify that the controllers given above stabilize G22 . On the other hand,
suppose K0 is a stabilizing controller; then Q0 := K0(I ; G22 K0 );1 2 RH1 , so K0 can
be expressed as K0 = Q0 (I + G22 Q0 );1 . Note that the invertibility in the last equation
is guaranteed by the well posedness of the interconnected system with controller K0
since I + D22 Q0 (1) = (I ; D22 K0 (1));1 .
2
Theorem 12.8 Let F and L be such that A + LC2 and A + B2F are stable, and then all
controllers that internally stabilize G can be parameterized as the transfer matrix from
y to u below
u
y
J
- Q
3
7
5
A non-constructive proof of the theorem can be given by using the same argument
as in the proof of Theorem 12.7, i.e., rst verify that any controller given by the formula
indeed stabilizes the system G, and then show that we can construct a stable Q for any
303
given stabilizing controller K . This approach, however, does not give much insight into
the controller construction and thus can not be generalized to other synthesis problems.
The conventional Youla approach to this problem is via coprime factorization [Youla
et al, 1976, Vidyasagar, 1985, Desoer et al, 1982], which will be adopted in the later
part of this chapter as an alternative approach.
In the following sections, we will present a novel approach to this problem without
adopting coprime factorizations. The idea of this approach is to reduce the output
feedback problem into some simpler problems, such as FI and OE or FC and DF which
admit simple solutions, and then to solve the output feedback problem by the separation
argument. The advantages of this approach are that it is simple and that many other
synthesis problems, such as H2 and H1 optimal control problems in Chapters 14 and 16,
can be solved by using the same machinery.
Readers should bear in mind that our objective here is to nd all admissible controllers for the OF problem. So at rst, we will try to build up enough tools for this
objective by considering the special problems. We will see that it is not necessary to parameterize all stabilizing controllers for these special problems to get the required tools.
Instead, we only parameterize some equivalent classes of controllers which generate the
same control action.
Denition 12.2 Two controllers K and K^ are of equivalent control actions if their
corresponding closed loop transfer matrices are identical, i.e. Fl (G; K ) = Fl (G; K^ ),
written as K
= K^ .
Algebraically, the controller equivalence is an equivalence relation. We will see that
for dierent special problems we have dierent rened versions of this relation. We
will also see that the characterizations of equivalent classes of stabilizing controllers for
special problems are good enough to construct the parameterization of all stabilizing
controllers for the OF problem. In the next two subsections, we mainly consider the
stabilizing controller characterizations for special problems. Also, we use the solutions
to these special problems and the approach provided in the last section to characterize
all stabilizing controllers of OF problems.
z
yFI
GFI
- KFI
w
u
where the transfer matrix GFI is given in section 12.2. The purpose of this subsection
is to characterize the equivalent classes of stabilizing controllers KFI that stabilize
304
internally GFI and to build up enough tools to be used later. For this problem, we
say two controllers KFI and K^ FI are equivalent if they produce the same closed-loop
transfer function from w to u. Obviously, this also guarantees that Fl (GFI ; KFI ) =
Fl (GFI ; K^ FI ). The characterization of equivalent classes of controllers can be suitably
called the control parameterization in contrast with controller parameterization. Note
that the same situation will occur in FC problems by duality.
Since we have full information for feedback, our controller will have the following
general form:
h
i
KFI = K1 (s) K2(s)
with K1 (s) stabilizing internally (sI ; A);1 B2 and arbitrary K2 (s) 2 RH1 . Note that
the stability of K2 is required to guarantee the internal stability of the whole system
since w is fed directly through K2.
Lemma 12.9 Let F be a constant matrix such that A + B2F is stable. Then all stabilizing controllers, in the sense of generating all admissible closed-loop transfer functions,
for FI can be parameterized as
h
KFI
= F Q
Proof. It is easy to see that the controller given in the above formula stabilizes the
system GFI . Hence we only need to show that the given set of controllers parameterizes
all stabilizing control action, u, i.e., there is a choice of Q 2 RH
h 1 such that thei transfer
functions from w to u for any stabilizing controller KFI = K1 (s) K2 (s) and for
h
i
0 = F Q are the same. To show this, make a change of control variable as
KFI
v = u ; Fx, where x denotes the state of the system GFI ; then the system with the
controller KFI will be as shown in the following diagram:
z
yFI
with
A + B2 F
6
6 C1 + D12 F
G~ FI = 66 " #
I
4
0
w
v
G~ FI
- K~ FI
B1
D
" 11#
0
B2
7
7
D
12
~ FI = KFI ; [F 0]:
"
# 7; K
0 7
0
305
shown by directly computing Q from equating the transfer matrices from w to u for the
0 . In fact, the transfer matrices from w to u with KFI and K 0
cases of KFI and KFI
FI
are given by
I ; K1(sI ; A);1 B2 ;1 K1(sI ; A);1 B1 + I ; K1 (sI ; A);1 B2 ;1 K2
(12:3)
I ; F (sI ; A);1 B2 ;1 F (sI ; A);1 B1 + I ; F (sI ; A);1 B2 ;1 Q;
(12:4)
and
Now we consider the dual FC problem; the system diagram pertinent to this case is
z
w
GFC
yFC
u
- KFC
Dually, we say controllers KFC and K^ FC are equivalent in the sense that the same
injection inputs yFC 's produce the same outputs z 's. This also guarantees the identity
of their resulting closed-loop transfer matrices from w to z . And we also have
Lemma 12.10 Let L be a constant matrix such that A + LC2 is stable. Then the
set of equivalent classes of all stabilizing controllers for FC in the above sense can be
parameterized as
"
#
L
KFC
=
Q
306
z
yDF
GDF
- KDF
w
u
The transfer matrix is given as in section 12.2.1. We will further assume that A ; B1 C2
is stable in this subsection. It should be pointed out that the existence of a stabilizing
controller for this system is guaranteed by the stabilizability of (A; B2 ) and detectability
of (C2 ; A). Hence this assumption is not necessary for our problem to be solvable;
however, it does simplify the solution.
We will now parameterize stabilizing controllers for GDF by invoking the relationship
between the FI problem and DF problem established in section 12.2. We say that the
controllers KDF and K^ DF are equivalent for the DF problem if the two transfer matrices
from w to u in the above diagram are the same. Of course, the resulting two closed-loop
transfer matrices from w to z are identical.
Remark 12.7 By the equivalence between FI and DF
it is heasy to show
h problems,
i
i
^
^
that if KDF = KDF in the DF structure, then KDF C2 I = KDF C2 I in
the corresponding FI structure. We also have that if KFI
= K^ FI , then Fl (PDF ; KFI )
=
^
Fl (PDF ; KFI ).
~
Now we construct the parameterization of equivalent classes of stabilizing controllers
in DF structure via the tool we have developed in section 12.2.
h
i
Let KDF (s) be a stabilizing control for DF ; then KFI (s) = KDF (s) C2 I
h
i
stabilizes the corresponding GFI . Assume KFI
= K^ FI = F Q for some Q 2
RH1 ; then K^ FI stabilizes GFI and F` (JDF ; Q) = F` (PDF ; K^ FI ) where
2
A + B2 F ; B1 C2 B1 B2
JDF = 64
F
0 I
I 0
;C2
3
7
5
307
Lemma 12.11 All stabilizing controllers for the DF problem can be characterized by
KDF = F` (JDF ; Q) with Q 2 RH1 , where JDF is given as above.
Proof. We have already shown that the controller KDF = F`(JDF ; Q) for any given
Q 2 RH1 does internally stabilize GDF . Now let KDF be any stabilizing controller for
GDF ; then F` (J^DF ; KDF ) 2 RH1 where
2
A B1 B2
6
^
JDF = 4 ;F 0 I 75 :
C2 I 0
(J^DF is stabilized by KDF since it has the same `G22 ' matrix as GDF .)
Let Q0 := F` (J^DF ; KDF ) 2 RH1 ; then F`(JDF ; Q0 ) = F`(JDF ; F`(J^DF ; KDF )) =:
F` (Jtmp ; KDF ), where Jtmp can be obtained by using the state space star product
formula given in Chapter 10:
3
A ; B1 C2 + B2 F ;B2F B1 B2
7
6
6
;B1 C2
A B1 B2 77
Jtmp = 66
F
;F 0 I 75
4
I 0
;C2
C2
3
2
A ; B1 C2 ;B2F B1 B2
7
6
6
0
A + B2 F 0 0 77
= 66
0
;F
0 I 75
4
I 0
0
C2
"
#
0 I
=
:
I 0
Hence F` (JDF ; Q0 ) = F`(Jtmp ; KDF ) = KDF . This shows that any stabilizing
controller can be expressed in the form of F` (JDF ; Q0 ) for some Q0 2 RH1 .
2
Now we turn to the dual OE case. The corresponding system diagram is shown as
below:
z
yOE
GOE
- KOE
w
u
We will assume that A ; B2 C1 is stable. Again this assumption is made only for
the simplicity of the solution, it is not necessary for the stabilization problem to be
308
A ; B2 C1 + LC2 L ;B2
6
JOE = 4
C1
0 I
C2
I 0
3
7
5
Lemma 12.12 All admissible controllers for the OE problem can be characterized as
F` (JOE ; Q0 ) with any Q0 2 RH1 , where JOE is dened as above.
Proof. The controllers in the form as stated in the theorem are admissible since the
corresponding FC controllers internally stabilize resulting GFC .
Now assume KOE is any stabilizing controller for GOE ; then F`(J^OE ; KOE ) 2 RH1
where
2
3
A ;L B2
J^OE = 64 C1
C2
7
5
We are now ready to give a complete proof for Theorem 12.8. We will assume the
results of the special problems and show how to construct all admissible controllers for
the OF problem from them. And we can also observe the separation argument as the
309
z
w
G
y
u
- K
with
A B1 B2
G(s) = 64 C1 D11 D12 75
C2 D21 D22
and that (A; B2 ) is stabilizable and (C2 ; A) is detectable.
Proof of Theorem 12.8. Without loss of generality, we shall assume D22 = 0. For
more general cases, i.e. D22 6= 0, the mapping
K^ (s) = K (s)(I ; D22 K (s));1
is well dened if the closed-loop system is assumed to be well posed. Then the system
in terms of K^ has the structure
z
y
where
w
G^
- K^
A B1 B2
6
^
G(s) = 4 C1 D11 D12 75 :
C2 D21 0
Now we construct the controllers for the OF problem with D22 = 0. Denote x the
state of system G; then the open-loop system can be written as
2
6
4
x_
A B1 B2
7 6
z 5 = 4 C1 D11 D12
y
C2 D21 0
32
76
54
x
w 75 :
u
310
v = u ; Fx:
Then the system can be broken into two subsystems:
"
"
x_ = A + B2 F B1 B2
z
C1 + D12 F D11 D12
and
2
6
4
A B1 B2
x_
7 6
v 5 = 4 ;F 0 I
C2 D21 0
y
32
76
54
2
6
4
x
w
v
3
7
5
x
w 75 :
u
G1 v
z
y
Gtmp
w
u
- K
with
and
"
G1 = A + B2 F B1 B2 2 RH1
C1 + D12 F D11 D12
2
A B1 B2
6
Gtmp = 4 ;F 0 I 75 :
C2 D21 0
Obviously, K stabilizes G if and only if K stabilizes Gtmp ; however, Gtmp is of OE
structure. Now let L be such that A + LC2 is stable. Then by Lemma 12.12 all
controllers stabilizing Gtmp are given by
K = F`(J; Q)
311
where
2
A + B2 F + LC2 L ;B2
J = 64
;F
0 I
C2
I 0
A + B2 F + LC2 ;L B2
7 6
=
F
0 I 75 :
5 4
;C2
I 0
Remark 12.8 We can also get the same result by applying the dual procedure to the
above construction, i.e., rst use an output injection to reduce the OF problem to a DF
problem. The separation argument is obvious since the synthesis of the OF problem
can be reduced to FI and FC problems, i.e. the latter two problems can be designed
independently.
~
Remark 12.9 Theorem 12.8 shows that any stabilizing controller K (s) can be characterized as an LFT of a parameter matrix Q 2 RH1 , i.e., K (s) = F` (J; Q). Moreover,
using the same argument as in the proof of Lemma 12.11, a realization of Q(s) in terms
of K can be obtained as
Q := F` (J;^ K )
where
3
2
A ;L B2
0 I
C2 I D22
J^ = 64 ;F
7
5
Now we can reconsider the characterizations of all stabilizing controllers for the
special problems with some reasonable assumptions, i.e. the stability conditions of
A ; B1 C2 and A ; B2 C1 for DF and OE problems which were assumed in the last
section can be dropped.
If we specify
"
#
"
#
"
#
I
0
0
C2 =
D21 =
D22 =
;
0
I
0
the OF problem, in its general setting, becomes the FI problem. We know that the
solvability conditions for the FI problem are reduced, because of its special structure,
to (A; B2 ) as stabilizable. By assuming this, we can get the following result from the
OF problem.
Corollary 12.13 Let L1 and F be such that A + L1 and A + B2 F are stable; then all
controllers that stabilize GFI can be characterized as F`(JFI ; Q) with any Q 2 RH1 ,
312
where
A + B2 F + L1 L ;B2
6
6
;F #
0 I
"
JFI = 66
I
4
I 0
3
7
7
7
7
5
0
and where L = (L1 L2) is the injection matrix for any L2 with compatible dimensions.
In the same way, we can consider the FC problem as the special OF problem by
specifying
h
i
h
i
h
i
B2 = I 0 D12 = 0 I D22 = 0 0 :
The DF (OE ) problem can also be considered as the special case of OF by simply
setting D21 = I (D12 = I ) and D22 = 0.
Corollary 12.14 Consider the DF problem, and assume that (C2; A; B2 ) is stabilizable
and detectable. Let F and L be such that A + LC2 and A + B2 F are stable, and then
all controllers that internally stabilize G can be parameterized as Fl (JDF ; Q) for some
Q 2 RH1 , i.e. the transfer function from y to u is shown as below
u
JDF
- Q
y
A + B2 F + LC2 ;L B2
6
JDF = 4
F
0 I 75 :
;C2
I 0
Remark 12.10 It would be interesting to compare this result with Lemma 12.11. It
can be seen that Lemma 12.11 is a special case of this corollary. The condition that
A ; B1 C2 is stable, which is required in Lemma 12.11, provides the natural injection
matrix L = ;B1 which satises a partial condition in this corollary.
~
C2 D22
We choose F and L so that A + B2 F and A + LC2 are stable. Dene J by the formula
in Theorem 12.8. Then the proper K 's achieving internal stability are precisely those
representable in Figure 12.3 and K = F` (J; Q) where Q 2 RH1 and I + D22 Q(1) is
invertible.
313
G
D22
?
c?
;
C2
-A
c c B2
66
-F
- ;L
u1
6
c
y1
- Q
Figure 12.3: Structure of Stabilizing Controllers
"
u =J y :
u1
y1
It is also easy to show that the transfer matrix from y1 to u1 is zero.
This diagram of all stabilizing controller parameterization also suggests an interesting interpretation: every internal stabilization amounts to adding stable dynamics to
the plant and then stabilizing the extended plant by means of an observer. The precise
statement is as follows: for simplicity of the formulas, only the cases of strictly proper
G22 and K are treated.
Theorem 12.15 Assume that G22 and K are strictly proper and the system is Figure 12.1 is internally stable. Then G22 can be embedded in a system
"
where
"
Ae Be
Ce 0
"
h
Ae = A 0 ; Be = B2 ; Ce = C2 0
0 Aa
0
(12:5)
314
K = Ae + Be Fe + Le Ce ;Le
Fe
0
where Ae + Be Fe and Ae + Le Ce are stable.
(12:6)
Q = Aa Ba :
Ca 0
Since Q 2 RH1 , Aa is stable. Let x and xa denote state vectors for J and Q, respectively, and write the equations for the system in Figure 12.3:
x_
u
u1
x_ a
y1
(A + B2 F + LC2 )x ; Ly + B2 y1
Fx + y1
;C2 x + y
Aa xa + Ba u1
=
=
=
=
=
Ca xa
x_ e = (Ae + Be Fe + Le Ce )xe ; Le y
u = Fe xe
where
"
"
h
i
xe := x ; Fe := F Ca ; Le := L
xa
; Ba
and where Ae ; Be ; Ce are as in (12.5).
Recall that the closed-loop transfer matrix from w to z is a linear fractional transformation F`(G; K ) and that K stabilizes G if and only if K stabilizes G22 . Elimination
of the signals u and y in Figure 12.3 leads to Figure 12.4 for a suitable transfer matrix
T . Thus all closed-loop transfer matrices are representable as in Figure 12.4.
(12:7)
315
- Q
Figure 12.4: Closed loop system
Theorem 12.16 Let F and L be such that A + BF and A + LC are stable. Then the
set of all closed-loop transfer matrices from w to z achievable by an internally stabilizing
proper controller is equal to
F` (T; Q) = fT11 + T12 QT21 : Q 2 RH1 ; I + D22 Q(1) invertibleg
where T is given by
2
"
T = T11
T21
;B2 F
B1
B2
# 6
7
6
0
A + LC2 B1 + LD21 0 77 :
T12 = 6
6
T22
;D12F
D11
D12 75
4 C1 + D12 F
0
C2
D21
0
A + B2 F
Proof. This is straightforward by using the state space star product formula and
follows from some tedious algebra. Hence it is left for the readers to verify.
An important point to note is that the closed-loop transfer matrix is simply an ane
function of the controller parameter matrix Q since T22 = 0.
316
for Ql 2 RH1 where U0 ; V0 ; U~0 ; V~0 2 RH1 satisfy the Bezout identities:
~ 0 ; NU
~ 0 = I:
V~0 M ; U~0 N = I; MV
Moreover, if U0 ; V0 ; U~0 , and V~0 are chosen such that U0V0;1 = V~0;1 U~0 , i.e.,
"
#"
# "
#
V~0 ;U~0
M U0 = I 0 :
;N~ M~
N V0
0 I
Then
K = (U0 + MQy )(V0 + NQy );1
= (V~0 + Qy N~ );1 (U~0 + Qy M~ )
= F` (Jy ; Qy )
where
"
#
;1
~0;1
U
V
0 V0
Jy :=
;1
;1
V0
;V0 N
(12.10)
(12:11)
and where Qy ranges over RH1 such that (I + V0;1 NQy )(1) is invertible
Proof. We shall prove the parameterization given in (12.8) rst. Assume that K has
317
Thus,
K = UV ;1
= (U0 + MQr )(V0 + NQr );1 :
To see that Qr belongs to RH1 , observe rst from (12.12) and then from (12.13) that
both MQr and NQr belong to RH1 . Then
Qr = (V~0 M ; U~0 N )Qr = V~0 (MQr ) ; U~0 (NQr ) 2 RH1 :
Finally, since V and Z evaluated at s = 1 are both invertible, so is V0 + NQr from
(12.13), hence so is I + V0;1 NQr .
Similarly, the parameterization given in (12.9) can be obtained.
To show that the controller can be written in the form of equation (12.10), note that
(U0 + MQy )(V0 + NQy );1 = U0 V0;1 + (M ; U0 V0;1 N )Qy (I + V0;1 NQy );1 V0;1
and that U0 V0;1 = V~0;1 U~0 . We have
and
(12:14)
~ ; NU:
~
Z := MV
Next, we shall establish the precise relationship between the above all stabilizing
controller parameterization and the parameterization obtained in the previous sections
via LFT framework.
N V0
= 64
F
I
C2 + D22 F D22
7
5
318
V~0 ;U~0
;N~ M~
Remark 12.11
Note that Jy is exactly the same as the J in Theorem 12.8 and that
;1
K0 := U0V0 is an observer-based stabilizing controller with
"
K0 := A + B2 F + LC2 + LD22 F ;L :
F
0
13
H := A R :
;Q ;A
(13:2)
A matrix of this form is called a Hamiltonian matrix. The matrix H in (13.2) will
be used to obtain the solutions to the equation (13.1). It is useful to note that (H )
(the spectrum of H) is symmetric about the imaginary axis. To see that, introduce the
2n 2n matrix:
"
#
0 ;I
J :=
I 0
having the property J 2 = ;I . Then
J ;1 HJ = ;JHJ = ;H
319
320
V = Im X1 :
X2
If X1 is invertible, then X := X2 X1;1 is a solution to the Riccati equation (13.1) and
(A + RX ) = (H jV ). Furthermore, the solution X is independent of a specic choice
of bases of V .
A R
;Q ;A
#"
A R
;Q ;A
#"
"
X1 = X1 :
X2
X2
#
"
I = I X X ;1:
X
X 1 1
(13:3)
321
to get
"
#"
A R
I
0 = ;X I
;Q ;A
X
= ;XA ; A X ; XRX ; Q;
h
which shows that X is indeed a solution of (13.1). Equation (13.3) also gives
A + RX = X1 X1;1;
therefore, (A + RX ) = (). But, by denition, is a matrix representation of the
map H jV , so (A + RX ) = (H jV ). Finally note that any other basis spanning V can
be represented as
"
#
"
#
X1 P = X1 P
X2
X2 P
for some nonsingular matrix P . The conclusion follows from the fact (X2 P )(X1 P );1 =
X2 X1;1.
2
As we would expect, the converse of the theorem also holds.
X1
X2
"
A R
;Q ;A
I
X
#"
"
I = I :
X
X
322
Remark 13.1 It is now clear that to obtain solutions to the Riccati equation, it is
necessary to be able to construct bases for those invariant subspaces of H . One way of
constructing those invariant subspaces is to use eigenvectors and generalized eigenvectors
of H . Suppose i is an eigenvalue of H with multiplicity k (then i+j = i for all
j = 1; : : : ; k ; 1), and let vi be a corresponding eigenvector and vi+1 ; : : : ; vi+k;1 be the
corresponding generalized eigenvectors associated with vi and i . Then vj are related
by
(H ; i I )vi
= 0
(H ; i I )vi+1
= vi
..
.
(H ; i I )vi+k;1 = vi+k;2 ;
and the spanfvj ; j = i; : : : ; i + k ; 1g is an invariant subspace of H .
~
"
"
"
A = ;3 2 R = 0 0 ; Q = 0 0 :
;2 1
0 ;1
0 0
The eigenvalues of H are 1; 1; ;1; ;1, and the corresponding eigenvector and generalized
eigenvector of 1 are
1
;1 7
6
7
6
6 2 7
6 ;3=2 7
7 ; v2 = 6
7:
v1 = 66
7
6
1 75
4 2 5
4
;2
0
The corresponding eigenvector and generalized eigenvector of ;1 are
2
v3 =
6
6
6
6
4
1
7
6
6
1 77
; v4 = 66
7
05
4
0
1
7
3=2 77
:
0 75
0
All solutions of the Riccati equation under various combinations are given below:
"
"
;10 6
6 ;4
;2 2
2 ;2
323
ers can verify that the matrices constructed from those vectors are not solutions
to the Riccati equation (13.1).
3
Up to this point, we have said nothing about the structure of the solutions given by
Theorem 13.1 and 13.2. The following theorem gives a sucient condition for a Riccati
solution to be hermitian (not necessarily real symmetric).
be such that
"
V = Im X1 :
X2
Then i + j 6= 0 for all i; j = 1; : : : ; n, i ; j 2 (H jV ) implies that X1 X2 is hermitian, i.e., X1X2 = (X1X2 ) . Furthermore, if X1 is nonsingular, then X = X2 X1;1 is
hermitian.
"
H X1 = X1 :
X2
X2
"
X1
Pre-multiply this equation by
X2
"
X1
X2
#
#
"
J to get
#
"
JH X1 = X1
X2
X2
#
"
J X1 :
X2
(13:4)
Note that JH is hermitian (actually symmetric since H is real); therefore, the left-hand
side of (13.4) is hermitian as well as the right-hand side:
"
i.e.,
X1
X2
#
"
"
J X1 = X1
X2
X2
#
J
"
"
X1 = ; X1
X2
X2
#
"
J X1
X2
324
This is a Lyapunov equation. Since i + j 6= 0, the equation has a unique solution:
;X1X2 + X2 X1 = 0:
Remark 13.2 It is clear from Example 13.1 that the condition i + j 6= 0 is not
necessary for the existence of a hermitian solution.
~
The following theorem gives necessary and sucient conditions for a solution to be
real.
"
"
X1 = X1 P
X2
X2
where the over bar X denotes the complex conjugate. Therefore, X = X2 X1;1 =
X2 P (X1 P );1 = X2 X1;1 = X is real as desired.
()) Dene X := X2 X1;1. By assumption, X 2 Rnn and
"
I = V;
Im
X
therefore, V is conjugate symmetric.
Example 13.2 This example is intended to show that there are non-real, non-hermitian
solutions to equation (13.1). It is also designed to show that the condition i + j 6=
0; 8 i; j , which excludes the possibility of having imaginary axis eigenvalues since if
l = j! then l + l = 0, is not necessary for the existence of a hermitian solution. Let
2
A = 64
1 0 0
;1 0 ;2
0 0 0
7
6
7
6
0 0 ;1 5 ; R = 4 0 0 0 5 ; Q = 4 0 0 0 75 :
0 1 0
;2 0 ;4
0 0 0
325
X = 64
2 0 0
0 0 0
0 0 0
3
7
5
X; (H ) = Im X1
X2
X; (H ); Im
"
(13:5)
326
property, respectively. This solution will be called the stabilizing solution. Thus, X =
Ric(H ) and
Ric : dom(Ric) R2n2n 7;! Rnn :
The following well-known results give some properties of X as well as veriable conditions under which H belongs to dom(Ric).
A X + XA + XRX + Q = 0;
(iii) A + RX is stable .
X1X2 is symmetric:
To prove this, note that there exists a stable matrix H; in Rnn such that
"
"
(13:6)
H X1 = X1 H; :
X2
X2
(H; is a matrix representation of H jX; (H ) .) Pre-multiply this equation by
"
to get
"
X1
X2
#
"
X1
X2
"
#
JH X1 = X1
X2
X2
#
"
J X1 H; :
X2
(13:7)
Since JH is symmetric, so is the left-hand side of (13.7) and so is the right-hand side:
(;X1X2 + X2X1 )H; = H; (;X1 X2 + X2 X1 )
;X1X2 + X2 X1 = 0:
327
"
H X1 = X1 H;
X2
X2
and post-multiply by X1;1 to get
"
"
H I = I X1 H; X1;1:
X
X
Now pre-multiply by [X ; I ]:
"
(13:8)
I = 0:
[X ; I ]H
X
This is precisely the Riccati equation.
(iii) Pre-multiply (13.8) by [I 0] to get
A + RX = X1 H; X1;1:
Now, we are going to state one of the main theorems of this section which gives the
necessary and sucient conditions for the existence of a unique stabilizing solution of
(13.1) under certain restrictions on the matrix R.
Theorem 13.6 Suppose H has no imaginary eigenvalues and R is either positive semidenite or negative semi-denite. Then H 2 dom(Ric) if and only if (A; R) is stabilizable.
"
are complementary. This requires a preliminary step. As in the proof of Theorem 13.5
dene X1 ; X2 ; H; so that
"
#
X; (H ) = Im X1
"
"
X2
#
H X1 = X1 H; :
X2
X2
(13:9)
328
We want to show that X1 is nonsingular, i.e., Ker X1 = 0. First, it is claimed that Ker
X1 is H; -invariant. To prove this, let x 2 Ker X1 . Pre-multiply (13.9) by [I 0] to get
AX1 + RX2 = X1 H; :
(13:10)
Pre-multiply by x X2 , post-multiply by x, and use the fact that X2 X1 is symmetric
(see (13.6)) to get
x X2RX2 x = 0:
Since R is semidenite, this implies that RX2 x = 0. Now post-multiply (13.10) by x to
get X1 H; x = 0, i.e. H; x 2 Ker X1 . This proves the claim.
Now to prove that X1 is nonsingular, suppose, on the contrary, that Ker X1 6= 0.
Then H; jKer X1 has an eigenvalue, , and a corresponding eigenvector, x:
H; x = x
(13:11)
Re < 0; 0 6= x 2 Ker X1 :
Pre-multiply (13.9) by [0 I ]:
; QX1 ; A X2 = X2 H; :
(13:12)
Post-multiply the above equation by x and use (13.11):
(A + I )X2 x = 0:
Recall that RX2 x = 0, we have
x X2 [A + I R] = 0:
Then stabilizability
"
# of (A; R) implies X2 x = 0. But if both X1 x = 0 and X2 x = 0, then
A ;BB :
H=
;C C ;A
Then H 2 dom(Ric) i (A; B ) is stabilizable and (C; A) has no unobservable modes on
the imaginary axis. Furthermore, X = Ric(H ) 0 if H 2 dom(Ric), and Ker(X ) = 0
if and only if (C; A) has no stable unobservable modes.
Note that Ker(X ) Ker(C ), so that the equation XM = C always has a solution
for M , and a minimum F -norm solution is given by X yC .
329
Proof. It is clear from Theorem 13.6 that the stabilizability of (A; B) is necessary,
and it is also sucient if H has no eigenvalues on the imaginary axis. So we only need
to show that, assuming (A; B ) is stabilizable, H has no imaginary eigenvalues i (C; A)
has no unobservable
modes on the imaginary axis. Suppose that j! is an eigenvalue
"
#
x
and 0 6=
is a corresponding eigenvector. Then
Ax ; BB z = j!x
;C Cx ; A z = j!z:
Re-arrange:
(A ; j!I )x = BB z
; (A ; j!I ) z = C Cx:
Thus
(13:13)
(13:14)
z [A ; j!I B ] = 0
"
A ; j!I x = 0:
C
A X + XA ; XBB X + C C = 0
or equivalently
(A ; BB X ) X + X (A ; BB X ) + XBB X + C C = 0:
(13:15)
330
X=
e(A;BB X ) t (XBB X + C C )e(A;BB X )t dt:
(13:16)
XAx = 0:
We conclude that Ker(X ) is an A-invariant subspace. Now if Ker(X ) 6= 0, then there
is a 0 6= x 2 Ker(X ) and a such that x = Ax = (A ; BB X )x and Cx = 0. Since
(A ; BB X ) is stable, Re < 0; thus is a stable unobservable mode. Conversely,
suppose (C; A) has an unobservable stable mode , i.e., there is an x such that Ax =
x; Cx = 0. By pre-multiplying the Riccati equation by x and post-multiplying by x,
we get
2Rex Xx ; x XBB Xx = 0:
Hence x Xx = 0, i.e., X is singular.
2
Example 13.3 This example shows that the observability of (C; A) is not necessary
for the existence of a positive denite stabilizing solution. Let
"
"
h
i
A= 1 0 ; B= 1 ; C= 0 0 :
0 2
1
X = 18 ;24 > 0
;24 36
Corollary 13.8 Suppose that (A; B) is stabilizable and (C; A) is detectable. Then the
Riccati equation
A X + XA ; XBB X + C C = 0
has a unique positive semidenite solution. Moreover, the solution is stabilizing.
331
Proof. It is obvious from the above theorem that the Riccati equation has a unique
stabilizing solution and that the solution is positive semidenite. Hence we only need
to show that any positive semidenite solution X 0 must also be stabilizing. Then
by the uniqueness of the stabilizing solution, we can conclude that there is only one
positive semidenite solution. To achieve that goal, let us assume that X 0 satises
the Riccati equation but that it is not stabilizing. First rewrite the Riccati equation as
(A ; BB X ) X + X (A ; BB X ) + XBB X + C C = 0
(13:17)
Ax = x; Cx = 0
i.e., (C; A) is not detectable, which is a contradiction. Hence Re() < 0, i.e., X 0 is
the stabilizing solution.
2
Lemma 13.9 Suppose D has full column rank and let R = DD > 0; then the following
statements are equivalent:
"
(i)
A ; j!I B
C
D
"
A ; j!I B
C
D
#"
;R;1D C I
#"
x = 0:
0
332
But this implies that
"
A ; j!I B
C
D
(13:18)
u 6= 0 such that
v
Now let
"
Then
"
and
A ; j!I B
C
D
#"
"
u = 0:
v
u =
I
0
v
;R;1D C I
#
"
x =
I
0
y
R;1D C I
#"
#"
x :
y
#
u 6= 0
v
(A ; BR;1 D C ; j!I )x + By = 0
(I ; DR;1D )Cx + Dy = 0:
Pre-multiply (13.20) by D to get y = 0. Then we have
(A ; BR;1 D C )x = j!x; (I ; DR;1 D )Cx = 0
;
i.e., j! is an unobservable mode of (I ; DR;1 D )C; A ; BR;1 D C .
(13:19)
(13:20)
H =
"
"
"
h
A
0
B
;1 D C B
;
R
;C C ;A
;C D
#
A ; BR;1 D C
;BR;1 B
:
;C (I ; DR;1 D )C ;(A ; BR;1 D C )
333
#
Proof. This is the consequence of the Lemma 13.9 and Theorem 13.7.
BR;1 D C ) implies the detectability (observability) of (C; A); however, the converse is
in general not true. Hence the existence of a stabilizing solution to the Riccati equation
in the above corollary is not guaranteed by the stabilizability of (A; B ) and detectability
of (C; A). Furthermore, even if a stabilizing solution exists, the positive deniteness of
the solution is not guaranteed by the observability of (C; A) unless D C = 0. As an
example, consider
"
#
"
#
"
#
"
#
0 1
0
1 0
1
A=
; B=
; C=
; D=
:
0 0
;1
0 0
0
Then (C; A) is observable, (A; B ) is controllable, and
"
#
h
i
0 1
A ; BD C =
; D? C = 0 0 :
1 0
A Riccati equation with the above data has a nonnegative denite stabilizing solution since (D? C; A ; BR;1 D C ) has no unobservable modes on the imaginary axis.
However, the solution is not positive denite since (D? C; A ; BR;1 D C ) has a stable
unobservable mode. On the other hand, if the B matrix is changed to
"
#
0
B=
;
1
then the corresponding Riccati equation has no stabilizing solution since, in this case,
(A ; BD C ) has eigenvalues on the imaginary axis although (A; B ) is controllable and
(C; A) is observable.
~
334
A = 64
1 0 0
1
h
i
7
6
0 2 0 5 ; B = 4 1 75 ; C = 0 0 0 :
0 0 ;3
1
"
A ;BB are
The eigenvalues of matrix H =
;C C ;A
1 = 1; 2 = ;1; 3 = 2; 4 = ;2; 5 = ;3; 6 = 3;
and their corresponding eigenvectors are
3
2
3
2
3
2
3
2
3
2
3
2
;
3
0
4
0
3
1
7
6
7
6
7
6
7
6
7
6
7
6
6 ;6 7
6 0 7
6 3 7
6 1 7
6 2 7
6 0 7
7
6
7
6
7
6
7
6
7
6
7
6
7
6
7
6
7
6
6 0 7
6 ;3 7
6 0 7
6 ;1 7
6 1 7
6 ;12 7
7
6
7
6
7
6
v1 = 66 77 ; v2 = 66
7:
7 ; v6 = 6
7 ; v5 = 6
7 ; v4 = 6
7 ; v3 = 6
6 0 7
6 0 7
6 0 7
6 0 7
7
7
6
7
6
7
6
7
6
6 6 7
6 0 7
7
6
7
6
7
6
7
6
6 0 7
6 0 7
4 0 5
4 0 5
4 12 5
4 0 5
5
4
5
4
6
0
0
0
0
0
There are four distinct nonnegative denite symmetric solutions depending on the chosen invariant subspaces:
2
3
"
#
0 0 0
X1 = [v1 v3 v5 ]; Y1 = X2 X1;1 = 64 0 0 0 75;
X2
0 0 0
"
"
2 0 0
0 0 0 75;
0 0 0
0 0 0
0 4 0 75;
0 0 0
X1 = [v v v ]; Y = X X ;1 = 6
4
2 3 5 2
2 1
X2
#
X1 = [v v v ]; Y = X X ;1 = 6
4
1 4 5 3
2 1
X2
2
18 ;24 0
X
1
6
;
1
= [v2 v4 v5 ]; Y4 = X2 X1 = 4 ;24 36 0 75.
X2
0
0 0
These solutions can be ordered as Y4 Yi Y1 ; i = 2; 3. Of course, this is only a partial
ordering since Y2 and Y3 are not comparable. Note also that only Y4 is a stabilizing
solution, i.e., A ; BB Y4 is stable. Furthermore, Y4 and Y1 are the \maximal" and
\minimal" solutions, respectively.
3
"
335
The partial ordering concept shown in Example 13.4 can be stated in a much more
general setting. To do that, again consider the Riccati equation (13.1). We shall call
a hermitian solution X+ of (13.1) a maximal solution if X+ X for all hermitian
solutions X of (13.1). Similarly, we shall call a hermitian solution X; of (13.1) a
minimal solution if X; X for all hermitian solutions X of (13.1). Clearly, maximal
and minimal solutions are unique if they exist.
To study the properties of the maximal and minimal solutions, we shall introduce
the following quadratic matrix:
Q(X ) := A X + XA + XRX + Q:
(13:21)
Theorem 13.11 Assume R 0 and assume there is a hermitian matrix X = X such
that Q(X ) 0.
(i) If (A; R) is stabilizable, then there exists a unique maximal solution X+ to the
Riccati equation (13.1). Furthermore,
X+ ; X; =
=
Z
Z
;1
;1
e(A+RX+ )t Re(A+RX+ ) t dt
;1
(iv) If Q(X ) > 0, the results in (i) and (ii) can be respectively strengthened to X+ > X ,
(A + RX+ ) C ; , and X; < X , (A + RX; ) C + .
Proof. Let R = ;BB for some B. Note the fact that (A; R) is stabilizable (controllable) i (A; B ) is.
(i): Let X be such that Q(X ) 0. Since (A; B ) is stabilizable, there is an F0 such
that
A0 := A + BF0
336
X0 A0 + A0 X0 + F0 F0 + Q = 0:
Then X0 is hermitian. Dene
F^0 := F0 + B X;
X0 X:
X0 X1 Xn;1 X;
Next, introduce
Ai = A + BFi ; is stable; i = 0; : : : ; n ; 1;
Fi = ;B Xi;1 ; i = 1; : : : ; n ; 1;
Xi Ai + Ai Xi = ;FiFi ; Q; i = 0; 1; : : : ; n ; 1:
(13:22)
Fn = ;B Xn;1 ;
An = A + BFn :
(13:23)
F^n := Fn + B X ;
then
(Xn;1 ; X )An + An (Xn;1 ; X ) = ;Q(X ) ; F^nF^n ; (Fn ; Fn;1 ) (Fn ; Fn;1 ): (13:24)
Now assume that An is not stable, i.e., there exists an with Re 0 and x 6= 0 such
that An x = x. Then pre-multiply (13.24) by x and postmultiply by x, and we have
2Rex (Xn;1 ; X )x = ;x fQ(X ) + F^n F^n + (Fn ; Fn;1 ) (Fn ; Fn;1 )gx:
337
Since it is assumed Xn;1 X , each term on the right-hand side of the above equation
has to be zero. So we have
(Fn ; Fn;1 )x = 0:
But now
Xn;1 Xn X:
We have a non-increasing sequence fXi g, and the sequence is bounded below by Xi X .
Hence the limit
Xf := nlim
!1 Xn
exists and is hermitian, and we have Xf X . Passing the limit n ! 1 in (13.25), we
get Q(Xf ) = 0. So Xf is a solution of (13.1). Since X is an arbitrary element satisfying
Q(X ) 0 and Xf is independent of the choice of X , we have
Xf X; 8 X such that Q(X ) 0:
In particular, Xf is the maximal solution of the Riccati equation (13.1), i.e., Xf = X+ .
To establish the stability property of the maximal solution, note that An is stable
for any n. Hence, in the limit, the eigenvalues of
A ; BB Xf
will have non-positive real parts. The uniqueness follows from the fact that the maximal
solution is unique.
(ii): The results follow by the following substitutions in the proof of part (i):
;A; X
;X ; X+
;X; :
338
(iii): The existence of X+ and X; follows from (i) and (ii). Let A+ := A + RX+ ;
we now show that X+ ; X; > 0 i (A+ ) C ; . It is easy to verify that
X+ ; X = ;
e(A+RX+ )t Re(A+RX+ ) t dt
;1
> 0:
(13:26)
Remark 13.5 The proof given above also gives an iterative procedure to compute
the maximal and minimal solutions. For example, to nd the maximal solution, the
following procedures can be used:
(i) nd F0 such that A0 = A + BF0 is stable;
(ii) solve Xi : Xi Ai + Ai Xi + Fi Fi + Q = 0;
(iii) if kXi ; Xi;1 k =specied accuracy, stop. Otherwise go to (iv);
(iv) let Fi+1 = ;B Xi and Ai+1 = A + BFi+1 go to (ii).
This procedure will converge to the stabilizing solution if the solution exists.
~
339
Corollary 13.12 Let R 0 and suppose (A; R) is controllable and X1, X2 are two
solutions to the Riccati equation (13.1). Then X1 > X2 implies that X+ = X1 , X; =
X2 , (A + RX1) C ; and that (A + RX2 ) C + .
The following example illustrates that the stabilizability of (A; R) is not sucient
to guarantee the existence of a minimal hermitian solution of equation (13.1).
"
"
"
A = 0 0 ; R = ;1 0 ; Q = 1 0 :
0 ;1
0 0
0 0
Then it can be shown that all the hermitian solutions of (13.1) are given by
"
1 0
;
0 0
"
;1
; 2 C:
; 12 jj2
"
X+ = 1 0 ;
0 0
however, there is no minimal solution.
The Riccati equation appeared in H1 control, which will be considered in the later
part of this book, often has R 0. However, these conditions are only a dual of the
above theorem.
340
X+ ; X; =
=
Z
Z
;1
;1
;1
e(A+RX+ )t Re(A+RX+ ) t dt
(iv) If Q(X ) < 0, the results in (i) and (ii) can be respectively strengthened to X+ > X ,
(A + RX+ ) C + , and X; < X , (A + RX; ) C ; .
Theorem 13.11 can be used to derive some comparative results for some Riccati
equations. More specically, let
and
"
"
;1
;BRs;1 B
Hs := A ; BRs;1S
;P + SRs S ;(A ; BRs;1 S )
~ ~ ~ ;1 ~
;B~ R~s;1 B~
H~ s := A~; B~R~s;1S~
;P + S Rs S ;(A~ ; B~ R~s;1S~ )
where P; P~ ; Rs , and R~s are real symmetric and Rs > 0, R~s > 0. We shall also make use
of the following matrices:
"
"
~ ~
T := P S ; T~ := ~P ~S :
S Rs
S Rs
We denote by X+ and X~+ the maximal solution to the Riccati equation associated with
Hs and H~ s , respectively:
(A ; BRs;1 S ) X + X (A ; BRs;1 S ) ; XBRs;1B X + (P ; SRs;1S ) = 0 (13:27)
(A~ ; B~ R~s;1 S~ ) X~ + X~ (A~ ; B~ R~s;1 S~) ; X~ B~ R~s;1B~ X~ + (P~ ; S~R~s;1 S~) = 0: (13:28)
Recall that JHs and J H~ s are hermitian where J is dened as
"
#
0 ;I
J=
:
I 0
Let
"
;1
;1
K := JHs = P ; SRs;1S (A ; BR;s1 S )
A ; BRs S
;BRs B
"
#
~ ; S~R~s;1 S~ (A~ ; B~ R~s;1S~ )
P
~
~
K := J Hs = ~ ~ ~ ;1 ~
:
A ; B Rs S
;B~ R~s;1 B~
341
(i) Assume that (13.28) has a hermitian solution and that K K~ (K > K~ ); then
X+ and X~+ exist, and X+ X~+ (X+ > X~+ ).
(ii) Let A = A~, B = B~ , and T T~ (T > T~). Assume that (13.28) has a hermitian
solution. Then X+ and X~+ exist, and X+ X~+ (X+ > X~+ ).
(iii) If T 0 (T > 0), then X+ exists, and X+ 0 (X+ > 0).
Qs (X ) := I
X
#
I
X
"
#
"
K~ I = 0:
X
#
"
K I = I
X
X
#
"
I 0:
(K ; K~ )
X
Now we use Theorem 13.11 to obtain the existence of X+ , and X+ X . Since X is
arbitrary, we get X+ X for all hermitian solutions X of (13.28). The existence of X~+
Qs (X ) = I
X
while (13.28) becomes
It is easy to show that
#
"
K I = A X + XA ; LRs L + P
X
A X + XA ; L~ R~s L~ + P~ = 0:
"
I
X
#
"
I
Qs (X ) =
(K ; K~ )
X
= P" ; P~# ; L Rs L "+ L~ #R~s L~
I
I + (L ; L~ ) R~ (L ; L~ ) 0:
=
(T ; T~)
s
L
L
Now as in part (i), there exist X+ and X~+ , and X+ X~+ .
(iii): The condition T 0 implies P ; SRs;1S 0, so we have Qs (0) = P ;
;
SRs 1 S 0. Apply Theorem 13.11 to get the existence of X+ , and X+ 0.
2
342
Lemma 13.15 Suppose R is nonsingular and either one of the following assumptions
is satised:
;1
;BR;1B
H = A ; BR ;1S
;(P ; SR S ) ;(A ; BR;1 S )
A^
(s) = ^
C
#
A 0
B^ := 6 ;P ;A
4
D^
S B
Then H = A^ ; B^ D^ ;1 C^ and
2
6
;1 (s) = 64
H
R;1S R;1 B
"
i
B
;S
R
;BR;1
SR;1
R;1
3
7
5
# 3
7
7
5
If (j!0 ) is singular, then j!0 is a zero of (s). Hence j!0 is a pole of ;1 (s), and j!0
is an eigenvalue of H .
343
(ii) ) (i): Suppose j!0 is an eigenvalue ofhH but is not a polei of ;1 (s). Then
j!0 must be either an unobservable mode of ( R;1S R;1 B ; H ) or an uncon"
"
x 6 0 such that
( R;1 S R;1 B ; H ). Then there exists an x0 = 1 =
x2
h
Hx0 = j!0 x0 ;
R;1 S R;1 B x0 = 0:
(13:30)
(13:31)
S x1 + B x2 = 0:
(13:32)
SR
Corollary 13.16 Suppose R > 0 and either one of the assumptions (A1) or (A2)
dened in Lemma 13.15 is true. Then the following statements are equivalent:
(i) (j!) > 0 for all 0 ! 1.
(ii) The Hamiltonian matrix
"
;1
;BR;1B
H = A ; BR ;1S
;(P ; SR S ) ;(A ; BR;1 S )
344
Proof. (i) ) (ii) follows easily from Lemma 13.15. To prove (ii) ) (i), note that
(j 1) = R > 0 and det (j!) 6= 0 for all ! from Lemma 13.15. Then the continuity of
2
"
P 0
0 I
#"
(sI ; A);1 B
satises
Proof. (i): ()) From Corollary 13.16, 0(j!) > 0 implies that a Hamiltonian matrix
"
#
A ;BB
;P ;A
has no eigenvalues on the imaginary axis. This in turn implies from Theorem 13.6 that
(13.33) has a stabilizing solution. Furthermore, since
"
we have
"
A ;BB = ;I
;P ;A
I
"
#"
A BB
P ;A
"
#"
;I
Ric A ;BB = ;Ric A BB 0:
;P ;A
P ;A
(() The suciency proof is omitted here and is a special case of (b) ) (a) of
Theorem 13.19 below.
(ii): ()) Let P = ;C C and G(s) := C (sI ; A);1 B . Then 0 (s) = I ; G (s)G(s).
Let 0 <
< 1 and dene
(s) := I ;
2 G (s)G(s):
345
Then
(j!) > 0; 8!. Thus from part (i), there is an X
= X
0 such that
A ; BB X
is stable and
A X
+ X
A ; X
BB X
;
2 C C = 0:
It is easy to see from Theorem 13.14 that X
is monotone-decreasing with
, i.e.,
X
1 X
2 if
1
2 . To show that lim
!1 X
exists, we need to show that X
is
bounded below for all 0 <
< 1.
In the following, it will be assumed that (A; B ) is controllable. The controllability
assumption will be removed later.
Let Y be the destabilizing solution to the following Riccati equation:
A Y + Y A ; Y BB Y = 0
with (A ; BB Y ) C + (note that the existence of such a solution is guaranteed by
the controllability assumption). Then it is easy to verify that
(A ; BB Y ) (X
; Y ) + (X
; Y )(A ; BB Y ) ; (X
; Y )BB (X
; Y ) ;
2C C = 0:
This implies that
X ; Y =
Thus X
is bounded below with Y as the lower bound, and lim
!1 X
exists. Let
X := lim
!1 X
; then from continuity argument, X satises the Riccati equation
A X + XA ; XBB X + P = 0
and (A ; BB X ) C ; .
Now suppose (A; B ) is not controllable, and then assume without loss of generality
that
"
#
"
#
h
i
A
B
11 A12
1
A=
; B=
; C = C1 C2
0 A22
0
so that (A11 ; B1 ) is controllable and A11 and A22 are stable. Then the Riccati equation
for
"
#
X
X
11
12
X
=
) X
(X12
22
can be written as three equations
346
and
A ; BB X =
"
is stable.
Let Y11 be the anti-stabilizing solution to
A22 X22 + X22 A22 + X12 A12 + A12 X12 ; X12 B1 B1 X12 ; C2 C2 = 0:
We have proven that there exists a unique X such that
A X + XA ; XBB X ; C C = 0
and (A ; BB X ) C ; .
(() same as in part (i).
"
"
P S
S R
P^ 0
0 I
#"
(sI ; A);1 B
#"
(sI ; A^);1 B^
"
P S = I SR;1=2
S R
0 R1=2
347
#"
P^ 0
0 I
#"
R;1=2 S R1=2
"
P^ 0
0 I
#"
(sI ; A);1 B
'(s)
Now we are ready to state and prove one of the main results of this section. The
following theorem and corollary characterize the relations among spectral factorizations,
Riccati equations, and decomposition of Hamiltonians.
"
P S
S R
#"
(sI ; A);1 B
satises
348
(c) The Hamiltonian matrix
"
;1
;BR;1B
H = A ; BR ;1S
;(P ; SR S ) ;(A ; BR;1 S )
Corollary 13.20 If either one of the conditions, (a){(e), in Theorem 13.19 is satised,
then there exists an M 2 Rp such that
= M RM:
A particular realization of one such M is
"
M= A B
;F I
Remark 13.6 If the stabilizability of (A; B) is changed into the stabilizability of (;A; B),
then the theorem still holds except that the solutions X in (b) and (e) are changed into
the destabilizing solution ((A ; BR;1 S ; BR;1 B X ) C + ) and the weakly destabilizing solution ((A ; BR;1 S ; BR;1 B X ) C + ), respectively.
~
Proof. (a) ) (c) follows from Corollary 13.16.
(c) ) (b) follows from Theorem 13.6 and Theorem 13.5.
(b) ) (a) Suppose 9X = X such that A ; BR;1S ; BR;1 B X = A ; BR;1 (S +
B X ) is stable. Let F = ;R;1(S + B X ) and
"
M= A B :
;F I
349
It is easily veried by use of the Riccati equation for X and by routine algebra that
= M RM . Since
"
#
A
+
BF
B
M ;1 =
;
M ;1 2 RH1 . Thus M (s) has no zeros on the imaginary axis and (j!) > 0.
(e) ) (d) follows the same procedure as the proof of (b) ) (a).
(d) ) (e): Assume S = 0 and R = I ; otherwise use Lemma 13.18 rst to get a
new function with such properties. Let P = C1 C1 ; C2 C2 with C1 and C2 square
nonsingular. Note that this decomposition always exists since P = I ; (I ; P ), with
> 0 suciently large, denes one such possibility. Let X1 be the positive denite
solution to
A X1 + X1 A ; X1 BB X1 + C1 C1 = 0:
By Theorem 13.7, X1 indeed exists and is stabilizing, i.e., A1 := A ; BB X1 is stable.
Let = X ; X1 . Then the equation in becomes
A1 + A1 ; BB ; C2 C2 = 0:
To show that this equation has a solution, recall Lemma 13.17 and note that A1 is
stable; then it is sucient to show that
C2 (sI ; A + BB X1 );1 B = C2 (sI ; A);1 B I + B X1 (sI ; A);1 B ;1 :
= I + B X1 (sI ; A);1 B I + B X1 (sI ; A);1 B :
It follows that
n
o
I ; I + B (;j!I ; A );1 X1 B ;1 B (;j!I ; A );1 C2
n
o
C2 (j!I ; A);1 B I + B X1 (j!I ; A);1 B ;1 0:
350
Consequently,
We may now apply Lemma 13.17 to the equation. Consequently, there exists a unique
solution such that (A1 ; BB ) = (A ; BB X ) C ; . This shows the existence
and uniqueness of X .
We shall now illustrate the proceeding results through a simple example. Note in
particular that the function can have poles on the imaginary axis.
H = 0 ;1
;1 0
does not have eigenvalues on the imaginary axis and X = 1 is the stabilizing solution
to the corresponding Riccati equation and the spectral factor is given by
#
"
M (s) = 0 1 = s +s 1 :
1 1
3
Some frequently used special spectral factorizations are now considered.
"
A B
C D
B
A
M (s) =
1
=
2
;R F R1=2
where
R =
2 I ; D D
F = R;1"(B X + D C )
A + BR;1 D C
BR;1B
X = Ric
;C (I + DR;1 D )C ;(A + BR;1 D C )
and X 0 if A is stable.
351
Proof. This is a special case of Theorem 13.19. In fact, the theorem follows by letting
P = ;C C; S = ;C D; R =
2I ; D D in Theorem 13.19 and by using the fact that
"
A + BR;1 D C
BR;1 B
Ric
=
;
1
;C (I + DR D )C ;(A + BR;1 D C )
"
;1
;BR;1 B
;Ric A + BR ;D1 C
:
C (I + DR D )C ;(A + BR;1 D C )
The spectral factorization for the dual case is also often used and follows by taking
the transpose of the corresponding transfer matrices.
"
Corollary 13.22 Assume G(s) := A B 2 RL1 and
> kG(s)k1. Then, there
C D
exists a transfer matrix M 2 RL1 such that MM =
2 I ; GG and M ;1 2 RH1 .
A particular realization of M is
"
1=2
M (s) = A ;LR1=2
C R
where
R =
2 I ; DD
L = (Y C" + BD )R;1
;1
C R;1 C
Y = Ric (A + BD R;1 C )
;B (I + D R D)B ;(A + BD R;1 C )
and Y 0 if A is stable.
For convenience, we also include the following spectral factorization results which
are again special cases of Theorem 13.19.
"
"
"
A ; j! B
C
D
A ; BR;1 D C
;BR;1 B
X = Ric
;C (I ; DR;1D )C ;(A ; BR;1 D C )
352
W W = G G
where W ;1 2 RH1 and
"
B :
W = ;1=2 A
R (D C + B X ) R1=2
(b) Suppose G(j!)G (j!) > 0 for all ! or
!. Let
"
A ; j! B
C
D
"
~ ;1
;C R~ ;1 C
Y = Ric (A ; BD R~ ;1 C )
;B (I ; D R D)B ;(A ; BD R~ ;1 C )
"
~ ;1=2
:
W~ = A (BD + ~Y1C=2 )R
R
C
A + BR;1 D C
BR;1 B
H :=
;C (I + DR;1D )C ;(A + BR;1 D C )
353
Proof. This follows from the fact that kGk1 <
is equivalent to that the following
function is positive denite for all !:
(j!) :=
2 I ; GT (;j!)G(j!)
h
= B (;j!I ; A );1 I
"
;C C ;C D
;DC
2 I ; D D
#"
(j!I ; A);1 B
> 0;
"
;I 0 H ;I 0 =
0
"
A + BR;1D C
;BR;1 B
:
C (I + DR;1 D )C ;(A + BR;1 D C )
2
The equivalence between (i) and (iv) in the above corollary is usually referred as
Bounded Real Lemma.
Theorem 13.25 Let A B be a state space realization of G(s) with A stable (not
C D
necessarily a minimal realization). Suppose there exist an X 0, Q, and W such that
XA + A X = ;QQ
BX + W Q = C
D + D = W W;
(13.34)
(13.35)
(13.36)
354
Proof.
B
6
G(s)+G (s) = 4 0 ;A ;C
C B D + D
0
7
5
= 64
"
0
;A ;(XB + Q W ) 75 :
BX + W Q B
W W
G(s) + G (s) =
6
4
"
XA + A X ;A
W Q
B
#"
A
;A ;Q
B W
Q
B
;QW
W W
#
B :
W
B
A
0
7 6
;QW
5 = 4 ;Q Q ;A
W Q B W W
3
7
5
and
"
Proof
. Since
# G(s) is assumed to be positive real, there exists a transfer matrix M (s) =
"
A1 B1
C1 D1
355
where A and A1 have the same dimensions. Now let X1 0 be the solution of the
following Lyapunov equation:
X1 A1 + A1 X1 = ;C1 C1 :
Then
2
3
#"
#
"
B1
A1
0
A1 B1 = 6
7
M (s)M (s) = ;A1 ;C1
4 ;C1 C1 ;A1 ;C1 D1 5
B1
D1
C1 D1
D1 C1
B1
D1 D1
A1
0
B1
X1 A1 + A1 X1 ;A1 ;C1 D1 75
D1 C1
B1 D1 D1
2
3
A1
0
B1
= 64
0
;A1 ;(X1 B1 + C1 D1 ) 75
B1 X + D1 C1 B1
D1 D1
"
# "
1 ;(B1 X + D1 C1 ) #
A
;
A
B
1
1
= D1 D1 +
+
:
B1 X + D1 C1 0
B1
0
But the realization for G(s) + G (s) is given by
# "
#
"
;C
;
A
A
B
+
:
G(s) + G (s) = D + D +
C 0
B 0
Since the realization for G(s) is minimal, there exists a nonsingular matrix T such that
A = TA1T ;1; B = TB1 ; C = (B1 X + D1 C1 )T ;1; D + D = D1 D1 :
=
6
4
356
Proof. This follows from Theorem 13.19 and from the fact that
G(j!) + G (j!) > 0
for all ! including 1.
The above corollary also leads to the following special spectral factorization.
"
Corollary 13.28 Let G(s) = A B 2 RH1 with D full row rank and G(j!)G (j!) >
C D
0 for all !. Let P be the controllability grammian of (A; B ):
PA + AP + BB = 0:
Dene
BW = PC + BD :
A BW
where M (s) =
CW DW
DW DW = DD
and
Proof. This corollary follows from Corollary 13.27 and the fact that G(j!)G (j!) > 0
and
G(s)G (s) =
"
A BW
C 0
"
;A ;C + DD :
+
BW
357
(13:37)
Then
(a) D C + B X = 0 implies N N = D D.
(b) (A; B ) controllable, and N N = D D implies D C + B X = 0.
B
A
0
6
6
N N = 4 ;C C ;A ;C D
D C B D D
"
"
N N =
6
6
4
2
6
6
4
#;1
"
3
7
7
5
A
0
B
;(A X + XA + C C ) ;A ;(XB + C D)
B X + D C
B
D D
3
A
0
B
7
0
;A ;(XB + C D) 75 :
B X + D C B
D D
3
7
7
5
358
"
A B
C D
y
N? = A ;X C D?
C
D?
359
G = A B is stabilizable and that A ; j!I B has full column rank for all
C
D
C D
! 2 R. Then a particular realization of the desired coprime factorization is
"
M := 6
4
N
where
and
A + BF
F
C + DF
BR;1=2
R;1=2
DR;1=2
3
7
5
2 RH1
R = D D > 0
F = ;R;1(B X + D C )
"
A ; BR;1 D C
;BR;1B
X = Ric
0:
;C (I ; DR;1 D )C ;(A ; BR;1 D C )
y
N? = A + BF ;X C D?
C + DF
D?
if p > m.
N = A + BF BZr
C + DF DZr
(13:38)
(BZr ) X + (DZr ) (C + DF ) = 0
(13:39)
360
(A + BF ) X + X (A + BF ) + (C + DF ) (C + DF ) = 0:
(13:40)
;
1
=
2
Clearly, we have that Zr = R U where R = D D > 0 and where U is any orthogonal
matrix. Take U = I and solve (13.39) for F to get
F = ;R;1(B X + D C ):
Then substitute F into (13.40) to get
0 = (A + BF ) X + X (A + BF ) + (C + DF ) (C + DF )
= (A ; BR;1 D C ) X + X (A ; BR;1 D C ) ; XBR;1B X + C D?D? C
where D? D? = I ; DR;1 D . To show that such choices indeed make sense, we need
to show that H 2 dom(Ric), where
"
;1 D C
;BR;1B
H = A ; BR
;C D?D? C ;(A ; BR;1 D C )
so X "= Ric(H ). However,
by Theorem 13.19, H 2 dom(Ric) is guaranteed by the fact
#
A
;
j!
B
that
has full column rank (or G (j!)G(j!) > 0.
C
D
The uniqueness of the factorization follows from coprimeness and N inner. Suppose
that G = N1M1;1 = N2 M2;1 are two right coprime factorizations and that both nu-
merators are inner. By coprimeness, these two factorizations are unique up to a right
2 in RH1 . That is, there exists a unit 2 RH1 such that
multiple
"
# which" is a unit
#
M1 = M2 . Clearly, is inner since = N N = N N = I .
1 1
2 2
N1
N2
The only inner units in RH1 are constant matrices, and thus the desired uniqueness
property is established. Note that the non-uniqueness is contained entirely in the choice
of a particular square root of R.
Finally, the formula for N? follows from Lemma 13.31.
2
Note that the important inner-outer factorization formula can be obtained from this
inner numerator coprime factorization if G 2 RH1 .
Corollary 13.33 Suppose G 2 RH1 ; then the denominator matrix M in Theorem 13.32
is an outer. Hence, the factorization G = N (M ;1) given in Theorem 13.32 is an innerouter factorization.
Remark 13.7 It is noted that the above inner-outer factorization procedure does not
apply to the strictly proper transfer matrix even if the factorization exists. For example,
;1 1 has inner-outer factorizations but the above procedure cannot be used.
G(s) = ss+1
s+2
The inner-outer factorization for the general transfer matrices can be done using the
method adopted in Section 6.1 of Chapter 6.
~
2 A function is called a unit in RH if ; ;1 2 RH .
1
1
361
Suppose that the system G is not stable; then a coprime factorization with an inner
denominator can also be obtained by solving a special Riccati equation. The proof of
this result is similar to the inner numerator case and is omitted.
"
F" = ;B X #
X = Ric A ;BB 0:
0 ;A
Dual results can be obtained when p m by taking the transpose of the transfer function
matrix. In these factorizations, output injection using the dual Riccati solution replaces
state feedback to obtain the corresponding left factorizations.
Theorem 13.35 Assume p m. Then there exists an lcf G = M~ ;1N~ such that N~ is
a co-inner if and only if GG > 0 on the j!-axis, including at 1. This factorization is
unique
Furthermore,
"
# assume that the realization of
" up to a#constant unitary multiple.
G = A B is detectable and that A ; j!I B has full row rank for all ! 2 R.
C
D
C D
Then a particular realization of the desired coprime factorization is
"
#
h
i
A
+
LC
L
B
+
LD
2 RH1
M~ N~ := ~ ;1=2 ~ ;1=2 ~ ;1=2
R C R
R D
where
and
R~ = DD > 0
L = ;(BD + Y C )R~;1
"
~ ;1 ~
;C R~ ;1 C
Y = Ric (A ; BD R~ ;1C )
0:
;B (I ; D R D)B ;(A ; BD R~ ;1 C )
Moreover, a complementary co-inner factor can be obtained as
#
"
A
+ LC B + LD
~
N? =
;D~ ? B Y y D~ ?
if p < m, where D~ ? is a full row rank matrix such that D~ ? D~ ? = I ; D R~ ;1D.
362
"
L = ;Y C #
Y = Ric A ;C C 0:
0 ;A
"
i.e., if
M
N
G= A B
C D
and dene
363
F = ;R;1 (B X + D C )
and
"
;1D C
;BR;1B
X = Ric A ; BR
0:
;C R~ ;1 C ;(A ; BR;1 D C )
(b) Suppose (C; A) is detectable and (A; B ) has no uncontrollable modes on the imaginary axis. Then there is a normalized left coprime factorization G = M~ ;1 N~
h
"
A + LC L B + LD
M~ N~ := ~ ;1=2 ~ ;1=2 ~ ;1=2
R C R
R D
i
where
L = ;(BD + Y C )R~;1
and
Y = Ric
"
(A ; BD R~ ;1C )
;BR;1B
;C R~ ;1 C
0:
;(A ; BD R~ ;1 C )
"
M are given
(c) The controllability grammian P and observability grammian Q of
N
by
P = (I + Y X );1Y; Q = X
h
while the controllability grammian P~ and observability grammian Q~ of M~ N~
are given by
P~ = Y; Q~ = (I + XY );1 X:
Proof. We shall only prove the rst part of (c). It is obvious that Q = X since the
Riccati equation for X can be written as
X (A + BF ) + (A + BF ) X +
"
F
C + DF
# "
F
C + DF
"
(A + BF )
0
P = Ric
:
;
1
;BR B ;(A + BF )
=0
364
"
I X ; then
Now let T =
0 I
"
(A + BF )
0
=T
;
1
;BR B ;(A + BF )
"
A ; BD R~ ;1 C
;C R~ ;1 C
T ;1:
;BR;1 B ;(A ; BD R~ ;1 C )
This shows that the stable invariant subspaces for these two Hamiltonian matrices are
related by
"
"
0
A ; BD R~ ;1 C
;C R~ ;1 C
X;
= T X;
;
1
;
1
;BR B ;(A + BF )
;BR B
;(A ; BD R~ ;1 C )
or
#
#
"
#
"
#
"
"
I
I
+ XY
I
I
:
= Im
= Im
= T Im
Im
Y (I + XY );1
Y
Y
P
Hence we have P = Y (I + XY );1 .
2
(A + BF )
14
H2 Optimal Control
In this chapter we treat the optimal control of linear time-invariant systems with a
quadratic performance criterion. The material in this chapter is standard, but the
treatment is somewhat novel and lays the foundation for the subsequent chapters on
H1 -optimal control.
x_ = Ax + B2 u; x(t0 ) = x0
(14:1)
where x0 is given but arbitrary. Our objective is to nd a control function u(t) dened
on [t0 ; T ] which can be a function of the state x(t) such that the state x(t) is driven to a
(small) neighborhood of origin at time T . This is the so-called Regulator Problem. One
might suggest that this regulator problem can be trivially solved for any T > t0 if the
system is controllable. This is indeed the case if the controller can provide arbitrarily
large amounts of energy since, by the denition of controllability, one can immediately
construct a control function that will drive the state to zero in an arbitrarily short
time. However, this is not practical since any physical system has the energy limitation,
i.e., the actuator will eventually saturate. Furthermore, large control action can easily
drive the system out of the region where the given linear model is valid. Hence certain
limitations have to be imposed on the control in practical engineering implementation.
365
H2 OPTIMAL CONTROL
366
The constraints on control u may be measured in many dierent ways; for example,
Z
t0
kuk dt;
t0
kuk2 dt;
sup kuk
t2[t0 ;T ]
i.e., in terms of L1 -norm, L2 -norm, and L1 -norm, or more generally, weighted L1 -norm,
L2 -norm, and L1 -norm
T
t0
kWu uk dt;
t0
sup kWu uk
t2[t0 ;T ]
T
t0
kWx xk dt;
t0
sup kWx xk
t2[t0 ;T ]
for some weighting matrix Wx . Hence the regulator problem can be posed as an optimal
control problem with certain combined performance index on u and x, as given above.
In this chapter, we shall be concerned exclusively with the L2 performance problem or
quadratic performance problem. Moreover, we will focus on the innite time regulator
problem, i.e., T ! 1, and, without loss of generality, we shall assume t0 = 0. In this
case, our problem is as follows: nd a control u(t) dened on [0; 1) such that the state
x(t) is driven to the origin at t ! 1 and the following performance index is minimized:
1 " x(t) # " Q S
min
u 0
u(t)
S R
Z
#"
x(t) dt
u(t)
(14:2)
Q S 0:
S R
(14:3)
Since R is positive denite, it has a square root, R1=2 , which is also positive-denite.
By the substitution
R1=2 u;
we may as well assume at the start that R = I . In fact, we can even assume S = 0 by
using a pre-state feedback u = ;S x + v provided some care is exercised; however, this
u
367
will not be assumed in the sequel. Since the matrix in (14.3) is positive semi-denite
with R = I , it can be factored as
"
"
Q S = C1
S I
D12
C1 D12 :
u2L2 [0;1)
u2L2 [0;1)
(14:4)
x_ = Ax + Bu; x(0) = x0
(14:5)
without explicitly mentioning the condition that the control should drive the state to the
origin. Instead some assumptions are imposed on Q; S , and R (or equivalently C1 and
D12 ) to ensure that the optimal control law u has this property. To see what assumption
one needs to make in order to ensure that the minimization problem formulated in (14.4)
and (14.5) has a sensible solution, let us consider a simple example with A = 1; B =
1; Q = 0; S = 0, and R = 1:
min
u2L2 [0;1) 0
u2 dt; x_ = x + u; x(0) = x0 :
H2 OPTIMAL CONTROL
368
(A4)
Remark 14.1 Assumption (A1) is clearly necessary for the existence of a stabilizing
control function u. The assumption (A2) is made for simplicity of notation and is
D12 = I . Note also that D? drops out when D12
actually a restatement that R = D12
is square. It is interesting to point out that (A3) is not needed in the Extended LQR
problem. The assumption (A3) enforces that the unconditional optimization problem
will result in a stabilizing control law. In fact, the assumption (A3) together with (A1)
guarantees that the input/output stability implies the internal stability, i.e., u 2 L2
and z 2 L2 imply x 2 L2 , which will be shown in Lemma 14.1. Finally note that (A4)
C1 ) has no unobservable modes
is equivalent to the condition that (D? C1 ; A ; B2 D12
on the imaginary axis and is weaker than the popular assumption of detectability of
C1 ). (A4), together with the stabilizability of (A; B2 ), guarantees by
(D? C1 ; A ; B2 D12
Corollary 13.10 that the following Hamiltonian matrix belongs to dom(Ric) and that
X = Ric(H ) 0:
H =
"
"
"
h
A
0
B2
;
;C1 C1 ;A
;C1 D12 D12 C1 B2
#
C1
A ; B2 D12
;B2B2
:
;C1 D? D? C1 ;(A ; B2 D12 C1 )
(14.8)
369
C1 =
Then (C1 ; A) is detectable and A ; B2 D12
"
1 0
0 ;2
(14:9)
(14:10)
AF := A + B2 F; CF := C1 + D12 F
and re-arrange equation (14.9) to get
x_ = AF x; x(0) = x0
z = CF x
or equivalently
x_ = AF x + x0 (t); x(0; ) = 0
z = CF x:
The associated transfer matrix is
"
Gc (s) = AF I
CF 0
and
The proof of the following theorem requires a preliminary result about internal
stability given input-output stability.
Lemma 14.1 If u; z 2 Lp [0; 1) for p 1 and (C1 ; A) is detectable in the system
described by equations (14.6) and (14.7), then x 2 Lp [0; 1). Furthermore, if p < 1,
then x(t) ! 0 as t ! 1.
H2 OPTIMAL CONTROL
370
Proof. Since (C1; A) is detectable, there exists L such that A + LC1 is stable. Let x^
be the state estimate of x given by
x^_ = (A + LC1 )^x + (LD12 + B2 )u ; Lz:
Then x^ 2 Lp [0; 1) since z and u are in Lp [0; 1). Now let e = x ; x^; then
e_ = (A + LC1 )e
and e 2 Lp [0; 1). Therefore, x = e + x^ 2 Lp [0; 1). It is easy to see that e(t) ! 0
as t ! 1 for any initial condition e(0). Finally, x(t) ! 0 follows from the fact that if
p < 1 then x^ ! 0.
2
Theorem 14.2 There exists a unique optimal control for the LQR problem, namely
u = Fx. Moreover,
min kz k2 = kGc x0 k2 :
u2L2 [0;1)
Note that the optimal control strategy is constant gain state feedback, and this gain
is independent of the initial condition x0 .
Proof. With the change of variable v = u ; Fx, the system can be written as
"
"
x_ = AF B2
z
CF D12
#"
x
v
x(0) = x0 :
(14:12)
Now if v 2 L2 [0; 1), then x; z 2 L2 [0; 1) and x(1) = 0 since AF is stable. Hence
u = Fx + v 2 L2 [0; 1). Conversely, if u; z 2 L2 [0; 1), then from Lemma 14.1 x 2
L2 [0; 1). So v 2 L2 [0; 1). Thus the mapping v = u ; Fx between v 2 L2 [0; 1) and
those u 2 L2 [0; 1) that make z 2 L2 [0; 1) is one-to-one and onto. Therefore,
min kz k2 = v2Lmin
kz k2 :
[0;1)
u2L2 [0;1)
By dierentiating x(t) Xx(t) with respect to t along a solution of the dierential equation (14.12) and by using (14.9) and the fact that CF D12 = ;XB2, we see that
d x Xx = x_ Xx + x X x_
dt
= x (AF X + XAF )x + 2x XB2 v
= ;x CF CF x + 2x XB2 v
= ;(CF x + D12 v) (CF x + D12 v) + 2x CF D12 v + v v + 2x XB2 v
= ; kz k2 + kvk2 :
(14.13)
371
This method of proof, involving change of variables and the completion of the square,
is a standard technique and variants of it will be used throughout this book. An alternative proof can be given in frequency domain. To do that, let us rst note the following
fact:
Lemma 14.3 Let a transfer matrix be dened as
"
U := AF B2 2 RH1 :
CF D12
Then U is inner and U Gc 2 RH?2 .
Proof. The proof uses standard manipulations of state space realizations. From U we
get
"
U (s) =
B2
U U = 64 0
B2
B2
I
AF
CF
D12
;AF ;CF :
D12
3
7
5
; U Gc = 64 0
B2
"
;AF ;CF CF 0
I ;X
0 I
I 75 :
AF
CF 0
D12
;AF
U U = 64 0
B2
2
U Gc = 64
;AF
0
B2
;X
AF
AF B2
0 I
3
7
5
3
7
5
=I
"
;AF ;X 2 RH? :
=
2
B2
H2 OPTIMAL CONTROL
372
Remark 14.2 It is clear that the LQR problem considered above is essentially equiva-
lent to minimizing the 2-norm of z with the input w = x0 (t) in the following diagram:
z
A I B2
C1 0 D12
I 0 0
373
i
Proof. The proof of this theorem is very similar to the proof of the standard LQR
problem except that, in this case, the input/output stability may not necessarily imply
the internal stability. Instead, the internal stability is guaranteed by the way of choosing
control law.
Suppose that u 2 L2 [0; 1) is such a control law that the system is stable, i.e.,
x 2 L2 [0; 1). Then v = u ; Fx 2 L2 [0; 1). On the other hand, let v 2 L2 [0; 1) and
consider
"
# "
#"
#
x_ = AF B2
x ; x(0) = x :
0
CF D12
v2L2 [0;1)
Using the same technique as in the proof of the standard LQR problem, we have
kz k22 = x0 Xx0 + kvk22 :
And the unique optimal control is v = 0, i.e., u = Fx.
H2 OPTIMAL CONTROL
374
- x_ = Ax + B2u
x-
B2 (;sI ; A );1 F + F (sI ; A);1 B2 ; B2 (;sI ; A );1 F F (sI ; A);1 B2
C1 (sI ; A);1 B2
+B2 (;sI ; A );1 C1 D12 + D12
+B2(;sI ; A );1 C1 C1 (sI ; A);1 B2 = 0:
D12 = I .
Then the result follows from completing the square and from the fact that D12
;
(14:14)
and
;
;
(14:15)
375
Note that the inequality (14.15) follows from taking the inverse of inequality (14.14).
Dene G(s) = ;F (sI ; A);1 B2 and assume for the moment that the system is single
input. Then the inequality (14.14) shows that the open-loop Nyquist diagram of the
system G(s) in Figure 14.1 never enters the unit disk centered at (;1; 0) of the complex
plane. Hence the system has at least the following stability margins:
kmin 12 ; kmax = 1; min ;60o; max 60o
i.e., the system has at least a 6dB (= 20 log 2) gain margin and a 60o phase margin in
both directions. A similar interpretation may be generalized to multiple input systems.
Next, it is noted that the inequality (14.15) can also be given some robustness
interpretation. In fact, it implies that the closed-loop system in Figure 14.1 is stable
even if the open-loop system G(s) is perturbed additively by a 2 RH1 as long as
kk1 < 1. This can be seen from the following block diagram and small gain theorem
where the transfer matrix from w to z is exactly I + F (j!I ; A ; B2 F );1 B2 .
- x_ = Ax + B2u - ?e -
z
y
w
A B1 B2
6
G(s) = 4 C1 0 D12 75 :
C2 D21 0
H2 OPTIMAL CONTROL
376
Notice the special o-diagonal structure of D: D22 is assumed to be zero so that G22
is strictly proper1; also, D11 is assumed to be zero in order to guarantee that the H2
problem properly posed.2 The case for D11 6= 0 will be discussed in Section 14.7.
The following additional assumptions are made for the output feedback H2 problem
in this chapter:
(i) (A; B2 ) is stabilizable and (C2 ; A) is detectable;
h
(ii) D12 has full column rank with D12 D? unitary, and D21 has full row rank
"
D
with ~21 unitary;
D?
"
A ; j!I B2
C1
D12
"
A ; j!I B1
C2
D21
(iii)
(iv)
The rst assumption is for the stabilizability of G by output feedback, and the third
and the fourth assumptions together with the rst guarantee that the two Hamiltonian matrices associated with the H2 problem below belong to dom(Ric). The rank
assumptions (ii) are necessary to guarantee that the H2 optimal controller is a nite
dimensional linear time invariant one, while the unitary assumptions are made for the
simplicity of the nal solution; they are not restrictions (see e.g., Chapter 17).
"
A
0
B2
H2 :=
;
;C1 C1 ;A
;C1 D12
"
"
A
0
C2
J2 :=
;
;B1 B1 ;A
;B1D21
C1 B2
D12
D21 B1 C2
1 As we have discussed in Section 12.3.4 of Chapter 12 there is no loss of generality in making this
assumption since the controller for D22 nonzero case can be recovered from the zero case.
2 Recall that a rational proper stable transfer function is an RH function i it is strictly proper.
2
377
and
F2
Moreover, min kTzw k22 = kGc B1 k22 + kF2 Gf k22 = kGc L2k22 + kC1 Gf k22 .
The controller Kopt has the well-known separation structure, which will be discussed
in more detail in Section 14.9. For comparison with the H1 results, it is useful to
describe all suboptimal controllers.
Theorem 14.8 The family of all admissible controllers such that kTzw k2 <
equals
the set of all transfer matrices from y to u in
u
M2
y
A^2 ;L2 B2
6
M2 (s) = 4 F2 0 I
;C2 I 0
3
7
5
H2 OPTIMAL CONTROL
378
z
y
y1
w
XXXX u
XXX
M2
G
u1
;B2 F2
B1 B2
7
6
6 0
AL2 B1L2 0 77 :
N = 66
0 D12 75
4 C1F2 ;D12 F2
D21 0
0
C2
Dene
"
We have
AF2
"
kTzw k22 = kGc B1 k22 + kUF2 Gf ; UQV k22 = kGc B1 k22 + kF2 Gf ; QV k22 :
379
It can also be shown easily by duality that Gf and V are orthogonal, i.e., Gf V 2 RH?2 ,
and V is a co-inner, so we have
kTzw k22 = kGc B1 k22 + kF2 Gf ; QV k22 = kGc B1 k22 + kF2 Gf k22 + kQk22 :
This shows clearly that Q = 0 gives the unique optimal control, so K = F`(M2 ; 0) is
the unique optimal controller. Note also that kTzw k2 is nite if and only if Q 2 RH2 .
Hence Theorem 14.7 and 14.8 follow easily.
It is interesting to examine the structure of Gc and Gf . First of all the transfer
matrix Gc can be represented as a xed system with the feedback matrix F2 wrapped
around it:
A I B2
C1 0 D12
I 0 0
F2
F2 is, in fact, an optimal LQR controller and minimizes the H2 norm of Gc . Similarly,
Gf can be represented as
A B1 I
I 0 0
C2 D21 0
L2
Let us consider the generalized system structure again with D11 not necessarily zero:
2
A B1 B2
6
G(s) = 4 C1 D11 D12
C2 D21 0
3
7
5
H2 OPTIMAL CONTROL
380
We shall consider the following question: what will happen and under what condition
will the H2 optimal control problem make sense if D11 6= 0?
Recall that F`(M2 ; Q) with Q 2 RH1 parameterizes all stabilizing controllers for G
regardless of D11 = 0 or not. Now again consider the closed loop transfer matrix with
the controller K = F`(M2 ; Q); then
Hence the H2 optimal control problem will make sense, i.e., having nite H2 norm, if
and only if there is a constant Q(1) such that
D11 D
Q(1) = ;D12
21
(14:16)
Note that the equation (14.16) is a very restrictive condition. For example, suppose
D12 =
"
D21 = 0 I
"
D1111 D1112 = 0 0
D1121 0
0 0
and that Q(1) = ;D1122 . So only D1122 can be nonzero for a sensible H2 problem.
Hence from now on in this section we shall assume that (14.16) holds and denotes
D11 D21
. To nd the optimal control law for the system G with D11 6= 0,
DK := ;D12
let us consider the following system conguration:
381
- DK
- ;DK
Then
and
-f
6
u1
-f
6
G^
K^
A + B2 DK C2 B1 + B2 DK D21 B2
6
^
G = 4 C1 + D12 DK C2
0
D12
C2
D21
0
3
7
5
^
K = DK + K:
It is easy to check that the system G^ satises all assumptions in Section 14.5; hence the
controller formula in Section 14.5 can be used. A little bit of algebra will show that
#
"
^2 ; B2 DK C2 ;(L2 ; B2 DK )
A
K^ =
F2 ; DK C2
0
is the H2 optimal controller for G^ . Hence the controller K for the original system G
will be given by
"
#
^2 ; B2 DK C2 ;(L2 ; B2 DK )
A
K=
= F` (M2 ; DK ):
F2 ; DK C2
DK
In this section we look at various H2 -optimization problems from which the output
feedback solutions of the previous sections will be constructed via a separation argument.
All the special problems in this section are to nd K stabilizing G and minimizing the
H2 -norm from w to z in the standard setup, but with dierent structures for G. As
H2 OPTIMAL CONTROL
382
in Chapter 12, we shall call these special problems, respectively, state feedback (SF),
output injection (OI), full information (FI), full control (FC), disturbance feedforward
(DF), and output estimation (OE). OI, FC, and OE are natural duals of SF, FI, and
DF, respectively. The output feedback solutions will be constructed out of the FI and
OE results.
The special problems SF, OI, FI, and FC are not, strictly speaking, special cases of
the output feedback problem since they do not satisfy all of the assumptions for output
feedback (while DF and OE do). Each special problem inherits some of the assumptions
(i)-(iv) from the output feedback as appropriate. The assumptions will be discussed in
the subsections for each problem.
In each case, the results are summarized as a list of three items; (in all cases, K
must be admissible)
1. the minimum of kTzw k2 ;
2. the unique controller minimizing kTzw k2 ;
3. the family of all controllers such that kTzw k2 <
, where
is greater than the
minimum norm.
Warning: we will be more specic below about what we mean about the uniqueness and
all controllers in the second and third item. In particular, the controllers characterized
here for SF, OI, FI and FC problems are neither unique nor all-inclusive. Once again we
regard all controllers that give the same control signal u, i.e., having the same transfer
function from w to u, as an equivalence class. In other words, if K1 and K2 generate
the same control signal u, we will regard them as the same, denoted as K1
= K2. Hence
the \unique controller" here means one of the controllers from the unique equivalence
class. The same comments apply to the \all" situation. This will be much clearer in
section 14.8.1 when we consider the state feedback problem. In that case we actually
give a parameterization of all the elements in the equivalence class of the \unique"
optimal controllers. Thus the unique controller is really not unique. We chose not to
give the parameterization of all the elements in an equivalence class in this book since it
is very messy, as can be seen in section 14.8.1, and not very useful. However, it will be
seen that this equivalence class problem will not occur in the general output feedback
case including DF and OE problems.
A B1 B2
6
GSF (s) = 4 C1 0 D12
I 0 0
with the following assumptions:
3
7
5
383
(iii)
A ; j!I B2
C1
D12
This is very much like the LQR problem except that we require from the start that u
be generated by state feedback and that the detectability of (C1 ; A) is not imposed since
the controllers are restricted to providing internal stability. The controller is allowed to
be dynamic, but it turns out that dynamics are not necessary.
State Feedback:
1. min kTzw k2 = kGc B1 k2 = (trace(B1 X2 B1 ))1=2
2. K (s)
= F2
Remark 14.3 The class of all suboptimal controllers for state feedback are messy and
are not very useful in this book, so they are omitted, as are the OI problems.
~
Proof. Let K be a stabilizing controller, u = K (s)x. Change control variables by
dening v := u ; F2 x and then write the system equations as
2
x_
AF2
B1 B2
7 6
z 5 = 4 C1F2
0 D12
v
(K ; F2 ) 0 0
6
4
32
76
54
x
w 75 :
v
z
AF2 B1 B2
C1F2 0 D12
I
0 0
- K ; F2
Let Tvw denote the transfer matrix from w to v. Notice that Tvw 2 RH2 because
K stabilizes G. Then
Tzw = Gc B1 + UTvw
H2 OPTIMAL CONTROL
384
"
Thus min kTzw k22 = kGc B1 k22 and the minimum is achieved i Tvw = 0. Furthermore,
K = F2 is a controller achieving this minimum, and any other controllers achieving
minimum are in the equivalence class of F2 .
2
Note that the above proof actually yields a much stronger result than what is needed.
The proof that the optimal Tvw is Tvw = 0 does not depend on the restriction that the
controller measures just the state. We only require that the controller produce v as a
causal stable function Tvw of w. This means that the optimal state feedback is also
optimal for the full information problem as well.
We now give some further explanation about the uniqueness of the optimal controller
that we commented on before. The important observation for this issue is that the
controllers making Tvw = 0 are not unique. The controller given above, F2 , is only one
of them. We will now try to nd all of those controllers that stabilize the system and
give Tvw = 0, i.e., all K (s)
= F2 .
Proposition 14.9 Let Vc be a matrix whose columns form a basis for KerB1 (VcB1 =
0). Then all H2 optimal state feedback controllers can be parameterized as Kopt =
F` (Msf ; ) with 2 RH2 and
"
I
Msf = F2
:
Vc (sI ; AF2 ) ;Vc B2
Proof. Since
"
Tvw = I ; AF2 B2
K ; F2 0
#!;1 "
AF2 B1 = 0;
K ; F2 0
we get
Pc (s) := AF2 B2 :
I 0
385
(14:18)
(14:19)
Therefore,
A
6
6 C1
GFI (s) = 66 " #
I
4
0
B1
"
0
0
B2
D
" 12#
0
0
3
7
7
7
7
5
The assumptions relevant to the FI problem are the same as the state feedback problem.
This is similar to the state feedback problem except that the controller now has more
information (w). However, as was pointed out in the discussion of the state feedback
problem, this extra information is not used by the optimal controller.
Full Information:
1. min kTzw k2 = kGc B1 k2 = (trace(B1 X2 B1 ))1=2
h
i
2. K (s)
= F2 0
h
3. K (s)
= F2 Q(s) , where Q 2 RH2 , kQk22 <
2 ; kGcB1 k22
H2 OPTIMAL CONTROL
386
Proof. Items 1 and 2 follow immediately from the proof of the state feedback results
because the argument that Tvw = 0 did not depend on the restriction to state feedback
only. Thus we only need to prove item 3. Let K be an admissible controller such that
kTzw k2 <
. As in the SF proof, dene a new control variable v = u ; F2 x; then the
closed-loop system is as shown below
z
y
with
AF2
6
6 C1F2
G~ FI = 66 " #
I
4
G~ FI
0
0
K~
B1
"
w
B2
7
7
D
12
~ = K ; [F2 0]:
"
# 7; K
7
0
0
I
0
Denote by Q the transfer matrix from w to v; it belongs to RH2 by internal stability
and the fact that D12 has full column rank and Thzw with zi= C1F2 x + Dh12 Qw hasi nite
H2 norm. Then u = F2 x + v = F2 x + Qw = F2 Q y so K
= F2 Q , and
2
2
2
2
kTzw k2 = kGc B1 + UQk2 = kGc B1 k2 + kQk2; hence,
Likewise, one can show that every controller of the form given in item no.3 is admissible
and suboptimal.
2
The results for DF, OI, FC, and OE follow from the parallel development of Chapter 12.
Disturbance Feedforward:
A B1 B2
6
GDF (s) = 4 C1 0 D12
C2 I 0
3
7
5
This problem inherits the same assumptions (i)-(iii) as in the state feedback problem,
in addition to the stability condition of A ; B1 C2 .
1. min kTzw k2 = kGc B1 k2
"
A + B2 F2 ; B1 C2 B1
2. K (s) =
F2
0
387
u
M2D
y
A + B2 F2 ; B1 C2 B1 B2
6
M2D (s) = 4
F2
0 I
;C2
I 0
Output Injection:
A B1 I
6
GOI (s) = 4 C1 0 0
C2 D21 0
"
D
(ii) D21 has full row rank with ~21
D?
"
(iii)
A ; j!I B1
C2
D21
3
7
5
unitary;
L
2. K (s)
= 2
Full Control:
A B1
6
6
GFC (s) = 64 C1 0
C2 D21
h
h
h
I 0
0 I
i 3
i 7
7
7
i 5
0 0
with the same assumptions as an output injection problem.
1. min kTzw k2 = kC1 Gf k2 = (trace(C1 Y2 C1 ))1=2
"
L
2. K (s)
= 2
0
3
7
5
H2 OPTIMAL CONTROL
388
3. K (s)
=
"
Output Estimation:
A B1 B2
GOE (s) = 64 C1 0 I
C2 D21 0
3
7
5
The assumptions are taken to be those in the output injection problem plus an additional
assumption that A ; B2 C1 is stable.
1. min kTzw k2 = kC1 Gf k2
"
A + L2 C2 ; B2 C1 L2
2. K (s) =
C1
0
u
y
M2O
-
A + L2 C2 ; B2 C1 L2 ;B2
M2O (s) = 64
C1
0 I
C2
I 0
3
7
5
where X2 := Ric(H2) and Y2 := Ric(J2 ) and the min is over all stabilizing controllers.
Note that F2 is the optimal state feedback in the Full Information problem and L2 is the
389
optimal output injection in the Full Control case. The well-known separation property
of the H2 solution is re
ected in the fact that K2 is exactly the optimal output estimate
of F2 x and can be obtained by setting C1 = F2 in OE.2. Also, the minimum cost is the
sum of the FI cost (FI.1) and the OE cost for estimating F2 x (OE.1).
The controller equations can be written in standard observer form as
z=
"
AF2 B1 B2
C1F2 0 D12
#
w = G B w + Uv
c 1
v
"
(14:20)
A + B2 F2 + L2C2 ;L2 ;
F2
0
(14:21)
H2 OPTIMAL CONTROL
390
Continuing with the development in the previous proof, we see that the set of all suboptimal controllers equals the set of all K 's such that kTvw k22 <
2 ; kGc B1 k22 . Apply
item OE.3 to get that such K 's are parameterized by
u
M2
y
A^2 ;L2 B2
6
M2 (s) = 4 F2 0 I
;C2 I 0
3
7
5
We have shown that the system with LQR controller has at least 60o phase margin
and 6dB gain margin. However, it is not clear whether these stability margins will be
preserved if the states are not available and the output feedback H2 (or LQG) controller
has to be used. The answer is provided here through a counterexample: there are no
guaranteed stability margins for a H2 controller.
Consider a single input and single output two state generalized dynamical system:
2
G(s) =
"
1 1
0 1
6
6
6
6 "
6
6
6
6
0
6
4
h
pq pq
0
1 0
p 0
p 0
"
"
"
0 1
0
1
0
1
0
X2 =
and
where
"
; Y2 =
"
"
F2 = ; 1 1 ; L2 = ; 1
1
h
= 2 + 4 + q ; = 2 + 4 + :
# 3
7
7
7
# 7
7
7
7
7
7
5
391
392
H2 OPTIMAL CONTROL
The detailed treatment of H2 related theory, LQ optimal control, Kalman ltering, etc.,
can be found in Anderson and Moore [1990] or Kwakernaak and Sivan [1972].
15
Linear Quadratic
Optimization
This chapter considers time domain characterizations of Hankel operators and Toeplitz
operators by means of some related quadratic optimizations. These characterizations
will be used to prove a max-min problem which is the key to the H1 theory considered
in the next chapter.
x_ = Ax + Bw
z = Cx + Dw:
(15.1)
Consider rst the problem of using an input w 2 L2; to maximize kP+ z k22. This is
exactly the standard problem of computing the Hankel norm of G, i.e., the induced
norm of the Hankel operator
P+ MG : H2? ! H2 ;
and the norm can be expressed in terms of the controllability Gramian Lc and observability Gramian Lo:
ALc + LcA + BB = 0
A Lo + LoA + C C = 0:
393
394
Although this result is well-known, we will include a time-domain proof similar in technique to the proofs of the optimal H2 and H1 control results.
Lemma 15.1 w2L
inf kwk22 x(0) = x0 = x0 y0 where y0 solves Lcy0 = x0 .
2;
Proof. Assume (A; B) is controllable; otherwise, factor out the uncontrollable subspace. Then Lc is invertible and y0 = L;c 1x0 . Moreover, w 2 L2; can be used to
produce any x(0) = x0 given x(;1) = 0. We need to show
2 x(0) = x = x L;1 x :
inf
k
w
k
(15:2)
0
2
0 c 0
w2L2;
To show this, we can dierentiate x(t) L;c 1 x(t) along the solution of (15.1) for any given
input w as follows:
d (x L;1 x) = x_ L;1 x + x L;1 x_ = x (A L;1 + L;1 A)x + 2hw; B L;1 xi:
c
c
c
c
c
c
dt
Using Lc equation to substitute for A L;c 1 + L;c 1 A and completion of the squares gives
d ;1
2
;1 2
dt (x Lc x) = kwk ; kw ; B Lc xk :
Integration from t = ;1 to t = 0 with x(;1) = 0 and x(0) = x0 gives
x0 L;c 1x0 = kwk22 ; kw ; B L;c 1 xk22 kwk22 :
If w(t) = B e;A t L;c 1x0 = B L;c 1e(A+BB L;c 1)t x0 on (;1; 0], then w 2 L2; , w =
B L;c 1 x and equality is achieved, thus proving (15.2).
2
Lemma 15.2
Proof. Given x(0) = x0 and w = 0, for t 0 the norm of z(t) = CeAtx0 can be found
from
kP+z k22 =
395
Remark 15.1 Another useful way to characterize the Hankel norm is to examine the
following quadratic optimization with initial condition x(;1) = 0:
sup kP+ z k22 ; 2 kwk22 :
w2L2;
sup
06=w2L2;
So the Hankel norm is equal to the smallest such that the above inequality holds.
Now
w2L2;
x0 2Rn
(
0; (Lo Lc) 2 ;
+1; (Lo Lc) > 2 :
Hence the Hankel norm is equal to the square root of (Lo Lc).
If a transfer matrix G 2 RH1 and kGk1 < 1, then by Corollary 13.24, the Hamiltonian
matrix
"
A + BR;1 D C
BR;1 B
H=
;
;
1
;C (I + DR D )C ;(A + BR;1 D C )
R = 2 I ; D D
A X + XA + (XB + C D)R;1 (B X + D C ) + C C = 0:
(15:3)
The following lemma oers yet another consequence of kGk1 < 1. (Recall that the H1
norm of a stable matrix is the Toeplitz operator norm.)
396
Proof. We can dierentiate x(t) Xx(t) as in the last section, use the Riccati equation
(15.3) to substitute for A X + XA, and complete the squares to get
= sup
8
>
<
>
:
06=x0 2Rn
max (X ;
2 X0 ) < 0;
max (X ;
2 X0 ) = 0;
= +1; max (X ;
2 X0 ) > 0:
< 0;
= 0;
GT (s) =
"
A C
B D
1 Note that since the system matrices are real, AT = A ; B T = B , etc. The conjugate transpose is
used here for the transpose for the sake of consistency in notation.
397
and kGk1 < 1 i kGT k1 < 1. Let J denote the following Hamiltonian matrix
"
"
h
J = A 0 + C R~ ;1 DB C
;BB ;A
;BD
Proof. Analogous to the proof of Lemma 15.3, we can dierentiate x(t) Y ;1x(t), use
the Riccati equation (15.5) to substitute for AY + Y A , and complete the squares to
get
d ;1
2
2
;1=2
;1
2
dt (x Y x) = ;kz k + kwk ; kR [Rw ; (B Y + D C )x]k
398
"
h
A
0
B2
;1 D C B
HW :=
+
R
2
2
2
;C C ;A
;C D2
w2
We are interested in a test for supw2BW kP+ z k2 < 1, or, equivalently,
sup k;wk2 < 1
w2BW
(15:8)
w
; 1
w2
= P+
= P+
h
h
"
w1
w1 2 H2? ; w2 2 L2
G1 G2
w2
i
G1 G2 P; w + P+ G2 P+ w2 :
i
Thus ; is the sum of the Hankel operator P+ MG : H2? H2? ! H2 and the Toeplitz
operator P+ MG2 : H2 ! H2 . The following lemma generalizes Corollary 13.24 (B1 =
0; D1 = 0) and Lemma 15.2 (B2 = 0; D2 = 0).
Lemma 15.6 supw2BW k;wk2 < 1 i the following two conditions hold:
(i) (D2 ) < 1 and HW 2 dom(Ric);
(ii) (WLc) < 1.
Proof. By Corollary 13.24, condition (i) is necessary for (15.8), so we will prove that
given condition (i), (15.8) holds i condition (ii) holds. We will do this by showing,
equivalently, that (WLc) 1 i supw2BW k;wk2 1. By denition of W , if w 2 W
then
kP+ z k22 ; kwk22 = kP+ z k22 ; kP+ w2 k22 ; kP; wk22 :
Note that the last term only contributes to kP+ z k22 through x(0). Thus if Lc is invertible,
then Lemma 15.1 and 15.3 yield
sup kP+ z k22 ; kwk22 x(0) = x0 = x0 Wx0 ; x0 L;c 1 x0
w2W
(15:9)
and the supremum is achieved for some w 2 W that can be constructed from the
previous lemmas. Since (WLc ) 1 i 9 x0 6= 0 such that the right-hand side of
399
If Lc is not invertible, we need only restrict x0 in (15.9) to Im(Lc ), and then the
above argument generalizes in a straightforward way.
2
In the dierential game problem considered later and in the H1 optimal control
problem, we will make use of the adjoint ; : H2 ! W , which is given by
; z =
"
"
P; (G1 z ) = P; G1 z
G2 z
G2
(15:10)
where P; Gz := P; (Gz ) = (P; MG)z . That the expression in (15.10) is actually the
adjoint of ; is easily veried from the denition of the inner product on vector-valued
L2 space. The adjoint of ; : W ! H2 is the operator ; : H2 ! W such that
< z; ;w >=< ; z; w > for all w 2 W , z 2 H2 . By denition, we have
A B1 B2
G(s) = G11 (s) G12 (s) = 64 C1 D11 D12
G21 (s) G22 (s)
C2 D21 D22
#
"
Denote
"
3
7
5
"
A B 2 RH :
=:
1
C D
h
i
D2 := D12
D2 := D21 D22
D22
Rx := I ; D2 D2 Ry := I ; D2 D2
400
"
"
h
i
A
0
B2
;1 D C B
HX :=
+
R
x
2
2
;C C ;A
;C D2
"
# "
#
h
i
A
0
C
2
;1 D2 B C2 :
HY :=
+
R
;BB ;A
;BD2 y
Dene W = H2? L2 , Z = H2 L2 , and ; : W ! Z as
"
"
#"
w
P 0
; 1 = +
w2
0 I
G11 G12
G21 G22
#"
w1 :
w2
Lemma 15.7 supw2BW k;wk2 < 1 holds i the following three conditions hold:
(i) (D2 ) < 1 and HX 2 dom(Ric);
(ii) (D2 ) < 1 and HY 2 dom(Ric);
(iii) (XY ) < 1 for X = Ric(HX ) and Y = Ric(HY ).
"
w1 = P G w1 + 6 h
4
+
P; G21 G22
w2
w2
;
Hence
"
w1
w2
# 7
5
w2BW
and
But
"
and
h
G21 G22
G12
G22
sup
w2 2H2
P+ G
"
w2
w2BW
These two inequalities then imply that (i) and (ii) are necessary. Analogous to the proof
of Lemma 15.6, we will show that given conditions (i) and (ii), supw2BW k;wk2 < 1
401
holds i condition (iii) holds. We will do this by showing, equivalently, that (XY ) 1
i supw2BW k;wk2 1. By denition of W , if w 2 W then
;
k;wk22 ; kwk22 = kP+ z k22 ; kP+ w2 k22 + kP; G21 G22 wk22 ; kP;wk22 :
Thus if Y is invertible, then Lemma 15.3 and 15.5 yield
sup k;wk22 ; kwk22 x(0) = x0 = x0 Xx0 ; x0 Y ;1 x0 :
w2W
Now the same arguments as in the proof of Lemma 15.6 give the desired conclusion. 2
x_ = Ax + B1 w + B2 u
z = C1 x + D12 u
with the following assumptions:
(i) (C1 ; A) is observable;
(ii) (A; B2 ) is stabilizable;
h
(15.11)
(15.12)
C1 D12 = 0 I .
(iii) D12
In this section, we are interested in answering the following question: when
sup umin
kz k2 < 1?
2L
w2BL2+
2+
Remark 15.3 This max-min problem is a game problem in the sense that u is chosen
to minimize the quadratic norm of z and that w is chosen to maximize the norm. In
other words, inputs u and w act as \opposing players". This linear quadratic max-min
problem can be reformulated in the traditional fashion as in the previous sections:
sup umin
2L
w2L2+
2+ 0
1
A conventional game problem setup would be to consider the min-max problem, i.e.,
switching the order of sup and min. However, it will be seen that they are equivalent
and that a saddle point exists. By saying that, we would like to warn readers that this
may not be true in the general case where z = C1 x + D11 w + D12 u and D11 6= 0. In that
case, it is possible that supw inf u < inf u supw . This will be elaborated in Chapter 17.
It should also be pointed out that the results presented here still hold, subject to
some minor modications, if the assumptions (i) and (iii) on the dynamical system are
relaxed to:
402
"
(i)'
A ; j!I B2
C1
D12
It is clear from the assumptions that H2 2 dom(Ric) and X2 = Ric(H2 ) > 0, where
"
A
;B2 B2 :
H2 =
;C1 C1 ;A
Let F2 = ;B2 X2 and D? be such that [D12 D? ] is an orthogonal matrix. Dene
#
"
"
"
;1
(15:13)
U? = AF2 ;X2 C1 D?
C1F2
D?
where AF2 = A + B2 F2 and C1F2 = C1 + D12 F2 . The following is easily proven using
Lemma 15.8 [U U?] is square and inner and a realization for Gc U U? is
h
Gc U U?
"
(15:14)
This implies that U and U? are each inner and that both U?Gc and U Gc are
in RH?2 . To answer our earlier question, dene a Hamiltonian matrix H1 and the
associated Riccati equation as
"
A
B1 B1 ; B2 B2
H1 :=
;C1 C1
;A
403
(15:15)
w2BH2 2H2
Since by Lemma 15.8 [U U? ] is square and inner, kz k2 = k[U U?] z k2, and
h
U U?
i
"
"
z = U Gc B1 w + = P; (U Gc B1 w)+ P+ (U Gc B1 w + ) :
U? Gc B1 w
U? Gc B1 w
Thus
sup min kz k2 =
w2BH2 2H2
"
sup min
2H
2
w2BH2
P; (U Gc B1 w) + P+ (U Gc B1 w + )
U? Gc B1 w
and the right hand of the above equation is minimized by = ;P+ (U Gc B1 w); we
have
sup min kz k2 =
w2BH2 2H2
=:
"
"
sup
w2BH2
P; (U Gc B1 w)
U? Gc B1 w
sup k; wk2 < 1
w2BH2
"
P; (U Gc B1 w) = P; U G B w
c 1
U?Gc B1 w
U?
W :=
("
q1
q2
q1 2 H2?; q2 2 L2 :
h
q
; 1 = P+ (B1 Gc (Uq1 + U? q2 )) = P+ B1 Gc U U?
q2
q2BW
"
q1
q2
404
This is just the condition (15.8), so from Lemma 15.6 and equation (15.14) we have
that
"
;1C C1 X ;1 #
A
X
F
2
2 1
2
HW :=
2 dom(Ric)
;X2B1 B1 X2
;AF2
"
;1
X; (H1 ) = T X; (HW ) = T Im I = Im I ; X2 W ;
W
X2
so H1 2 dom(Ric) and X1 = X2 (X2 ; W );1 X2 > 0.
(() If H1 2 dom(Ric) and X1 = Ric(H1 ) > 0, A + B1 B1 X1 ; B2 B2 X1 is
stable. Dene
Theorem 15.9 will be used in the next chapter to solve the FI H1 control problem.
16
z
y
K
405
w
406
where the plant G and controller K are assumed to be real-rational and proper. It will
be assumed that state space models of G and K are available and that their realizations
are assumed to be stabilizable and detectable. Recall again that a controller is said to
be admissible if it internally stabilizes the system. Clearly, stability is the most basic
requirement for a practical system to work. Hence any sensible controller has to be
admissible.
is minimized.
It should be noted that the optimal H1 controllers as dened above are generally not
unique for MIMO systems. Furthermore, nding an optimal H1 controller is often both
numerically and theoretically complicated, as shown in Glover and Doyle [1989]. This
is certainly in contrast with the standard H2 theory, in which the optimal controller
is unique and can be obtained by solving two Riccati equations without iterations.
Knowing the achievable optimal (minimum) H1 norm may be useful theoretically since
it sets a limit on what we can achieve. However, in practice it is often not necessary
and sometimes even undesirable to design an optimal controller, and it is usually much
cheaper to obtain controllers that are very close in the norm sense to the optimal ones,
which will be called suboptimal controllers. A suboptimal controller may also have other
nice properties over optimal ones, e.g., lower bandwidth.
Now suppose K is a stabilizing controller for the system G. Then the internal stability
guarantees Tzw = F` (G; K ) 2 RH1 , but the latter does not necessarily imply the
internal stability. The following lemma provides the additional (mild) conditions to the
equivalence of Tzw = F`(G; K ) 2 RH1 and internal stability of the closed-loop system.
To state the lemma, we shall assume that G and K have the following stabilizable and
detectable realizations, respectively:
2
A B1 B2
6
G(s) = 4 C1 D11 D12 75 ;
C2 D21 D22
"
^ ^
K (s) = A^ B^ :
C D
407
Lemma 16.1 Suppose that the realizations for G and K are both stabilizable and detectable. Then the feedback connection Tzw = F` (G; K ) of the realizations for G and K
is
"
A ; I B2
(a) detectable if
C1 D12
"
A ; I B1
(b) stabilizable if
C2 D21
Moreover, if (a) and (b) hold, then K is an internally stabilizing controller i Tzw 2
RH1 .
^ 1 D21
^ 1 C2
B1 + B2 DL
B2 L2 C^
A + B2 DL
7
6
^ 1 C2
^ 1 D22 C^
^ 1 D21
7
BL
A^ + BL
BL
F` (G; K ) = 64
5
^ 2
^ 1 D21
C1 + D12 L2 DC
D12 L2 C^
D11 + D12 DL
"
#
B
A
c
c
=:
Cc Dc
^ 22 );1 .
where L1 := (I ; D22 D^ );1 , L2 := (I ; DD
Suppose F`(G; K ) has undetectable state (x0 ; y0)0 and mode Re 0; then the PBH
test gives
"
and
Now if
Ac ; I
Cc
A ; I B2
C1 D12
#"
#"
x = 0:
y
#
x
^
^ =0
DL1 C2 x + L2 Cy
^ 1(C2 x + D22 Cy
^ ) + Ay
^ ; y = 0:
BL
"
A ; I B2
C1 D12
^ = 0. This implies Ay
^ = y. Since (C;
^ A^) is
has full column rank, then x = 0 and Cy
detectable, we get y = 0, which is a contradiction. Hence part (a) is proven, and part
(b) is a dual result.
2
These relations will be used extensively below to simplify our development and to
enable us to focus on input/output stability only.
408
One of the keys to the entire development of H1 theory is the fact that the contraction
and internal stability is preserved under an inner linear fractional transformation.
z
r
w
"
Suppose that P P = I , P21;1 2 RH1 and that Q is a proper rational matrix. Then
the following are equivalent:
(a) The system is internally stable, well-posed, and kTzw k1 < 1.
(b) Q 2 RH1 and kQk1 < 1.
Proof. (b) ) (a). Note that since P; Q 2 RH1, the system internal stability is guaranteed if (I ; P22 Q);1 2 RH1 . Therefore, internal stability and well-posedness follow
easily from kP22 k1 1, kQk1 < 1, and a small gain argument. (Note that kP22 k1 1
follows from the fact that P22 is a compression of P .)
To show that kTzw k1 < 1, consider the closed-loop system at any frequency s = j!
with the signals xed as complex constant vectors. Let kQk1 =: < 1 and note that
Twr = P21;1 (I ; P22 Q) 2 RH1 . Also let := kTwr k1 . Then kwk krk, and P inner
implies that kz k2 + krk2 = kwk2 + kvk2. Therefore,
kz k2 kwk2 + (2 ; 1)krk2 [1 ; (1 ; 2 );2 ]kwk2
which implies kTzw k1 < 1 .
(a) ) (b). To show that kQk1 < 1, suppose there exist a (real or innite) frequency
! and a constant nonzero vector r such that at s = j! , kQrk krk. Then setting
w = P21;1 (I ; P22 Q)r, v = Qr gives v = Tvw w. But as above, P inner implies that
kz k2 + krk2 = kwk2 + kvk2 and, hence, that kz k2 kwk2 , which is impossible since
kTzw k1 < 1. It follows that max (Q(j!)) < 1 for all !, i.e., kQk1 < 1 since Q is
rational.
To show Q 2 RH1 , let Q = NM ;1 with N; M 2 RH1 be a right coprime factorization, i.e., there exist X; Y 2 RH1 such that XN + Y M = I . We shall show that
M ;1 2 RH1 . By internal stability we have
Q(I ; P22 Q);1 = N (M ; P22 N );1 2 RH1
and
409
Thus
In this chapter, we discuss a simplied version of H1 theory. The general case will be
considered in the next chapter. The main reason for doing so is that the general case
has its unique features but is much more involved algebraically. Involved algebra may
distract attention from the essential ideas of the theory and therefore lose insight into
the problem. Nevertheless, the problem considered below contains the essential features
of the H1 theory.
The realization of the transfer matrix G is taken to be of the form
2
A B1 B2
6
G(s) = 4 C1 0 D12 75 :
C2 D21 0
The following assumptions are made:
(i) (A; B1 ) is stabilizable and (C1 ; A) is detectable;
(ii) (A; B2 ) is stabilizable and (C2 ; A) is detectable;
h
"
C1 D12 = 0 I ;
(iii) D12
"
(iv)
B1 D = 0 .
D21 21
I
Assumption (i) is made for a technical reason: together with (ii) it guarantees that
the two Hamiltonian matrices (H2 and J2 in Chapter 14) in the H2 problem belong
to dom(Ric). This assumption simplies the theorem statements and proofs, but if
it is relaxed, the theorems and proofs can be modied so that the given formulae are
still correct, as will be seen in the next chapter. An important simplication that is
a consequence of the assumption (i) is that internal stability is essentially equivalent
to input-output stability (Tzw 2 RH1 ). This equivalence enables us to focus only on
input/output stability and is captured in Corollary 16.3 below. Of course, assumption
(ii) is necessary and sucient for G to be internally stabilizable, but is not needed
410
to prove the equivalence of internal stability and Tzw 2 RH1 . (Readers should be
clear that this does not mean that the realization for G need not be stabilizable and
detectable. In point of fact, the internal stability and input-output stability can never
be equivalent if either G or K has unstabilizable or undetectable modes.)
Corollary 16.3 Suppose that assumptions (i), (iii), and (iv) hold. Then a controller
K is admissible i Tzw 2 RH1 .
Proof. The realization for plant G is stabilizable and detectable by assumption (i).
We only need to verify that the rank conditions of the two matrices in Lemma 16.1 are
satised.
Nowi suppose assumptions (i) and (iii) are satised and let D? be such that
h
D12 D? is a unitary matrix. Then
"
A ; I B2
rank
C1 D12
rank 64
2
= rank 64
So
"
3
"
# 7
5
D12
D?
A ; I #
"
D? C1
A ; I B2
C1 D12
"
A ; I
D? C1
A ; I B2
C1 D12
"
B2 #
I 75 :
0
has full column rank. However, the last matrix has full rank for all Re 0 i (D? C1 ; A)
)C1 = C1 , (D C1 ; A) is detectable i
is detectable. Since D? (D? C1 ) = (I ; D12 D12
?
(C1 ; A) is detectable. The rank condition for the other matrix follows by duality. 2
Two additional assumptions that are implicit in the assumed realization for G(s) are
that D11 = 0 and D22 = 0. As we have mentioned many times, D22 6= 0 does not pose
any problem since it is easy to form an equivalent problem with D22 = 0 by a linear
fractional transformation on the controller K (s). However, relaxing the assumption
D11 = 0 complicates the formulae substantially, as will be seen in the next chapter.
In this subsection, we present the necessary and sucient conditions for the existence
of an admissible controller K (s) such that kTzw k1 <
for a given
, and, furthermore,
411
if the necessary and sucient conditions are satised, we characterize all admissible
controllers that satisfy the norm condition. The proofs of these results will be given in
the later sections. Let
opt := min fkTzw k1 : K (s) admissibleg, i.e., the optimal level.
Then, clearly,
must be greater than
opt for the existence of suboptimal H1 controllers. In Section 16.9 we will brie
y discuss how to nd an admissible K to minimize
kTzw k1 . Optimal H1 controllers are more dicult to characterize than suboptimal
ones, and this is one major dierence between the H1 and H2 results. Recall that
similar dierences arose in the norm computation problem as well.
The H1 solution involves the following two Hamiltonian matrices:
"
A
;2 B1 B1 ; B2 B2 ;
H1 :=
;C1 C1
;A
"
A
;2 C1 C1 ; C2 C2 :
J1 :=
;B1 B1
;A
The important dierence here from the H2 problem is that the (1,2)-blocks are not sign
denite, so we cannot use Theorem 13.7 in Chapter 13 to guarantee that H1 2 dom(Ric)
or Ric(H1 ) 0. Indeed, these conditions are intimately related to the existence of
H1 suboptimal controllers. Note that the (1,2)-blocks are a suggestive combination of
expressions from the H1 norm characterization in Chapter 4 (or bounded real ARE
in Chapter 13) and from the H2 synthesis of Chapter 14. It is also clear that if
approaches innity, then these two Hamiltonian matrices become the corresponding H2
control Hamiltonian matrices. The reasons for the form of these expressions should
become clear through the discussions and proofs for the following theorem.
Theorem 16.4 There exists an admissible controller such that kTzw k1 <
i the
following three conditions hold:
(i) H1 2 dom(Ric) and X1 := Ric(H1) 0;
(ii) J1 2 dom(Ric) and Y1 := Ric(J1 ) 0;
(iii) (X1 Y1 ) <
2 .
Moreover, when these conditions hold, one such controller is
"
^
Ksub (s) := A1 ;Z1 L1
F1
0
where
The H1 controller displayed in Theorem 16.4, which is often called the central
controller or minimum entropy controller, has certain obvious similarities to the H2
controller as well as some important dierences. Although not as apparent as in the
H2 case, the H1 controller also has an interesting separation structure. Furthermore,
412
Theorem 16.5 If conditions (i) to (iii) in Theorem 16.4 are satised, the set of all
admissible controllers such that kTzw k1 <
equals the set of all transfer matrices from
y to u in
u
y
M1
A^1 ;Z1 L1 Z1 B2
6
M1 (s) = 4 F1
0
I
I
0
;C2
3
7
5
As in the H2 case, the suboptimal controllers are parameterized by a xed linearfractional transformation with a free parameter Q. With Q = 0 (at the \center" of the
set kQk1 <
), we recover the central controller Ksub (s).
(16:1)
Let x denote the state of G with respect to a given input w, and then we can dierentiate
x(t) X1 x(t):
d
dt (x X1 x) = x_ X1 x + x X1 x_
413
d
2
2
2
2
;2
2
2
dt (x X1 x) = ;kz k +
kwk ;
kw ;
B1 X1 xk + ku + B2 X1 xk : (16:2)
Assume x(0) = x(1) = 0, w 2 L2+ , and integrate (16.2) from t = 0 to t = 1:
kz k22 ;
2 kwk22 = ku + B2 X1 xk22 ;
2 kw ;
;2B1 X1 xk22 = kvk22 ;
2 krk22 (16:3)
where
v := u + B2 X1 x;
r := w ; ;2 B1 X1 x:
(16:4)
With these new dened variables, the closed-loop system can be expressed as two interconnected subsystems below:
2
6
4
and
x_
z
r
A F1
;1 B1 B2
7 6
C1F1
0
D12
5=4
;
1
;
B1 X1 I
0
32
x
r
u
6
4
x_
Atmp B1 B2
7 6
=
v 5 4 ;F1 0 I
y
C2 D21 0
32
76
54
76
54
x
w
v
AF1 := A + B2 F1
C1F1 := C1 + D12 F1
7
5
Atmp := A + ;2 B1 B1 X1
7
5
z
r
w
P
XXXXX v
XX
r
Gtmp
-
414
where
"
P := P11 P12
P21 P22
and
AF1
;1B1 B2
6
= 4 C1F1
0
D12
;
;1B1 X1 I
0
2
3
7
5
(16:5)
Atmp B1 B2
6
Gtmp = 4 ;F1 0 I 75 :
(16:6)
C2 D21 0
The equality (16.3) motivates the change of variables to r and v as in (16.4), and these
variables provide the connection between Tzw and Tvr . Note that Tz(
w) =
;1 Tzw and
Tv(
r) =
;1Tvr . It is immediate from equality (16.3) that kTzw k1
i kTvr k1
.
While this is the basic idea behind the proof of Lemma 16.8 below, the details needed
for strict inequality and internal stability require a bit more work.
Note that wworst :=
;2 B1 X1 x is the worst disturbance input in the sense that
it maximizes the quantity kz k22 ;
2 kwk22 in (16.3) for the minimizing value of u =
;B2 X1 x; that is, the u making v = 0 and the w making r = 0 are values satisfying a
saddle point condition. (See Section 17.8 for more interpretations.) It is also interesting
to note that wworst is the optimal strategy for w in the corresponding LQ game problem
(see the dierential game problem in the last chapter). Equation (16.3) also suggests
that u = ;B2 X1 x is a suboptimal control for a full information (FI) problem if the
state x is available. This will be shown later. In terms of the OE problem for Gtmp , the
output being estimated is the optimal FI control input F1 x and the new disturbance r
is oset by the \worst case" FI disturbance input wworst .
Notice the structure of Gtmp : it is an OE problem. We will show below that the
output feedback can indeed be transformed into the OE problem. To show this we rst
need to prove some preliminary facts.
Lemma 16.6 Suppose H1 2 dom(Ric) and X1 = Ric(H1). Then AF1 = A + B2F1
is stable i X1 0.
Proof. Re-arrange the Riccati equation for X1 and use the denition of F1 and
C1F1 to get
"
# "
C1F1
C1F1
= 0:
(16:7)
;
;1 B1 X1
;
;1B1 X1
Since H1 2 dom(Ric), (AF1 +
;2 B1 B1 X1 ) is stable and hence (B1 X1 ; AF1 ) is
detectable. Then from standard results on Lyapunov equations (see Lemma 3.19), AF1
is stable i X1 0.
2
AF1 X1 + X1 AF1 +
415
This suggests that P might be inner when X1 0, which is veried by the following
lemma.
follows from Lemma 13.29 upon noting that the observability Gramian of P is X1 (see
(16.7)) and
"
#"
# "
#
;1 B1
C1F1
0 I
+
X1 = 0:
0
D12
;
;1B1 X1
B2
Finally, the state matrix for P21;1 is (AF1 +
;2 B1 B1 X1 ), which is stable by denition.
Thus, P21;1 2 RH1 .
2
The following lemma connects these two systems Tzw and Tvr , which is the central
part of the separation argument in Section 16.8.2. Recall that internal and inputoutput stability are equivalent for admissibility of K in the output feedback problem
by Corollary 16.3 .
Lemma 16.8 Assume H1 2 dom(Ric) and X1 = Ric(H1) 0. Then K is admissible for G and kTzw k1 <
i K is admissible for Gtmp and kTvr k1 <
.
Proof. We may assume without loss of generality that the realization of K is stabilizable and detectable. Recall from Corollary 16.3 that internal stability for Tzw is
equivalent to Tzw 2 RH1 . Similarly since
"
# "
#"
#
Atmp ; I B1 = A ; I B1
I
0
C2
D21
C2
D21
;2 B1 X1 I
A ; I B2 = det(A + B F ; I ) 6= 0
det tmp
tmp
2 1
;F1
I
for all Re() 0 by the stability of Atmp + B2 F1 , we have that
"
Atmp ; I B2
;F1
I
has full column rank. Hence by Lemma 16.1 the internal stability of Tvr , i.e., the internal
stability of the subsystem consisting of Gtmp and controller K , is also equivalent to
Tvr 2 RH1 . Thus internal stability is equivalent to input-output stability for both
416
G and Gtmp . This shows that K is an admissible controller for G if and only if it
is admissible for Gtmp . Now it follows
from
Theorem
16.2
and Lemma 16.7 along
with the above block diagram that
Tz(
w)
1 < 1 i
Tv(
r)
1 < 1 or, equivalently,
kTzw k1 <
i kTvr k1 <
.
2
From the previous analysis, it is clear that to solve the output feedback problem we
need to show
(a) H1 2 dom(Ric) and X1 = Ric(H1 ) 0;
(b) kTvr k1 <
.
To show (a), we need to solve a FI problem. The problem (b) is an OE problem which
can be solved by using the relationship between FC and OE problems in Chapter 12,
while the FC problem can be solved by using the FI solution through duality. So in the
sections to follow, we will focus on these special problems.
z
y
with
w
A
6
6
G(s) = 66 " C1 #
I
4
B1
"
0
0
B2
7
D 7
" 12# 7 :
7
0 5
0
I
0
The H1 problem corresponding to this setup again is not, strictly speaking, a special
case of the output feedback problem because it does not satisfy all of the assumptions.
In particular, it should be noted that for the FI (and FC in the next section) problem,
internal stability is not equivalent to Tzw 2 RH1 since
2
6
4
A" ; I#
I
0
B
" 1#
0
3
7
5
417
can never have full row rank, although this presents no diculties in solving this problem. We simply must remember that in the FI case, K admissible means internally
stabilizing, not just Tzw 2 RH1 .
We have seen that in the H2 FI case, the optimal controller uses just the state x even
though the controller is provided with full information. We will show below that, in the
H1 case, a suboptimal controller exists which also uses just x. This case could have
been restricted to state feedback, which is more traditional, but we believe that, once
one gets outside the pure H2 setting, the full information problem is more fundamental
and more natural than the state feedback problem.
One setting in which the full information case is more natural occurs when the
parameterization of all suboptimal controllers is considered. It is also appropriate when
studying the general case when D11 6= 0 in the next chapter or when H1 optimal
(not just suboptimal) controllers are desired. Even though the optimal problem is not
studied in detail in this book, we want the methods to extend to the optimal case in a
natural and straightforward way.
The assumptions relevant to the FI problem which are inherited from the output
feedback problem are
(i) (C1 ; A) is detectable;
(ii) (A; B2 ) is stabilizable;
h
C1 D12 = 0 I .
(iii) D12
Assumptions (iv) and the second part of (ii) for the general output feedback case have
been eectively strengthened because of the assumed structure for C2 and D21 .
Theorem 16.9 There exists an admissible controller K (s) for the FI problem such that
kTzw k1 <
if and only if H1 2 dom(Ric) and X1 = Ric(H1 ) 0. Furthermore,
if these conditions are satised, then the equivalence class of all admissible controllers
satisfying kTzw k1 <
can be parameterized as
h
i
F1 ;
;2 Q(s)B1 X1 Q(s)
K (s) =
(16:8)
where Q 2 RH1 , kQk1 <
.
418
h
Tzw = AF1 B1
C1F1 0
AF1 = A + B2 F1
C1F1 = C1 + D12 F1 :
(16:9)
Furthermore, since Tzw =
P11 and P11 P11 = I ; P21 P21 by P P = I , we have
kP11 k1 < 1 and kTzw k1 =
kP11 k1 <
. The further condition that X1 0 is
equivalent, by Lemma 16.6, to this K stabilizing Tzw .
intuitive way to see why all the equivalence classes of controllers can be parameterized
in the form of (16.8). Recall the following equality from equation (16.3):
Proof. ()) For simplicity, in the proof to follow we assume that the system is normalized such that
= 1. Further, we will show that we can, without loss of generality,
strengthen the assumption on (C1 "; A) from detectable
to observable. Suppose there ex#
^ B^1 B^2
A
ists an admissible controller K^ = ^ ^ ^ such that kTzw k1 < 1. If (C1 ; A) is
C D1 D2
"
x
detectable but not observable, then change coordinates for the state of G to 1 with
x2
x2 unobservable, (C11 ; A11 ) observable, and A22 stable, giving the following closed-loop
state equations:
2
6
6
6
6
6
6
4
x_ 1
x_ 2
x^_
z
u
A11
7 6
7 6 A21
7 6
7=6 B
^
7 6 11
7 6
5 4 C11
D^ 11
A22
B^12
0
^
D12
0
0
A^
0
C^
B11 B21
B12 B22
B^2 0
0 D12
^
D2 0
32
76
76
76
76
76
76
54
x1
x2 777
x^ 77 :
w 75
u
419
If we take a new plant Gobs with state x1 and output z and group the rest of the
equations as a new controller Kobs with the state made up of x2 and x^, then
2
A11
6
6
Gobs (s) = 66 "C11#
I
4
B11
"
0
0
B21
D
" 12#
0
0
3
7
7
7
7
5
still satises the assumptions of the FI problem and is stabilized by Kobs with the
closed-loop H1 norm kTzw k1 < 1 where
2
X^1 = Ric
then
"
; B21 B21
A11
B11 B11
;
;C11 C11
;A11
"
^
Ric(H1 ) = X1 = X1 0 0
0
exists for G. We can therefore assume without loss of generality that (C1 ; A) is observable.
We shall suppose that there exists an admissible controller such that kTzw k1 <
1. But note that the existence of an admissible controller such that kTzw k1 < 1 is
equivalent to that the admissible controller makes sup kz k2 < 1; hence, it is necessary
w2BL2+
that
kz k2 < 1
sup umin
2L
w2BL2+
2+
since the latter is always no greater than the former by the fact that the set of signals
u generated from admissible controllers is a subset of L2+ . But from Theorem 15.9,
the latter is true if and only if H1 2 dom(Ric) and X1 = Ric(H1 ) > 0. Hence the
necessity is proven.
(() Suppose H1 2 dom(Ric) and X1 = Ric(H1 ) 0 and suppose K (s) is an
admissible controller such that kTzw k1 < 1. Again change variables to v := u ; F1 x
420
z
r
w
"
P = P11 P12
P21 P22
- Tvr
AF1 B1 B2
= 64 C1F1
0 D12
;B1 X1 I 0
3
7
5
By Lemma 16.7, P is inner and P21;1 2 RH1 . By Theorem 16.2 the system is internally
stable and kTzw k1 < 1 i Tvr 2 RH1 and kTvr k1 < 1. Now denote Q := Tvr , then
v = Qr and
"
x
v = u ; F1 x = K ; F1 0
w
"
#
h
i x
r = w ; B1 X1 x = ;B1 X1 I
:
w
Hence we have
K ; F1 0
or
i
"
i
x = Qh
;B1 X1 I
w
i
"
x
w
K ; F1 0
= Q ;B1 X1 I :
Thus
K (s)
= F1 ; Q(s)B1 X1 Q(s) ; Q 2 RH1 ; kQk1 < 1:
Remark 16.2 It should be emphasized that the set of controllers given above does
not parameterize all controllers although it is sucient for the purpose of deriving the
output feedback results, and that is why
class" is used. It is clear that
h \equivalence
i
there is a suboptimal controller K1 = F1 0 with F1 6= F1 ; however, there is no
choice of Q such that K1 belongs to the set. Nevertheless, this problem will not occur
in output feedback case.
~
The following theorem gives all full information controllers.
Theorem 16.10 Suppose the condition in Theorem 16.9 is satised; then all admissible
controllers satisfying kTzw k1 <
can be parameterized as K = F` (MFI ; Q):
u
y
MFI
-
421
2
MFI (s) =
6
6
6
6
6
4
A + B2 F1
"
;I
h
"
0 B1
F1 0
i
i
I
0
;
2
;
B1 X1 I
B2
I
0
3
7
7
7
7
7
5
Proof. We only need to show that F`(MFI ; Q) with kQ2k1 <
parameterizes all
FI H1 suboptimal controllers. To show that, we shall make a change of variables as
before:
v = u + B2 X1 x;
r = w ;
;2 B1 X1 x:
Then the system equations can be written as follows:
2
3 2
32
3
x_
AF1
;1 B1 B2
x
6
7 6
C1F1
0
D12 75 64
w 75
4 z 5=4
r
;
;1B1 X1 I
0
v
and
2
3
2
3
Atmp
B1 B2 2 3
x_
6
7 x
;
F
0 # I 77 6 7
1
6
7 6
6
"
#
"
4 v 5=6
74 r 5
I
0
4
0 5 u
y
;2
B1 X1
z
r
w
XXXXX v
X r
G^ FI
u
- K
P
w
422
where P is as given in equation (16.5) and
2
6
6
G^ FI = 66
4
"
Atmp
;F1
I
;
2
B1 X1
B1
#
"
0
0
B2
7
I 77 :
7
So from Theorem 16.2 and Lemma 16.8, we conclude that K is an admissible controller
for G and kTzw k1 <
i K is an admissible" controller for# G^ FI and kTvr k1 <
.
h
I
6 Atmp + B2 F1 + L
;
2
B1 X1
6
6
M = 66
F1
"
#
6
I
4
; ;2
B1 X1
;L B2
0
3
7
7
7
7
7
7
5
With this parameterization of all controllers, the transfer matrix from r to v is Tvr =
N = 64
"
0
0
I
0
3
7
5
and Tvr = F` (N; ) = 2 . Hence kTvr k1 <
if and only if k2 k1 <
. This
implies
thati all FI H1 controllers can be parameterized as K = F`(M; ) with =
h
1 2 2 RH1 ; k2 k1 <
and
h
A + 2B2 F1
6
6
M = 66 " F1
4
; ;2 I
B1 X1
Now let
; B2 F1 ;B1
#
1 = F1 ;
;2 Q2 B1 X1 + Q1;
Then it is easy to show that F` (M; ) = F`(MFI ; Q).
B2
7
I 77 :
7
0
2 = Q2 :
423
A B1
6
6
G(s) = 64 C1 0
C2 D21
h
h
h
I 0
0 I
0 0
i 3
i 7
7
7
i 5
This problem is dual to the Full Information problem. The assumptions that the FC
problem inherits from the output feedback problem are just the dual of those in the FI
problem:
(i) (A; B1 ) is stabilizable;
(ii) (C2 ; A) is detectable;
"
(iv)
"
B1 D = 0 .
I
D21 21
Theorem 16.11 There exists an admissible controller K (s) for the FC problem such
that kTzw k1 <
if and only if J1 2 dom(Ric) and Y1 = Ric(J1 ) 0. Moreover,
if these conditions are satised then an equivalence class of all admissible controllers
satisfying kTzw k1 <
can be parameterized as
"
As expected, the condition in Theorem 16.11 is the same as that in (ii) of Theorem 16.4.
A B1 B2
G(s) = 64 C1 0 D12
C2 I 0
3
7
5
This problem inherits the same assumptions (i)-(iii) as in the FI problem, but for internal
stability we shall add that A ; B1 C2 is stable. With this assumption, it is easy to check
that the condition in Lemma 16.1 is satised so that the internal stability is again
equivalent to Tzw 2 RH1 , as in the output feedback case.
424
Theorem 16.12 There exists an admissible controller for the DF problem such that
kTzw k1 <
if and only if H1 2 dom(Ric) and X1 = Ric(H1 ) 0. Moreover, if
these conditions are satised then all admissible controllers satisfying kTzw k1 <
can
be parameterized as the set of all transfer matrices from y to u in
u
y
M1
-
A + B2 F1 ; B1 C2 B1 B2
6
M1 (s) = 4
F1
0 I
;
2
;C2 ;
B1 X1 I 0
3
7
5
v = u + B2 X1 x;
r = w ; ;2 B1 X1 x:
and
x_
z
r
2
6
4
AF1
;1 B1 B2
7 6
C1F1
0
D12
5=4
;
1
;
B1 X1 I
0
3
x_
Atmp
B1 B2
v 75 = 64
;F1
0 I
;
2
y
C2 +
B1 X1 I 0
32
76
54
32
76
54
x
w
v
3
3
7
5
x
r 75 :
u
425
w
P
XXXX v
X
XX
r
G^ DF
-
where
and
AF1
;1B1 B2
P = 64 C1F1
0
D12
;
1
0
;
B1 X1 I
2
3
7
5
B1 B2
Atmp
6
^
GDF = 4
;F1
0 I 75 :
C2 +
;2 B1 X1 I 0
Since Atmp ; B1 (C2 +
;2 B1 X1 ) = A ; B1 C2 and Atmp + B2 F1 are stable, the rank
conditions of Lemma 16.1 for system G^ DF are satised. So from Theorem 16.2 and
Lemma 16.7, we conclude that K is an admissible controller for G and kTzw k1 <
i K
is an admissible controller for G^ DF and kTvr k1 <
. Now it is easy to see by comparing
this formula with the controller parameterization in Theorem 12.8 that F` (M1 ; Q)
with Q 2 RH1 (no norm constraint) parameterizes all stabilizing controllers for G^ DF ;
however, simple algebra shows that Tvr = F` (G^ DF ; F`(M1 ; Q)) = Q. So kTvr k1 <
i kQk1 <
, and F` (M1 ; Q) with Q 2 RH1 and kQk1 <
parameterizes all
suboptimal controllers for G.
2
A B1 B2
6
G(s) = 4 C1 0 I
C2 D21 0
3
7
5
426
This problem is dual to DF, just as FC was to FI. Thus the discussion of the DF problem
is relevant here, when appropriately dualized. The OE assumptions are
(i) (A; B1 ) is stabilizable and A ; B2 C1 is stable;
(ii) (C2 ; A) is detectable;
"
(iv)
"
B1 D = 0 .
D21 21
I
Assumption (i), together with (iv), imply that internal stability is again equivalent to
Tzw 2 RH1 , as in the output feedback case.
Theorem 16.13 There exists an admissible controller for the OE problem such that
kTzw k1 <
if and only if J1 2 dom(Ric) and Ric(J1 ) 0. Moreover, if these
conditions are satised then all admissible controllers satisfying kTzw k1 <
can be
parameterized as the set of all transfer matrices from y to u in
u
y
M1
-
3
7
5
427
Proof. Let K be an admissible controller for which kTzw k1 <
. The controller
K [C2 D21 ] solves the FI problem; hence from Theorem 16.9, H1 2 dom(Ric) and
X1 := Ric(H1 ) 0. Condition (ii) follows by the dual argument.
2
We would also expect some condition beyond these two, and that is provided by
(iii), which is an elegant combination of elements from FI and FC. Note that all the
conditions of Theorem 16.4 are symmetric in H1 , J1 , X1 , and Y1 , but the formula
for the controller is not. Needless to say, there is a dual form that can be obtained by
inspection from the above formula. For a symmetric formula, the state equations above
;1 and put in descriptor form. A simple substitution
can be multiplied through by Z1
from the Riccati equation for X1 will then yield a symmetric, though more complicated,
formula:
(I ;
;2 Y1 X1 )x^_ = As x^ ; L1 y
(16.10)
u = F1 x^
(16.11)
where As := A + B2 F1 + L1 C2 +
;2Y1 A X1 +
;2 B1 B1 X1 +
;2Y1 C1 C1 .
To emphasize its relationship to the H2 controller formulae, the H1 controller can
be written as
x^_ = Ax^ + B1 w^worst + B2 u + Z1 L1 (C2 x^ ; y)
428
u = F1 x^;
It has been shown from Lemma 16.14 that conditions (i) and (ii) are necessary for
kTzw k1 <
. Hence we only need to show that if conditions (i) and (ii) are satised,
condition (iii) is necessary and sucient for kTzw k1 <
. As in section 16.3, we dene
new disturbance and control variables
r := w ;
;2 B1 X1 x;
Then
"
B B
A
v = 6 tmp 1 2
0 I
4 ;F1
y
C2 D21 0
#
z
y
3
7
5
"
w
v := u + B2 X1 x:
"
r =G
r
tmp
u
u
Atmp := A + ;2 B1 B1 X1 :
v
y
r
Gtmp
-
Recall from Lemma 16.8 that K is admissible for G and kTzw k1 <
i K is admissible
for Gtmp and kTvr k1 <
.
While Gtmp has the form required for the OE problem, to actually use the OE
results, we will need to verify that Gtmp satises the following assumptions for the OE
problem:
429
"
B1 D = 0 .
(iv)
D21 21
I
Assumption (iv) and that (Atmp ; B1 ) is stabilizable follow immediately from the corresponding assumptions for Theorem 16.4. The stability of Atmp + B2 F1 follows from
the denition of H1 2 dom(Ric). The following lemma gives conditions for assumption
(ii) to hold. Of course, the existence of an admissible controller for Gtmp immediately
implies that assumption (ii) holds. Note that the OE Hamiltonian matrix for Gtmp is
"
;2
Jtmp := Atmp
F1 F1 ; C2 C2 :
;B1 B1
;Atmp
Lemma 16.15 If Jtmp 2 dom(Ric) and Ytmp := Ric(Jtmp) 0, then (C2 ; Atmp) is
detectable.
Proof. The lemma follows from the dual to Lemma 16.6, which gives that (Atmp ;
Ytmp C2 C2 ) is stable.
Proof of Theorem 16.4 (Suciency) Assume the conditions (i) through (iii) in the
theorem
" statement hold.
# Using the Riccati equation for X1 , one can easily verify that
;2X1
I
;
T :=
provides a similarity transformation between Jtmp and J1 , i.e.,
0
I
T ;1Jtmp T = J1 . So
"
"
;2
X; (Jtmp ) = T X; (J1 ) = T Im I = Im I ;
X1 Y1
Y1
Y1
and (X1 Y1 ) <
2 implies that Jtmp 2 dom(Ric) and Ytmp := Ric(Jtmp ) = Y1 (I ;
;2 X1 Y1 );1 = Z1 Y1 0. Thus by Lemma 16.15 the OE assumptions hold for Gtmp ,
and by Theorem 16.13 the OE problem is solvable. From Theorem 16.13 with Q = 0,
one solution is
"
but this is precisely Ksub dened in Theorem 16.4. We conclude that Ksub stabilizes
Gtmp and that kTvr k1 <
. Then by Lemma 16.8, Ksub stabilizes G and that kTzw k1 <
.
430
(Necessity) Let K be an admissible controller for which kTzw k1 <
. By Lemma 16.14,
H1 2 dom(Ric), X1 := Ric(H1 ) 0, J1 2 dom(Ric), and Y1 := Ric(J1) 0.
From Lemma 16.8, K is admissible for Gtmp and kTvr k1 <
. This implies that the OE
assumptions hold for Gtmp and that the OE problem is solvable. Therefore, from Theorem 16.13 applied to Gtmp , we have that Jtmp 2 dom(Ric) and Ytmp = Ric(Jtmp) 0.
Using the same similarity transformation formula as in the suciency part, we get that
Ytmp = (I ;
;2 Y1 X1 );1 Y1 0. We shall now show that Ytmp 0 implies that
(X1 Y1 ) <
2 . We shall consider two cases:
Y1 is nonsingular: in this case Ytmp 0 implies that I ;
;2 Y11=2 X1 Y11=2 > 0.
So (Y11=2 X1 Y11=2 ) <
2 or (X1 Y1 ) <
2 .
Y1 is singular: there is a unitary matrix U such that
#
"
Y
11 0
U
Y =U
1
0 0
with Y11 > 0. Let UX1 U be partitioned accordingly,
UX1 U =
"
X11 X12 :
X21 X22
Ytmp = U
"
We now see exactly why the term involving w^worst appears and why the \observer"
gain is Z1 L1. Both terms are consequences of estimating the optimal Full Information
(i.e., state feedback) control gain. While an analogous output estimation problem arises
in the H2 output feedback problem, the resulting equations are much simpler. This is
because there is no \worst-case" disturbance for the H2 Full Information problem and
because the problem of estimating any output, including the optimal state feedback, is
equivalent to state estimation.
We now present a separation interpretation for H1 suboptimal controllers. It will
be stated in terms of the central controller, but similar interpretations could be made
for the parameterization of all suboptimal controllers (see the proofs of Theorems 16.4
and 16.5).
The H1 output feedback controller is the output estimator of the full information control law in the presence of the \worst-case" disturbance wworst .
Note that the same statement holds for the H2 optimal controller, except that
wworst = 0.
431
From Lemma 16.8, the set of all admissible controllers for G such that kTzw k1 <
equals the set of all admissible controllers for Gtmp such that kTvr k1 <
. Apply
Theorem 16.13.
2
In this section, we will discuss, without proof, the behavior of the H1 suboptimal
solution as
varies, especially as
approaches the inmal achievable norm, denoted by
opt . Since Theorem 16.4 gives necessary and sucient conditions for the existence of
an admissible controller such that kTzw k1 <
,
opt is the inmum over all
such that
conditions (i)-(iii) are satised. Theorem 16.4 does not give an explicit formula for
opt ,
but, just as for the H1 norm calculation, it can be computed as closely as desired by a
search technique.
Although we have not focused on the problem of H1 optimal controllers, the assumptions in this book make them relatively easy to obtain in most cases. In addition to
describing the qualitative behavior of suboptimal solutions as
varies, we will indicate
why the descriptor version of the controller formulae from Section 16.8.1 can usually
provide formulae for the optimal controller when
=
opt . Most of these results can be
obtained relatively easily using the machinery that is developed in the previous sections.
The reader interested in lling in the details is encouraged to begin by strengthening
;1
assumption (i) to controllable and observable and considering the Hamiltonians for X1
;
1
and Y1 .
As
! 1, H1 ! H2 , X1 ! X2 , etc., and Ksub ! K2 . This fact is the result of
the particular choice for the central controller (Q = 0) that was made here. While it
could be argued that Ksub is a natural choice, this connection with H2 actually hints at
deeper interpretations. In fact, Ksub is the minimum entropy solution (see next section)
as well as the minimax controller for kz k22 ;
2kwk22 .
If
2
1 >
opt , then X1 (
1 ) X1 (
2 ) and Y1 (
1 ) Y1 (
2 ). Thus X1 and
Y1 are decreasing functions of
, as is (X1 Y1 ). At
=
opt , anyone of the three
conditions in Theorem 16.4 can fail. If only condition (iii) fails, then it is relatively
straightforward to show that the descriptor formulae for
=
opt are optimal, i.e., the
optimal controller is given by
;2 Y1 X1 )x^_ = As x^ ; L1 y
(I ;
opt
(16.12)
u = F1 x^
(16.13)
;2 Y1 A X1 +
;2 B1 B X1 +
;2 Y1 C C1 . See
where As := A + B2 F1 + L1 C2 +
opt
opt
1
opt
1
the example below.
The formulae in Theorem 16.4 are not well-dened in the optimal case because the
;2 X1 Y1 ) is not invertible. It is possible but far less likely that conditions
term (I ;
opt
(i) or (ii) would fail before (iii). To see this, consider (i) and let
1 be the largest
432
for which H1 fails to be in dom(Ric) because the H1 matrix fails to have either the
stability property or the complementarity property. The same remarks will apply to (ii)
by duality.
If complementarity fails at
=
1 , then (X1 ) ! 1 as
!
1 . For
<
1 ,
H1 may again be in dom(Ric), but X1 will be indenite. For such
, the controller
u = ;B2 X1 x would make kTzw k1 <
but would not be stabilizing. See part 1) of
the example below. If the stability property fails at
=
1 , then H1 62 dom(Ric) but
Ric can be extended to obtain X1 and the controller u = ;B2 X1 x is stabilizing and
makes kTzw k1 =
1 . The stability property will also not hold for any
1 , and
no controller whatsoever exists which makes kTzw k1 <
1 . In other words, if stability
breaks down rst, then the inmum over stabilizing controllers equals the inmum over
all controllers, stabilizing or otherwise. See part 2) of the example below. In view of
this, we would typically expect that complementarity would fail rst.
Complementarity failing at
=
1 means (X1 ) ! 1 as
!
1 , so condition (iii)
would fail at even larger values of
, unless the eigenvectors associated with (X1 ) as
!
1 are in the null space of Y1 . Thus condition (iii) is the most likely of all to fail
rst. If condition (i) or (ii) fails rst because the stability property fails, the formulae
in Theorem 16.4 as well as their descriptor versions are optimal at
=
opt . This
is illustrated in the example below for the output feedback. If the complementarity
condition fails rst, (but (iii) does not fail), then obtaining formulae for the optimal
controllers is a more subtle problem.
Example 16.1 Let an interconnected dynamical system realization be given by
2
G(s) =
6 "
6
6
6
6
4
a
1
0
#
h
c2
1 0
0
0 1
b2 #
"
0
1
0
3
7
7
7
7
7
5
with jc2 j jb2 j > 0. Then all assumptions for the output feedback problem are satised
and
"
"
1;b22
2 #
1;c22
2 #
a
a
2
2
H1 =
; J1 =
:
;1 ;a
;1 ;a
The eigenvalues of H1 and J1 are given, respectively, by
( p
)
( p
)
2 + b2 )
2 ; 1
2 + c2 )
2 ; 1
(
a
(
a
2
2
(H ) =
; (J ) =
:
1
X; (H1 ) = Im
"
433
(a2 + b22 ) 2 ; 1 ; a = 0
when 2 =
> a2 +1 b2 .
2
b22
=
X1 = p 2 2
2
(a + b2 )
; 1 ; a
2
2 2
2
A+B2 F1 = ; a + b2
b2(
a2 ;+ 1b2 )
; 1 =
2
{ Suppose full information (or states) are available for feedback and let
u = F1 x:
Then the closed-loop transfer matrix is given by
2
Tzw =
"
A + B2 F1 B1 = 66
6
4
C1 + D12 F1 0
p
a+b22
(a2 +b22 )
2 ;1
;
b22
2 ;1
3
2
4
;b2
1 0
i 3
7
7
7
5
and Tzw is stable for all
2 > b12 and is not stable for a2 +1 b2 <
2 < b12 . Fur2
2
2
1 and
2 6= 1 .
thermore, it can be shown that kTzw k <
for all
2 > a2 +
b22
b22
It is clear that the optimal H1 norm is b122 but is not achievable.
434
{ Suppose the states are not available; then output feedback must be considered. Note that if
2 > b12 , then H1 2 dom(Ric), J1 2 dom(Ric), and
2
X1 = p
(a2 + b22 ) 2 ; 1 ; a
>0
> 0:
(a2 + c22 )
2 ; 1 ; a
Hence conditions (i) and (ii) in Theorem 16.4 are satised, and need to check
condition (iii). Since
Y1 = p
(X1 Y1 ) =
2p
;
( (a2 + b22 )
2 ; 1 ; a
)( (a2 + c22 )
2 ; 1 ; a
)
p
X1 = p
>0
Y1 = p
>0
(a2 + b22) 2 ; 1 ; a
1 .
for
2 > a2 +
b2
(a2 + c22 ) 2 ; 1 ; a
X; (H1 ) = Im ;a
435
2
a + ba2
6 "
Tzw = 64
1 0
0
b2
a
i 3
7
7
5
2 RH1
(X1 Y1 ) =
2
<
2
( (a2 + b22 )
2 ; 1 ; a
)2
p
2
2
> a +b22b2 + a > p 21 2 :
a + b2
2
So condition (iii) of Theorem 16.4 will fail before either (i) or (ii) fails.
In both a > 0 and a < 0 cases, the optimal
for the output feedback is given by
p
2
2
opt = a +b22b2 + a
2
if jb2 j = jc2 j; and the optimal controller given by the descriptor formula in equations (16.12) and (16.13) is a constant. In fact,
uopt = ; q
opt
2 ; 1 ; a
opt
(a2 + b22 )
opt
y:
436
Let T be a transfer matrix with kT k1 <
. Then the entropy of T (s) is dened by
2Z 1 ;
ln det I ;
;2T (j!)T (j!) d!:
I (T;
) = ; 2
;1
;1 i
and I (T;
) 0, where i (T (j!)) is the ith singular value of T (j!). It is also easy to
show that
1 Z 1 X 2 (T (j!)) d! = kT k2 :
lim
I
(
T;
)
=
i
2
!1
2
;1 i
Thus the entropy I (T;
) is in fact a performance index measuring the tradeo between
the H1 optimality (
! kT k1 ) and the H2 optimality (
! 1).
It has been shown in Glover and Mustafa [1989] that the central controller given in
Theorem 16.4 is actually the controller that satises the norm condition kTzw k1 <
and minimizes the following entropy:
2Z 1
; 2
;1
Therefore, the central controller is also called the minimum entropy controller (maximum entropy controller if the entropy is dened as I~(T; ) = ;I (T; )).
Although [Whittle, 1981] treats a nite horizon, discrete time, stochastic control problem, his separation result has a clear interpretation for the present innite horizon,
437
then by the completion of the squares and by the saddle point argument of Section 16.3,
where u is not constrained to be a function of the measurements (FI case), we obtain
S+ (x) = x X1 x:
The past stress, S; (x), is a function of the past inputs and observations, u(t); y(t) for
;1 < t < 0, and of the present state, x, and is produced by the worst case disturbance,
w, that is consistent with the given data:
S; (x) := sup(kP; z k22 ;
2kP; wk22 ):
In order to evaluate S; we see that w can be divided into two components, D21 w
? w , with x dependent only on D21
? w (since B1 D21 = 0) and D21 w = y ; C2 x.
and D21
The past stress is then calculated by a completion of the square and in terms of a lter
output. In particular, let x be given by the stable dierential equation
x_ = Ax + B2 u + L1 (C2 x ; y) + Y1 C1 C1 x with x(;1) = 0:
Then it can be shown that the worst case w is given by
? w = D? B Y ;1 (x(t) ; x(t)) for t < 0
D21
21 1 1
and that this gives, with e := x ; x,
S; (x) = ;
2e(0) Y1;1 e(0) ;
2 kP; (y ; C2 x)k22 + kP;(C1 x)k22 + kP; uk22:
The worst case disturbance will now reach the value of x to maximize the total
stress, S; (x) + S+ (x) , and this is easily shown to be achieved at the current state of
x^ = Z1 x(0):
The denitions of X1 and Y1 can be used to show that the state equations for
the central controller can be rewritten with the state x := Z1 x^ and with x as dened
above. The control signal is then
u = F1 x^ = F1 Z1 x:
The separation is between the evaluation of future stress, which is a control problem
with an unconstrained input, and the past stress, which is a ltering problem with
known control input. The central controller then combines these evaluations to give a
worst case estimate, x^, and the control law acts as if this were the perfectly observed
state.
438
such that
and
"
X11 2 R2nn
X12
#
"
such that
and
(iii)
"
"
Y11 2 R2nn
Y12
#
439
Remark
16.4 It is simple to show that if X11 and Y11 are nonsingular and if X1 =
X12 X1;11 and Y1 = Y12 Y1;11 , then condition (iii) in the above theorem is equivalent to
X1 0, Y1 0, and (Y1 X1 )
2 . So in this case, the conditions for the existence
of an optimal controller can be obtained from \taking the limit" of the corresponding
conditions in Theorem 16.4. Moreover, the controller given above is reduced to the
descriptor form given in equations (16.12) and (16.13).
~
440
17
442
z
y
w
where, as usual, G and K are assumed to be real rational and proper with K constrained
to provide internal stability. The controller is said to be admissible if it is real-rational,
proper, and stabilizing. Although we are taking everything to be real, the results
presented here are still true for the complex case with some obvious modications. We
will again only be interested in characterizing all suboptimal H1 controllers.
The realization of the transfer matrix G is taken to be of the form
2
A B1 B2
G(s) = 64 C1 D11 D12
C2 D21 0
3
7
5
"
A B
=
C D
which is compatible with the dimensions z (t) 2 Rp1 , y(t) 2 Rp2 , w(t) 2 Rm1 , u(t) 2
Rm2 , and the state x(t) 2 Rn . The following assumptions are made:
(A1) (A; B2 ) is stabilizable and (C2 ; A) is detectable;
(A2) D12 =
and D21 = 0 I ;
"
A ; j!I B2
C1
D12
"
A ; j!I B1
C2
D21
(A3)
(A4)
"
Assumption (A1) is necessary for the existence of stabilizing controllers. The assumptions in (A2) mean that the penalty on z = C1 x + D12 u includes a nonsingular,
normalized penalty on the control u, that the exogenous signal w includes both plant
disturbance and sensor noise, and that the sensor noise weighting is normalized and
nonsingular. Relaxation of (A2) leads to singular control problems; see Stroorvogel
[1990]. For those problems that have D12 full column rank and D21 full row rank but
do not satisfy assumption (A2), a normalizing procedure is given in the next section so
that an equivalent new system will satisfy this assumption.
443
Assumptions (A3) and (A4) are made for a technical reason: together with (A1)
they guarantee that the two Hamiltonian matrices in the corresponding H2 problem
belong to dom(Ric), as we have seen in Chapter 14. It is tempting to suggest that
(A3) and (A4) can be dropped, but they are, in some sense, necessary for the methods
presented in the last chapter to be applicable. A further discussion of the assumptions
and their possible relaxation will be discussed in Section 17.3.
The main result is now stated in terms of the solutions of the X1 and Y1 Riccati
equations together with the \state feedback" and \output injection" matrices F and L.
R := D1 D1 ;
"
"
2Im1 0 ; where D := [D D ]
1
11 12
0 0
#
"
2
R~ := D1 D1 ;
Ip1 0 ; where D1 := D11
D21
0 0
"
# "
#
h
i
A
0
B
;1 D C1 B
H1 :=
;
R
1
;C1 C1 ;A
;C1 D1
"
# "
#
h
i
A
0
C
;1 D1 B C
~
J1 :=
;
R
1
;B1 B1 ;A
;B1D1
X1 := Ric(H1)
F :=
L :=
"
h
"
Y1 := Ric(J1 )
Remark 17.1 In the above matrix partitioning, some matrices may not exist depending on whether D12 or D21 is square. This issue will be discussed further later. For the
time being, we shall assume all matrices in the partition exist.
~
Theorem 17.1 Suppose G satises the assumptions (A1){(A4).
(a) There exists an admissible controller K (s) such that jjF` (G; K )jj1 <
(i.e.
kTzw k1 <
) if and only if
444
; D1121
]);
(i)
> max( [D1111 ; D1112 ; ]; [D1111
(ii) H1 2 dom(Ric) with X1 = Ric(H1 ) 0;
(iii) J1 2 dom(Ric) with Y1 = Ric(J1 ) 0;
(iv) (X1 Y1 ) <
2 .
(b) Given that the conditions of part (a) are satised, then all rational internally
stabilizing controllers K (s) satisfying jjF` (G; K )jj1 <
are given by
where
K = F`(M1 ; Q)
A^ B^1 B^2 7
6
M1 = 64 C^1 D^ 11 D^ 12 75
C^2 D^ 21 0
(
2 I ; D1111 D );1 D1112 ; D1122 ;
D^ 11 = ;D1121 D1111
1111
D^ 12 2 Rm2 m2 and D^ 21 2 Rp2 p2 are any matrices (e.g. Cholesky factors) satisfying
= I ; D1121 (
2 I ; D D1111 );1 D ;
D^ 12 D^ 12
1111
1121
2
;
1
^
^
D21 D21 = I ; D1112 (
I ; D1111 D1111 ) D1112 ;
and
where
(Note that if D11 = 0 then the formulae are considerably simplied.)
In this case
1. in part (a), (i) becomes
> (D1121 ).
2. in part (b)
D^ 11 = ;D1122
= I ;
;2 D1121 D
D^ 12 D^ 12
1121
D^ 21 = I:
D^ 21
445
Case 2: D21 = I
In this case
1. in part (a), (i) becomes
> (D1112 ).
2. in part (b)
D^ 11 = ;D1122
=I
D^ 12 D^ 12
D^ 21 = I ;
;2 D D1112 :
D^ 21
1112
D^ 11 = ;D1122
=I
D^ 12 D^ 12
D^ 21 = I:
D^ 21
Let a given problem have the following diagram where zp (t) 2 Rp1 , yp (t) 2 Rp2 , wp (t) 2
Rm1 , and up (t) 2 Rm2 :
zp
yp
- Kp
wp
up
The plant P has the following state space realization with Dp12 full column rank and
Dp21 full row rank:
2
3
Ap Bp1 Bp2
P (s) = Cp1 Dp11 Dp12
Cp2 Dp21 Dp22
6
4
7
5
The objective is to nd all rational proper controllers Kp(s) which stabilize P and
jjF` (P; Kp )jj1 < . To solve this problem we rst transfer it to the standard one
treated in the last section. Note that the following procedure can also be applied to the
H2 problem (except the procedure for the case D11 6= 0).
446
Dp12 = Up
"
such that Up and U~p are square and unitary. Now let
zp = Up z; wp = U~p w; yp = R~p y; up = Rp u
and
K (s) = Rp Kp (s)R~p
G(s) =
"
2
6
6
4
2
=:
Then
6
4
"
0
U~p 0
P
(
s
)
0 R~p;1
0 Rp;1
Bp1 U~p
Bp2 Rp;1
Ap
Up Cp1 Up Dp11 U~p Up Dp12 Rp;1
R~p;1 Cp2 R~p;1Dp21 U~p R~p;1Dp22 Rp;1
Up
A B1 B2
C1 D11 D12
C2 D21 D22
D12 =
"
"
A B :
=
C D
h
7
5
7
7
5
D21 = 0 I ;
z
w
Suppose K (s) is a controller for G with D22 set to zero. Then the controller for D22 6= 0
is K (I + D22 K );1 . Hence there is no loss of generality in assuming D22 = 0.
447
We can even assume that D11 = 0. In fact, Theorem 17.1 can be shown by rst
transforming the general problem to the standard problem considered in the last chapter
using Theorem 16.2. This transformation is called loop-shifting. Before we go into the
detailed description of the transformation, let us rst consider a simple unitary matrix
dilation problem.
Lemma 17.2 Suppose D is a constant matrix such that kDk < 1; then
N=
"
;D
(I ; DD )1=2
(I ; D D)1=2
D
is a unitary matrix, i.e., N N = I .
This result can be easily veried, and the matrix N is called a unitary dilation of D
(of course, there are many other ways to dilate a matrix to a unitary matrix).
Consider again the feedback system shown at the beginning of this chapter; without
loss of generality, we shall assume that the system G has the following realization:
2
6
6
6
6
4
G = C1
with
"
C2
D12 =
"
B1
D1111 D1112
D1121
D1122
h
i
0 I
0
"
B2 #
0
3
7
7
7
7
5
; D21 = 0 I :
Suppose there is a stabilizing controller K such that kF` (G; K )k1 < 1. (Suppose we
have normalized
to 1). In the following, we will show how to construct a new system
transfer matrix M (s) and a new controller K~ such that
the D11
matrix for M (s) is zero,
and, furthermore, kF` (G; K )k1 < 1 if and only if
F`(M; K~ )
< 1. To begin with,
1
note that kF` (G; K )(1)k < 1 by the assumption
kF`(G; K )(1)k
Let
D1111
D1112
:
D1121 D1122 + K (1)
D1 2 X : D1111 D1112
< 1 :
D1121 D1122 + X
D1111 );1 D1111
D1112 and dene
For example, let D1 = ;D1122 ; D1121 (I ; D1111
"
#
D
D
1111
1112
~
D11 :=
;
D1121 D1122 + D1
448
~ K~ )
F`(G; K ) = F` (G; K~ + D1 ) = F` (G;
where
G~ =
6
6
6
6
4
2
=:
Let
K (s) = K~ (s) + D1
6
6
4
A + B2 D1 C2
"
C1 + D12 D1 C2
A~
C~1
C~2
C2
B~1
D~ 11
D~ 21
"
B1 + B2 D1 D21
D1111
D1112
D1121h D1122i+ D1
0 I
"
B2 #
0
3
7
7
7
7
5
B~2 7
D~ 12 75 :
0
;D~ 11
)1=2
(I ; D~ 11 D~ 11
N=
D~ 11 )1=2
(I ; D~ 11
D~ 11
and then N N = I . Furthermore, by Theorem 16.2, we have that K stabilizes G and
that kF` (G; K )k1 < 1 if and only if F`(G; K ) internally stabilizes N and
kF`(N; F` (G; K ))k1 < 1:
Note that
F`(N; F` (G; K )) = F`(M; K~ )
with
2
D~ 12 3
C~1 B~1 R1;1=2 B~2 + B~1 R1;1D~ 11
A~ + B~1 R1;1D~ 11
7
6
7
M (s) = 64
R~1;1=2 C~1
0
R~1;1=2 D~ 12
5
D~ 12
C~1 D~ 21 R1;1=2
D~ 21 R1;1 D~ 11
C~2 + D~ 21 R1;1 D~ 11
D~ 11 and R~1 = I ; D~ 11 D~ 11
. In summary, we have the following
where R1 = I ; D~ 11
lemma.
Lemma 17.3 There is a controller K that internally
stabilizes
G and kF` (G; K )k1 < 1
~
~
if and only if there is a K that stabilizes M and
F`(M; K )
< 1.
"
Corollary 17.4 Let G(s) = A B 2 RH1 . Then kG(s)k1 < 1 if and only if
C D
kDk < 1,
#
"
A
+ B (I ; D D);1 D C B (I ; D D);1=2
2 RH1 ;
M (s) =
(I ; DD );1=2 C
0
449
Relaxing A3 and A4
Suppose that
G = 64
0 0 1
0 0 1
1 1 0
3
7
5
which violates both A3 and A4 and corresponds to the robust stabilization of an integrator. If the controller u = ;x where > 0 is used, then
Relaxing A1
If assumption A1 is violated, then it is obvious that no admissible controllers exist.
Suppose A1 is relaxed to allow unstabilizable and/or undetectable modes on the j!
axis and internal stability is also relaxed to also allow closed-loop j! axis poles, but
A2-A4 is still satised. It can be easily shown that under these conditions the closedloop H1 norm cannot be made nite and, in particular, that the unstabilizable and/or
undetectable modes on the j! axis must show up as poles in the closed-loop system,
see Lemma 16.1.
450
Sensible control problems can be posed which violate A1 and either or both of A3 and
A4. For example, cases when A has modes at s = 0 which are unstabilizable through B2
and/or undetectable through C2 arise when an integrator is included in a weight on a
disturbance input or an error term. In these cases, either A3 or A4 are also violated, or
the closed-loop H1 norm cannot be made nite. In many applications, such problems
can be reformulated so that the integrator occurs inside the loop (essentially using the
internal model principle) and is hence detectable and stabilizable. We will show this
process in the next section.
An alternative approach to such problems which could potentially avoid the problem
reformulation would be to pursue the techniques in the last chapter but to relax internal
stability to the requirement that all closed-loop modes be in the closed left half plane.
Clearly, to have nite H1 norms, these closed-loop modes could not appear as poles
in Tzw . The formulae given in this chapter will often yield controllers compatible with
these assumptions. The user would then have to decide whether closed-loop poles on
the imaginary axis were due to weights and hence acceptable or due to the problem
being poorly posed, as in the above example.
A third alternative is to again introduce perturbations so that A1, A3, and A4 are
satised. Roughly speaking, this would produce sensible answers for sensible problems,
but the behavior as ! 0 could be problematic.
Relaxing A2
In the cases that either D12 is not full column rank or that D21 is not full row rank,
improper controllers can give a bounded H1 -norm for Tzw , although the controllers will
not be admissible by our denition. Such singular ltering and control problems have
been well-studied in H2 theory and many of the same techniques go over to the H1 -case
(e.g. Willems [1981], Willems, Kitapci, and Silverman [1986], and Hautus and Silverman
[1983]). In particular, the structure algorithm of Silverman [1969] could be used to make
the terms D12 and D21 full rank by the introduction of suitable dierentiators in the
controller. A complete solution to the singular problem can be found in [Stroorvogel,
1990].
451
i.e., K (s) has a pole at s = 0. (In general, K is required to have poles on the imaginary
axis.)
6z2
Wu
r
-g ;6
Wd
- g? y- We -z1
We = W~ e (s)M (s)
where M (s) is proper, containing all the imaginary axis poles of We , and M ;1 (s) 2
RH1 , W~ e (s) is stable and minimum phase. Now suppose there exists a controller K (s)
which contains the same imaginary axis poles that achieves the performance. Then
without loss of generality, K can be factorized as
452
6z2
Wu
K^
Wd
- ?g -
y1- ~
We
-z1
z1
W~ e
z2
Wu
Wd
?g P
y1
K^
"
"
"
W~ e = Ae Be ; Wu = Au Bu ; Wd = Ad Bd
Ce De
Cu Du
"
Cd Dd
"
M = Am Bm ; P = Ap Bp :
Cm Dm
Cp Dp
Then a realization for the generalized system is given by
2 "
G(s) = 64
W~ e MWd
0
MWd
"
W~ e MP
Wu
MP
# 3
7
5
453
3
Ae 0 Be Cm Be DmCd Be Dm Cp Be Dm Dd Be Dm Dp
6
6 0 Au
0
Bu 777
0
0
0
6
6 0
0 Am
Bm Cd
Bm Cp
Bm Dd
Bm Dp 77
6
7
6
7
6 0
Bd
0
0
0
Ad
0
7:
= 66
0
0
0
Ap
0
Bp 77
6 0
7
6
6 Ce
0 De Cm De Dm Cd De DmCp De DmDd De Dm Dp 77
6
6
0
0
0
0
Du 75
4 0 Cu
0 0 Cm
Dm Cd
Dm Cp
Dm Dd
Dm Dp
We shall illustrate the above design through a simple numerical example. Let
2
3
0 1 0
; 2 = 64 3 2 1 75 ; W = 1;
P = (s +s1)(
d
s ; 3)
;2 1 0
"
10 = ;100 ;90 ; W = 1 :
Wu = ss++100
e s
1
1
454
17.5 H1 Filtering
In this section we show how the ltering problem can be solved using the H1 theory
developed early. Suppose a dynamic system is described by the following equations
x_ = Ax + B1 w(t); x(0) = 0
(17.1)
y = C2 x + D21 w(t)
(17.2)
z = C1 x + D11 w(t)
(17.3)
The ltering problem is to nd an estimate z^ of z in some sense using the measurement
of y. The restriction on the ltering problem is that the lter has to be causal so
that it can be realized, i.e., z^ has to be generated by a causal system acting on the
measurements. We will further restrict our lter to be unbiased, i.e., given T > 0 the
estimate z^(t) = 0 8t 2 [0; T ] if y(t) = 0; 8t 2 [0; T ]. Now we can state our H1 ltering
problem.
H1 Filtering: Given a
> 0, nd a causal lter F (s) 2 RH1 if it exists
such that
2
J := sup kz ; z^2k2 <
2
with z^ = F (s)y.
w2L2 [0;1)
kwk2
17.5. H1 Filtering
455
?
j
z
z^
F (s)
6
4
A B1
C1 D11
C2 D21
3
7
5
z
y
w
G(s)
A B1 0
G(s) = 64 C1 D11 ;I
C2 D21 0
z^
- F (s)
2
sup kzk22 <
2 :
w2L2
3
7
5
(17:4)
kwk2
Hence the ltering problem can be regarded as a special H1 problem. However, comparing with control problems there is no internal stability requirement in the ltering
problem. Hence the solution to the above ltering problem can be obtained from the H1
solution in the previous sections by setting B2 = 0 and dropping the internal stability
requirement.
Theorem 17.5 Suppose (C2 ; A) is detectable and
"
A ; j!I B1
C2
D21
has full row rank for all !. Let D21 be normalized and D11 partitioned conformably as
"
"
"
D11
D21
#"
D11
D21
#
"
2
;
I 0
456
J1 :=
"
"
"
A
0
C1
C2
;1 D11 B1 C1 :
~
;
R
;B1 B1 ;A
;B1D11 ;B1D21
D21 B1 C2
Moreover, if the above conditions are satised, then a rational causal lter F (s) satisfying J <
2 is given by
"
+ Y1 C1 B1 D21
+ Y1 C2 R~ ;1 :
L11 L21 := ; B1 D11
"
z^ = A ; Y1 C2 C2 Y1 C2 y
C1
0
A B1 B2
G = 64 C1 D11 D12
C2 D21 D22
3
7
5
with the same assumptions as before. But for convenience, we will allow the assumption
(A2) to be relaxed to the following
457
(A20 ) D12 is full column rank with D12 D? unitary, and D21 is full row rank with
"
#
D21
D~ ?
unitary.
Next, we will outline the steps required to solve the H2 and H1 control problems.
Because of the similarity between the H2 and H1 problems, they are developed in
parallel below.
Parameterization: Recall that all controllers stabilizing the plant G can be expressed
as
K = F` (M2 ; Q); I + D22 Q(1) invertible
with Q 2 RH1 and
2
3
7
5
where
C1 )
F2 = ;(B2 X2 + D12
)
L2 = ;(Y"2 C2 + B1 D21
#
C1
A
;
B
;
B
2 D12
2 B2
X2 = Ric
0
;C1 D? D? C1 ;(A ; B2 D12 C1 )
"
#
C2 )
C2
(A ; B1 D21
;
C
2
Y2 = Ric
0:
;B1 D~ ? D~ ? B1 ;(A ; B1 D21 C2 )
Then the closed-loop transfer matrix from w to z can be written as
F`(G; K ) = To + UQV; I + D22 Q(1) invertible
where
and
AF2
;B2 F2
B1
To = 64 0
AL2 B1L2 75 2 RH1
C1F2 ;D12F2 D11
#
"
#
"
A
A
B
B
L
F
2
1L2
2
2
;V=
U=
C1F2 D12
C2 D21
AF2 := A + B2 F2 ; AL2 := A + L2 C2
C1F2 = C1 + D12 F2 ; B1L2 = B1 + L2 D21 :
It is easy to show that U is an inner, U U = I , and V is a co-inner, V V = I .
458
"
"
+
U? = AF2 ;X2 C1 D? ; V? =
C1F2
D?
"
are
AL2
B1L2
:
; D~ ?B1 Y2+ D~ ?
Since the multiplication of square inner matrices do not change H2 and H1 norms,
we have for = 2 or 1
kF` (P; K )k = kTo + UQV k
=
=
where
h
"
"
"
V
U U? (To + UQV )
V?
#
R11 + Q R12
R21 R22
i
h
R = U To V V?
U?
#
with the obvious partitioning. It can be shown through some long and tedious
algebra that R is antistable and has the following representation:
2
3
;AF2
EB1L2
;ED21
;E D~ ?
6
6
C2
;Y2+ B1 D~ ? 777
0
;AL2
6
R=6
D11 B
~ ? 75
B2
F2 Y2 ; D12
4
1L2 D12 D11 D21 D12 D11 D
D D11 D~
;D? C1 X2+
;D? D11 B1L2
D? D11 D21
?
?
kF`(G; K )k22
"
R11 + Q R12
R21 R22
#
2
459
"
kR11 + Qk22 +
R21
R12
R22
#
2
Furthermore,
kR11 + Qk22 = kR11 ; D12 D11 D21 k22 + kQ + D12 D11 D21 k22
D11 D21
) 2 H2? and (Q + D12
D11 D21
) 2 H1 . In fact,
since (R11 ; D12
(Q + D12 D11 D21 ) has to be in H2 to guarantee that the 2-norm be nite.
Hence the unique optimal solution is given by a projection to H1 :
D11 D :
Qopt = [;R11 ]+ = ;D12
21
The optimal controller is given by
Kopt = F` (M2 ; Qopt):
In particular, if D11 = 0 then Kopt = F` (M2 ; 0).
"
#!
where
0 := max
R11 + Q R12
R21 R22
R21 R22
"
>
0
R12
R22
#
)
R11 + Q R12
R21 R22
460
where
R22 );1=2
W = R12 (
2 I ; R22
);1=2 R21 :
Z = (
2 I ; R22 R22
The key here is to nd the spectral factors (I ; WW )1=2 and (I ; Z Z )1=2
with stable inverses such that if M = (I ; WW )1=2 and N = (I ; Z Z )1=2 ,
then M; M ;1 ; N; N ;1 2 H1 ; MM = I ; WW , and N N = (I ; Z Z ).
Now let
Z )N ;1 ;
Q^ := M ;1QN ;1 ; G := M ;1 (R11 + WR22
then the problem is reduced to nding Q^ 2 H1 such that
G + Q^ 1 (< ):
(17:5)
^ 2 H1 i Q^ 2 H1 .
Note that Q = M QN
The nal step in the H1 problem involves solving (17.5) for Q^ 2 H1 . This is
a standard Nehari problem and is solved in Chapter 8. The optimal control
law will be given by
Kopt = F` (J; Qopt ):
17.7 Connections
This section considers the connections between the Youla parameterization approach
and the current state space approach discussed in the last chapter. The key is Lemma 15.7
and its connection with the formulae given in the last section.
To see how Lemma 15.7 might be used in the last section to prove Theorems 16.4
and 16.5 or Theorem 17.1, we begin with G having state dimension n. For simplicity,
we assume further that D11 = 0. Then from the last section, we have
"
+
0
"
11 +
21
"
+
Q 0
kTzw k1 = R
=
=
1#
Q R12
R
R22
1
#
0
Q
R
N11 N12 := R :
N21 N22
17.7. Connections
Then
461
2
"
N11 N12
N21 N22
AF2
0
B2 ;X2+C1 D?
6
6 ;B
BX
AL2
Y2 F2
0
= 66 1L2 1 2
;C2
0
0
4 D21 B1 X2
+
~
~
0
0
D? B1 X2 D? B1 Y2
3
7
7
7
7
5
2 RH1 ;
(17:6)
"
w
Let w = 1
w2
"
N11 + Q N12
N21
N22
< :
N11 + Q N12
N21
N22
"
sup
w2BL2
#
2
#"
# 2
N11 + Q N12
w1
:=
N21
N22
w2
2
1
"
#"
N + Q N
w1
11
12
=
sup
N21
N22
w2
fw2BL2 g\fw1 2H?2 g
#
2
P+ 0
0 I
#"
N11 N12
N21 N22
#"
#
2
w1
w2
P+ 0
0 I
#"
N11 N12
N21 N22
#"
w1
w2
< :
But this is exactly the same operator as the one in Lemma 15.7. Lemma 15.7 may be
applied to derive the solvability conditions and some additional arguments to construct
a Q 2 RH1 from X and Y such that kTzw (Q)k1 <
. In fact, it turns out that X in
Lemma 15.7 for N is exactly W in the FI proof or in the dierential game problem.
The nal step is to obtain the controller from M2 and Q. Since M2 has state
dimension n and Q has 2n, the apparent state dimension of K is 3n, but some tedious
state space manipulations produce cancelations resulting in the n dimensional controller
462
formulae in Theorems 16.4 and 16.5. This approach is exactly the procedure used in
Doyle [1984] and Francis [1987] with Lemma 15.7 used to solve the general distance
problem. Although this approach is conceptually straightforward and was, in fact, used
to obtain the rst proof of the current state space results in this chapter, it seems
unnecessarily cumbersome and indirect.
x_ = Ax + B1 w + B2 u
z = C1 x + D11 w + D12 u
(17.7)
(17.8)
where z (t) 2 Rp1 , y(t) 2 Rp2 , w(t) 2 Rm1 , u(t) 2 Rm2 , and x(t) 2 Rn . The following
assumptions are made:
(AS1) (A; B2 ) is stabilizable;
h
(AS3)
A ; j!I B2
C1
D12
We are interested in the following two related quadratic min-max problems: given
> 0,
check if
sup
min kz k2 <
and
min
sup
kz k2 < :
The rst problem can be regarded as a full information control problem since the control
signal u can be a function of the disturbance w and the system state x. On the other
hand, the optimal control signal in the second problem cannot depend on the disturbance
w (the worst disturbance w can be a function of u and x). In fact, it will be shown that
the control signal in the latter problem depends only on the system state; hence, it is
equivalent to a state feedback control.
463
2
R := D1 D1 ;
Im1 0 ; where D1 := [D11 D12 ]
"
"
h
A
0
B
;1 D C1 B
H1 :=
;
R
1
;C1 C1 ;A
;C1 D1
h
where B := B1 B2 .
(a)
sup
wworst = F11 x
where
(b)
min
"
Proof. The condition for part (a) can be shown in the same way as in Chapter 15
and is, in fact, the solution to the FI problem for the general case. We now prove the
condition for part (b). It is not hard to see that (D11 ) <
is necessary since control
u cannot feed back w directly, so the D11 term cannot be (partially) eliminated as in
464
sup
sup
w2B L2 [0;1)
kz k2:
It is easy to verify directly that the optimal control and the worst disturbance can be
chosen in the given form.
2
Dene
D? D D11 =
2
R0 = I ; D11
?
D11 (
2 I ; D D11 );1 D D12
R~0 = I + D12
11
11
2
^
R0 = I ; D11 D11 =
:
; 2 R01=2 (w ; F11 x)
if conditions in (b) are satised. Integrating both equations from t = 0 to 1 with
x(0) = x(1) = 0 gives
and that an optimal control law and a worst disturbance for problem (b) are
u = F21 x w = wsfworst :
465
Moreover, if problem (b) has a solution for a given
, then the corresponding dierential
game problem
2 ;
2 kwk2
min
sup
k
z
k
2
2
u2L [0;1)
2
w2L2 [0;1)
sup
w2L2 [0;1)
kz k22 ; 2 kwk22
since the solvability of problem (b) implies the solvability of problem (a). However,
the converse may not be true. In fact, it is easy to construct an example so that the
problem (a) has a solution for a given
and problem (b) does not. On the other hand,
the problems (a) and (b) are equivalent if D11 = 0.
A B1 B2
G = 64 C1 0 D12 75 :
I 0 0
Note that the state feedback H1 problem is not a special case of the output feedback
problem since D21 = 0. Hence the parameterization cannot be obtained from the output
feedback.
Theorem 17.7 Suppose that the assumptions (AS 1) ; (AS 3) are satised and that B1
has the following SVD:
"
#
0
B1 = U
V ; UU = In ; V V = Im1 ; 0 < 2 Rrr :
0 0
There exists an admissible controller K (s) for the SF problem such that kTzw k1 <
if
and only if H1 2 dom(Ric) and X1 = Ric(H1) 0. Furthermore, if these conditions
are satised, then all admissible controllers satisfying kTzw k1 <
can be parameterized
as
(
"
#
);1 "
#
;1
0
;1
0
;
1
K = F1 + Im2 + Q
U B2 Q
U ;1 (sI ; A^)
0 Im1 ;r
0 Im1 ;r
h
i
C1 +B2 X1 ), A^ = A+
;2B1 B1 X1 +B2 F1 , and Q = Q1 Q2 2
where F1 = ;(D12
RH2 with kQ1 k1 <
. The dimensions of Q1 and Q2 are m2 r and m2 (m1 ; r),
respectively.
466
Proof. The conditions for the existence of the state feedback control law have been
shown in the last section and in Chapter 15. We only need to show that the above
parameterization indeed satises the H1 norm condition and gives all possible state
feedback H1 control laws. As in the proof of the FI problem in Chapter 16, we make
the same change of variables to get G^ SF instead of G^ FI :
z
r
w
v
XXXXX
X r
G^ SF
u
- K
P
w
Atmp B1 B2
6
^
GSF = 4 ;F1 0 I 75 :
I
0 0
So again from Theorem 16.2 and Lemma 16.8, we conclude that K is an admissible
controller for G and kTzw k1 <
i K is an admissible controller for G^ SF and kTvr k1 <
. Now let L = B2 F1 , and then Atmp + L = Atmp + B2 F1 is stable. All controllers
that stabilize G^ SF can be parameterized as K = F`(Mtmp ; ); 2 RH1 where
2
Atmp + B2 F1 + L ;L B2
Mtmp = 64
F1
0 I 75 :
;I
I 0
Then Tvr = F` (G^ SF ; F` (Mtmp ; )) =: F` (Ntmp ; ). It is easy to show that
3
Atmp + B2 F1 B1 0
6
Ntmp = 4
;F1
0 I 75 :
0 0
I
Now let = F1 + ^ , and we have
"
F`(Ntmp ; ) = ^ Atmp + B2 F1 B1 :
I
0
^ = Q1 Q2
"
467
;1 0 ;1
U (sI ; A^):
0 I
i
A B1 B2
6
Gg (s) = 4 C1 D11 D12 75 :
I 0 0
Let
N=
"
;D11
)1=2
(I ; D11 D11
D11 )1=2
(I ; D11
D11
and then N N = I . Furthermore, by Theorem 16.2, we have that K stabilizes Gg and
kF`(Gg ; K )k1 < 1 if and only if F`(Gg ; K ) internally stabilizes N and
Note that
with
R;1 C1 B1 (I ; D11
D11 );1=2 B2 + B1 D11
R;1D12
A + B1 D11
1
1
6
;
1
=
2
;
1
=
2
6
M (s) = 4
R1 C1
0
R1 D12
I
0
0
3
7
7
5
468
The detailed derivation of the H1 solution for the general case is treated in Glover
and Doyle [1988, 1989]. The loop-shifting procedures are given in Safonov, Limebeer,
and Chiang [1989]. The idea is also used in Zhou and Khargonekar [1988] for state
feedback problems. A fairly complete solution to the singular H1 problem is obtained
in Stoorvogel [1992]. The H1 ltering and smoothing problems are considered in detail
in Nagpal and Khargonekar [1991]. The Youla parameterization approach is treated
very extensively in Doyle [1984] and Francis [1987] and in the references therein. The
presentation of the state feedback H1 control in this chapter is based on Zhou [1992].
18
H1 Loop Shaping
This chapter introduces a design technique which incorporates loop shaping methods
to obtain performance/robust stability tradeos, and a particular H1 optimization
problem to guarantee closed-loop stability and a level of robust stability at all frequencies. The proposed technique uses only the basic concept of loop shaping methods and
then a robust stabilization controller for the normalized coprime factor perturbed system is used to construct the nal controller. This chapter is arranged as follows: The
H1 theory is applied to solve the stabilization problem of a normalized coprime factor
perturbed system in Section 18.1. The loop shaping design procedure is described in
Section 18.2. The theoretical justication for the loop shaping design procedure is given
in Section 18.3.
~ N;
~ ~ M ; ~ N 2 RH1 and
~ N ~ M
< . The transfer matrices (M;
~ N~ )
with M;
1
are assumed to be a left coprime factorization of P (i.e., P = M~ ;1N~ ), and K internally
stabilizes the nominal system.
469
H1 LOOP SHAPING
470
z1
r
-f ;6
- ~ N
- f;
- ?f - M~ ;1
N~
z2
~ M
K (I + PK );1 M~ ;1
1=:
I
1
Finding a controller such that the above norm condition holds is an H1 norm minimization problem which can be solved using H1 theory developed in the previous
chapters.
Suppose P has a stabilizable and detectable state space realization given by
"
P= A B
C D
and let L be a matrix such that A + LC is stable then a left coprime factorization of
P = M~ ;1N~ is given by
#
"
h
i
B
+
LD
L
A
+
LC
:
N~ M~ =
C
D
I
Denote
K^ = ;K
then the system diagram can be put in an LFT form as in Figure 18.2 with the generalized plant
2 "
G(s) =
6
4
0
~
M ;1
M~ ;1
"
I
P
P
# 3
7
5
6 "
6
6
6
4
2
=:
6
4
A
0
C
C
A B1
C1 D11
C2 D21
;L #
"
0
"
I
I
3
B2
D12 75 :
D22
B
I
D
D
# 7
7
7
7
5
471
M~ ;1
?i
N~
K^
; 12 K~
K^# = (I "+ D D
)
#
z1 = U z^1 :
z2
z^2
Then kTzw k1 = kU Tzw k1 = kTz^w k1 and the problem becomes of nding a controller
K~ so that kTz^w k1 <
with the following generalized plant
"
"
0
G^ = U 0 G I
0 I
0 (I + D D); 21
2
6 "
6
6
6
4
A
;(I + DD ); 12 C
(I + D D); 12 D C
C
"
;L
;(I + DD ); 12
B#
"
0
3
7
7
7
7
5
:
I
I
D(I + D D); 12
Now the formulae in Chapter 17 can be applied to1 G^ to obtain a controller K~ and then
the K can be obtained from K = ;(I + D D); 2 K~ . We shall leave the detail to the
=
(I + D D); 12 D
H1 LOOP SHAPING
472
reader. In the sequel, we shall consider the case D = 0. In this case, we have > 1 and
LC ) X ; X (BB ; LL )X +
2 C C = 0
)
+
(
A
;
X1 (A ;
2LC
2
;1
;1 1 1
2 ; 1 1
2 ; 1
Y1 (A + LC ) + (A + LC )Y1 ; Y1 C CY1 = 0:
2
Z1 = I; D^ 11 = 0; D^ 12 = I; D^ 21 =
; 1 I:
Theorem 18.1 Let D = 0 and let L be such that A + LC is stable then there exists a
controller K such that
"
K (I + PK );1 M~ ;1
<
I
1
LC ) X ; X (BB ; LL )X +
2 C C = 0:
)
+
(
A
;
X1 (A ;
2LC
;1
2 ; 1 1 1
2 ; 1 1
2 ; 1
Moreover, if the above conditions hold a central controller is given by
"
A ; BB X1 + LC L
K=
:
; B X1
0
It is clear that the existence of a robust stabilizing controller depends upon the choice
of the stabilizing matrix L, i.e., the choice of the coprime factorization. Now let Y 0
be the stabilizing solution to
AY + Y A ; Y C CY + BB = 0
~ N~ ) given by
and let L = ;Y C . Then the left coprime factorization (M;
h
"
A ; Y C C B ;Y C
N~ M~ =
C
0
I
i
473
"
inf
stabilizing
K (I + PK );1 M~ ;1
=
I
1
p
1 ; max (Y Q)
h
1 ; N~ M~
i
2 ;1=2
=: min
is given by
where
K (I + PK );1 M~ ;1
<
I
1
#
"
A ; BB X1 ; Y C C ;Y C
K (s) =
; B X1
0
X1 = 2 ; 1 Q I ; 2 ; 1 Y Q
;1
where
I ;
2
;1 Y
H1 =
2
0
2
;1 I
"
"
#;1
I ;
2
;1 Y
Hq
2
0
2
;1 I
#
0
Hq = A ; YC C
:
;C C ;(A ; Y C C )
It is clear that the stable invariant subspace of Hq is given by
"
#
I
X; (Hq ) = Im
Q
H1 LOOP SHAPING
474
and the stable invariant subspace of H1 is given by
"
X; (H1 ) = I ;
22;1 Y
0
2 ;1 I
"
X; (Hq ) = Im I ;
22;1 Y Q :
2 ;1 Q
Hence there is a nonnegative denite stabilizing solution to the algebraic Riccati equation of X1 if and only if
2
I ;
2
; 1 Y Q > 0
or
1
1 ; max (Y Q)
and the solution, if it exists, is given by
>p
;1
2
2
X1 =
2 ; 1 Q I ;
2 ; 1 Y Q :
Note
that iY and Q are the controllability Gramian and the observability Gramian
ofi
h
h
~
~
~
~
respectively.
Therefore,
we
also
have
that
the
Hankel
norm
of
N M
N M
p
is max (Y Q).
2
~ N ~ M
1 < :
Then there is a robustly stabilizing controller for P if and only if
r
1 ; max (Y Q) = 1 ; N~ M~
i
2
The solutions to the normalized left coprime factorization stabilization problem are
also solutions to a related H1 problem which is shown in the following lemma.
"
K (I + PK );1M~ ;1
=
K (I + PK );1 h
I P
I
I
1
475
and
M~ N~
M~ N~
"
=
=
This implies
"
"
"
"
"
"
K
I
K
I
K
I
K
I
K
I
#
#
#
#
#
ih
i
M~ N~
=
M~ N~
1
=I
i
K (I + PK );1M~ ;1
I
1
h
(I + PK );1 M~ ;1 M~ N~
h
(I + PK );1 M~ ;1 M~ N~
h
(I + PK );1 I P
ih
(I + PK );1 M
1
i
M~ N~
1
M N~
(I + PK );1 M
i
M~ N~
1
h
~ ;1
~
1
~ ;1
= 1:
"
K (I + PK );1M~ ;1
=
K (I + PK );1 h
I P
I
I
1
Corollary 18.5 A controller solves the normalized left coprime factor robust stabilization problem if and only if it solves the following H1 control problem
"
and
K
"
inf
stabilizing
I (I + PK );1 h
I P
K
I (I + PK );1 h
I P
K
=
=
p
<
1
1 ; max (Y Q)
1;
N~ M~
i
2 ;1=2
H1 LOOP SHAPING
476
The solution Q can also be obtained in other ways. Let X 0 be the stabilizing
solution to
XA + A X ; XBB X + C C = 0
then it is easy to verify that
Q = (I + XY );1 X:
Hence
min = p
1 ; max (Y Q)
= 1;
N~ M~
i
2 ;1=2
= 1 + max (XY ):
Similar results can be obtained if one starts with a normalized right coprime factorization. In fact, a rather strong relation between the normalized left and right coprime
factorization problems can be established using the following matrix fact.
Lemma 18.6 Let M and N be any compatibly dimensioned complex matrices such
that MM = M , NN = N , and M + N = I . Then i (M ) = i (N ) for all i such that
0 < i (M ) =
6 1.
Proof. We rst show that the eigenvalues of M and N are either 0 or 1 and M and N
are diagonalizable. In fact, assume that is an eigenvalue of M and x is a corresponding
eigenvector, then x = Mx = MMx = M (Mx) = Mx = 2 x, i.e., (1 ; )x = 0.
This implies that either = 0 or = 1. To show that M is diagonalizable, assume
M = TJT ;1 where J is a Jordan canonical form, it follows immediately that J must
be diagonal by the condition M = MM . The proof for N is similar.
Next, assume that M is diagonalized by a nonsingular matrix T such that
"
M = T I 0 T ;1:
0 0
Then
"
N = I ; M = T 0 0 T ;1:
0 I
Dene
"
A B := T T
B D
M ; I ) = 0
det(M
"
#
"
#
I
0 I 0
det(
T T
; T T ) = 0
0 0
0 0
(1 ; )A ;B
477
det
det(;D ; 1 ; B A;1 B ) = 0
det( 1 ; D + B A;1 B ) = 0
,
,
,
;B
"
;D
=0
;A ;B = 0
det
;B (1 ; )D
det(N N ; I ) = 0:
This implies that all nonzero eigenvalues of M M and N N that are not equal to 1 are
equal, i.e., i (M ) = i (N ) for all i such that 0 < i (M ) 6= 1.
2
Using this matrix fact, we have the following corollary.
Corollary 18.7 Let K and P be any compatibly dimensioned complex matrices. Then
"
I (I + PK );1 h
I P
K
Proof. Dene
"
"
I (I + KP );1 h
I K
P
"
h
i
h
i
M = I (I + PK );1 I P ; N = ;P (I + KP );1 ;K I :
K
I
"
"
i
I (I + KP );1 h
0 I
N 0 ;I :
I K =
P
;I 0
I 0
Corollary 18.8 Let P = M~ ;1N~ = NM ;1 be respectively the normalized left and right
coprime factorizations. Then
"
K (I + PK );1M~ ;1
=
M ;1 (I + KP );1 h
I K
I
1
H1 LOOP SHAPING
478
Proof. This follows from Corollary 18.7 and the fact that
M ;1 (I + KP );1 I K
"
I (I + KP );1 h
I K
P
This corollary says that any H1 controller for the normalized left coprime factorization is also an H1 controller for the normalized right coprime factorization. Hence
one can work with either factorization.
This section considers the H1 loop shaping design. The objective of this approach is to
incorporate the simple performance/robustness tradeo obtained in the loop shaping,
with the guaranteed stability properties of H1 design methods. Recall from Section 5.5
of Chapter 5 that good performance controller design requires that
;
(PK ) 1; (KP ) 1; (K ) 1
in some low frequency range and
(PK ) 1; (KP ) 1; (K ) M
in some high frequency range where M is not too large.
The design procedure is stated below.
;6
di
- ?e
479
- ?e y?en
max =
=
"
inf
stabilizing
h
!;1
I (I + G K );1 M~ ;1
s
s
K
1
1 ; N~s M~ s
i
2
<1
and M~ s ; N~s dene the normalized coprime factors of Gs such that Gs = M~ s;1 N~s
and
M~ s M~ s + N~s N~s = I:
If max 1 return to (1) and adjust W1 and W2 .
b) Select max, then synthesize a stabilizing controller K1 , which satises
"
I (I + G K );1 M~ ;1
;1 :
s 1
s
K1
1
(3) The nal feedback controller K is then constructed by combining the H1 controller K1 with the shaping functions W1 and W2 such that
K = W1 K1 W2 :
A typical design works as follows: the designer inspects the open-loop singular values
of the nominal plant, and shapes these by pre- and/or postcompensation until nominal
performance (and possibly robust stability) specications are met. (Recall that the
open-loop shape is related to closed-loop objectives.) A feedback controller K1 with
associated stability margin (for the shaped plant) max, is then synthesized. If max
is small, then the specied loop shape is incompatible with robust stability requirements,
and should be adjusted accordingly, then K1 is reevaluated.
In the above design procedure we have specied the desired loop shape by W2 GW1 .
But, after Stage (2) of the design procedure, the actual loop shape achieved is in fact
H1 LOOP SHAPING
480
;6
- K1
- W1 -
- W2 -
- W1 -
- W2
Gs
;6
- W2 - K1 - W1
K
Figure 18.4: The Loop Shaping Design Procedure
given by W1 K1W2 G at plant input and GW1 K1 W2 at plant output. It is therefore
possible that the inclusion of K1 in the open-loop transfer function will cause deterioration in the open-loop shape specied by Gs . In the next section, we will show
that the degradation in the loop shape caused by the H1 controller K1 is limited at
frequencies where the specied loop shape is suciently large or suciently small. In
particular, we show in the next section that can be interpreted as an indicator of the
success of the loop shaping in addition to providing a robust stability guarantee for the
closed-loop systems. A small value of max (max 1) in Stage (2) always indicates
incompatibility between the specied loop shape, the nominal plant phase, and robust
closed-loop stability.
Remark 18.1 Note that, in contrast to the classical loop shaping approach, the loop
shaping here is done without explicit regard for the nominal plant phase information.
That is, closed-loop stability requirements are disregarded at this stage. Also, in contrast with conventional H1 design, the robust stabilization is done without frequency
weighting. The design procedure described here is both simple and systematic, and only
assumes knowledge of elementary loop shaping principles on the part of the designer.
Remark 18.2 The above robust stabilization objective can also be interpreted as the
more standard H1 problem formulation of minimizing the H1 norm of the frequency
weighted gain from disturbances on the plant input and output to the controller input
"
#
I (I + Gs K1);1M~ s;1
K1
1
=
=
=
=
481
"
#
I (I + GsK1);1 h I Gs i
K1
1
"
#
W2 (I + GK );1 h W2;1 GW1 i
W1;1K
1
" #
I (I + K1Gs );1 h I K1 i
Gs
1
" ;1 #
W1 (I + KG);1 h W1 GW2;1 i
W2 G
1
This shows how all the closed-loop objectives in (18.1) and (18.2) are incorporated. ~
The objective of this section is to provide justication for the use of parameter as a
design indicator. We will show that is a measure of both closed-loop robust stability
and the success of the design in meeting the loop shaping specications.
We rst examine the possibility of loop shape deterioration at frequencies of high
loop gain (typically low frequency). At low frequency (in particular, ! 2 (0; !l )), the deterioration in loop shape at plant output can be obtained by comparing (GW1 K1 W2 )
to (Gs ) = (W2 GW1 ). Note that
(GK ) = (GW1 K1 W2 ) = (W2;1 W2 GW1 K1 W2 ) (W2 GW1 ) (K1 )=(W2 )
(18:3)
where () denotes condition number. Similarly, for loop shape deterioration at plant
input, we have
(KG) = (W1 K1W2 G) = (W1 K1 W2 GW1 W1;1 ) (W2 GW1 ) (K1 )=(W1 ):
(18:4)
In each case, (K1 ) is required to obtain a bound on the deterioration in the loop shape
at low frequency. Note that the condition numbers (W1 ) and (W2 ) are selected by
the designer.
Next, recalling that Gs denotes the shaped plant, and that K1 robustly stabilizes
the normalized coprime factorization of Gs with stability margin , then we have
"
#
I (I + Gs K1);1M~ s;1
;1 :=
K1
1
(18:5)
H1 LOOP SHAPING
482
is explicitly bounded by functions of and (Gs ), the minimum singular value of the
shaped plant, and hence by (18.3) and (18.4) K1 will only have a limited eect on the
specied loop shape at low frequency.
Theorem 18.9 Any controller K1 satisfying (18.5), where Gs is assumed square, also
satises
(Gs (j!)) ;
2 ; 1
(K1 (j!)) p
2 ; 1 (Gs (j!)) + 1
Now
can be rewritten as
"
#
I (I + Gs K1);1 M~ s;1
K1
1
(18:6)
We will next show that K1 is invertible. Suppose that there exists an x such that
K1 x = 0, then x (18:6) x gives
;2 x x x M~ s M~ s x
which implies that x = 0 since M~ sM~ s <
;2I , and hence K1 is invertible. Equation (18.6) can now be written as
; K ;1 + I )
2 (K ; ; G )M~ M~ s (K ;1 ; Gs ):
(K1
(18:7)
1
1
s s
1
Dene W such that
483
;1 yields
and completing the square in (18.7) with respect to K1
; + N )(WW );1 (K ;1 + N ) (
2 ; 1)RR
(K1
1
where
p
W ;1 (K1;1 + N )R;1
2 ; 1:
;
;1 + N )R;1 (W ;1 )(K1
;1 + N )(R;1 ) to get
Use W ;1 (K1
o;1
(18:8)
4
s )
(N N ) = (1 ;
2+(Gs(G
G G ))2
s s
(R R) = 1 ;1
+2 +(G(sGGsG) )
s s
therefore
1 + 2 (Gs ) 1=2
p
(Gs Gs ) 1=2 =
(W ) = max (WW ) = 1 ;1
+2 +min
min (Gs Gs )
1 ;
2 + 2 (Gs )
(N ) =
(R) =
max (N N ) =
max (R R) =
1 + 2(G ) 1=2
1 + min (Gs Gs ) 1=2 =
s
:
1 ;
2 + min (Gs Gs )
1 ;
2 + 2 (Gs )
H1 LOOP SHAPING
484
The main implication of Theorem 18.9 is that the bound on (K1 ) depends only
on the selected loop shape, and the stability margin of the shaped plant. The value of
(= ;1) directly determines the frequency range over which this result is valid{a small
(large ) is desirable, as we would expect. Further, Gs has a suciently large loop
gain, then so also will Gs K1 provided
(= ;1) is suciently small.
In an analogous manner, we now examine the possibility of deterioration in the loop
shape at high frequency due to the inclusion of K1 . Note that at high frequency (in
particular, ! 2 (!h ; 1)) the deterioration in plant output loop shape can be obtained
by comparing (GW1 K1W2 ) to (Gs ) = (W2 GW1 ). Note that, analogously to (18.3)
and (18.4) we have
loop shape at high frequency. In an identical manner to Theorem 18.9, we now show
that (K1 ) is explicitly bounded by functions of
, and (Gs ), the maximum singular
value of the shaped plant.
Proof. The proof of Theorem 18.10 is similar to that of Theorem 18.9, and is only
sketched here: As in the proof of Theorem 18.9, we have M~ s M~ s = (I + Gs Gs );1 and
K1 )
2(I ; K G )(M~ M~ s )(I ; Gs K1 ):
(I + K1
(18:9)
1 s s
485
p
V (K1 + M )Y ;1 2 ; 1
2
(K1 ) (V )
(;Y ;11 ) + (M ):
(18:10)
The results in Theorem 18.9 and 18.10 conrm that
(alternatively ) indicates the
compatibility between the specied loop shape and closed-loop stability requirements.
Theorem 18.11 Let G be the nominal plant and let K = W1K1W2 be the associated
controller obtained from loop shaping design procedure in the last section. Then if
"
#
I (I + Gs K1);1 M~ s;1
K1
1
H1 LOOP SHAPING
486
we have
K (I + GK );1 G
where
(18.13)
(18.14)
(18.15)
(18.16)
(N~s ) = (Ns ) =
(18:17)
1 + 2 (W2 GW1 )
1=2
1
(18:18)
(M~ s ) = (Ms ) = 1 + 2 (W
2 GW1 )
and (N~s ; M~ s ), respectively, (Ns ; Ms ), is a normalized left coprime factorization, respectively, right coprime factorization, of Gs = W2 GW1 .
The proof for the normalized right coprime factorization is similar. All other inequalities
follow from noting
"
#
I (I + Gs K1);1 M~ s;1
and
K1
"
#
I (I + Gs K1);1M~ s;1
K1
1
=
=
"
#
W2 (I + GK );1 h W2;1 GW1 i
W1;1K
1
" ;1 #
W1 (I + KG);1 h W1 GW2;1 i
W2 G
1
487
2
This theorem shows that all closed-loop objectives are guaranteed to have bounded
magnitude and the bounds depend only on
, W1 ; W2 , and G.
The H1 loop shaping using normalized coprime factorization was developed by McFarlane and Glover [1990, 1992]. In the same references, some design examples were
also shown. The method has been applied to the design of scheduled controllers for
a VSTOL aircraft in Hyde and Glover [1993]. The robust stabilization of normalized
coprime factors is closely related to the robustness in the gap metric and graph topology, see El-Sakkary [1985], Georgiou and Smith [1990], Glover and McFarlane [1989],
McFarlane, Glover, and Vidyasagar [1990], Qiu and Davison [1992a, 1992b] Vinnicombe
[1993], Vidyasagar [1984, 1985], Zhu [1989], and references therein.
488
H1 LOOP SHAPING
19
490
performance degradation is minimized with the reduced order controllers. The purpose
of this chapter is to introduce several controller reduction methods that can guarantee
the closed-loop stability and possibly the closed-loop performance as well.
# 2 A B1 B2 3
G12 = 6
4 C1 D11 D12 75
G22
"
G = G11
G21
"
A B2
C2 D22
C2 D21 D22
z
y
G(s)
- K (s)
w
491
(19:1)
kWa k1 < 1:
(19:2)
Proof. Since
I ; G22 K^ = I ; G22 K ; G22 = (I ; G22 K )(I ; (I ; G22 K );1 G22 )
by small gain theorem, the system is stable if kWa k1 < 1. On the other hand,
^ 22 = I ; KG22 ; G22 = (I ; (I ; G22 K );1 G22 )(I ; KG22 )
I ; KG
so the system is stable if kWa k1 < 1.
"
K = Ak Bk :
Ck Dk
Then
Wa =
=
3"
75 ; A B2
S 0 Ip ; G22
Iq K
C2 D22
Ck Iq Dk
2
3
;1 C2
~ ;1 Ck
~ ;1
A
+
B
D
R
B
R
B
R
2
k
2
2
66 B R;1C
7
Ak + Bk D22 R~ ;1 Ck Bk D22 R~ ;1 75
k
2
4
"
02 A
k
=SB
@64 0
R;1 C2
0 Bk
0 Ip
R;1 D22 Ck
#1
CA
D22 R~ ;1
492
the same number of unstable poles. Suppose K is unstable, then in order to make sure
that K^ and K have the same number of right half plane poles, K is usually separated
into stable and unstable parts as
K = K+ + K;
where K+ is stable and a reduction is done on K+ to obtain a reduced K^ + , the nal
reduced order controller is given by K^ = K^ + + K; .
We shall illustrate the above procedure through a simple example.
0 0 ;3
20
75 ; B1 = C1 = 64 1
=
D11 = 0; D12 = D21
2 3
75
0
1
7
6
0 5 ; B2 = C2 = 4 1
0 0
1
" #
0
; D22 = 0:
1
148:79(s + 1)(s + 3)
K = ; (s + 31
:74)(s + 3:85)(s ; 9:19)
with kTzw k2 = 55:09. Since K is unstable, we need to separate K into stable part and
antistable part K = K+ + K; with
:15(s + 3:61)
;34:64
K+ = ; (s 114
+ 31:74)(s + 3:85) ; K; = s ; 9:19 :
Wa (K+ ; K^ +)
1 ;
we have
and
:085
K^ + = ; s 117
+ 34:526
:72(s + 0:788)
K^ := K^ + + K; = ; (s 151
+ 34:526)(s ; 9:19) :
The kTzw k2 with the reduced order controller K^ is 61:69. On the other hand, if K+
is reduced directly by balanced truncation without stability weighting Wa , then the
reduced order controller does not stabilize the closed-loop system. The results are
summarized in Table 19.1 where both weighted and unweighted errors are listed.
493
K ; K^
1
Wa (K ; K^ )
1 kTzw k2
0.1622
0.1461
2.5295
0.471
unstable
61.69
Table 19.1: BT: Balance and Truncate, WBT: Weighted Balance and Truncate
It may seem somewhat strange that the unweighted error
K ; K^
resulted from
1
weighted balanced reduction is actually smaller than that from unweighted balanced
reduction. This happens because the balanced model reduction
is not optimal
in L1
^
norm. We should also point out that the stability criterion
Wa (K ; K )
< 1 (or
1
(K ; K^ )Wa
1 < 1) is only sucient. Hence having
Wa (K ; K^ )
1 1 does not
necessarily imply that K^ is not a stabilizing controller.
3
and
~ ; NU
~ );1 N~ M~ = V ;1 (I ; G22 K );1 G22 I
N~n M~ n := (MV
"
# "
"
Nn := N (V~ M ; UN
~ );1 = G22 (I ; KG22 );1 V~ ;1 :
Mn
M
I
Note that M~ n ; N~n; Mn ; Nn do not depend upon the specic coprime factorizations of
G22 .
^ V^ 2 RH1 be the reduced right coprime factors of U and V . Then
Lemma 19.2 Let U;
h
" # " #!
;N~n M~ n i U ; U^
< 1:
V
V^
1
(19:3)
494
h
"
U~ V~ i ; h U^~ V^~ i ;Nn
Mn
#
< 1:
1
(19:4)
Proof. We shall only show the results for the right coprime controller reduction. The
case for the left coprime factorization is analogous. It is well known that K^ := U^ V^ ;1
stabilizes the system if and only if (M~ V^ ; N~ U^ );1 2 RH1 . Since
" h
i " U ; U^ ##
~
^
~
^
~
~
~
~
M V ; N U = (MV ; NU ) I ; ;Nn Mn
V ; V^
the stability of the closed-loop system is guaranteed if
h
"
#
;N~n M~ n i U ; U^
< 1:
V ; V^
1
(19:5)
(19:6)
(19:7)
(19:8)
It is clear from this lemma that the coprime factors of the controller should be reduced so
that the weighted errors in (19.3) and (19.4) are small. Note that there is no restriction
495
on the right half plane poles of the controller. In particular, K and K^ may have dierent
number of right half plane poles. It is also not hard to see that the additive reduction
method in the last subsection may be regarded as a special case of the coprime factor
reduction if the controller K is stable by taking V = I or V~ = I .
Let L; F; Lk and Fk be any matrices such that A + LC2 ; A + B2 F; Ak + Lk Ck and
Ak + Bk Fk are stable. Dene
3
2
i " A + LC2 B2 + LD22 L # " N # 6 A + B2 F B2 7
;
= 4 C2 + D22 F D22 5 ;
N~ M~ =
M
C2
D22
I
F
I
" # 2 Ak + Bk Fk Bk 3 h
i " Ak + Lk Ck Bk + Lk Dk Lk #
U =6
7
:
4 Ck + Dk Fk Dk 5 ; U~ V~ =
V
Ck
Dk
I
F
I
h
"
K = A + B2 F + LC2 + LD22 F ;L ;
F
0
letting Fk = ;(C2 + D22 F ) and Lk = ;(B2 + LD22), we get
"
# 2 A + B2 F ;L 3 h
i " A + LC2 ;L ;(B2 + LD22) #
U =6
7
;
0 5 ; U~ V~ =
4 F
V
F
0
I
C2 + D22 F
~ ; NU
~ = I; V~ M ; UN
~ = I:
MV
496
"
"
i h
i
;Nn = ;N
Therefore, we can chose ;N~n M~ n = ;N~ M~ and
Mn
M
which have the same orders as that of the plant G.
We shall also illustrate the above procedure through a simple example.
Example 19.2 Consider a system with
2 ;1 0 4 3
20 03
213
A = 64 0 ;2 0 75 ; B1 = C1 = 64 1 0 75 ; B2 = C2 = 64 1 75
0 0 ;3
0 0
1
=
D11 = 0; D12 = D21
" #
0
; D22 = 0:
1
3
2
;
1
;
8:198
4
0
66 ;8:198 ;18:396 ;8:198 8:198 77
67:2078(s + 1)(s + 3)
77 = ;
K = 66
(s + 23:969)(s + 3:7685)(s ; 5:3414)
0 5
4 0 ;8:198 ;3
0
;8:198
0
0
with kTzw k2 = 37:02. Since the controller is an observer-based controller, a natural
coprime factorization of K is given by
2 ;1 ;8:198 4 0 3
" # 666 0 ;10:198 0 8:198 777
U =6
66 0 ;8:198 ;3 0 777 :
V
4 0 ;8:198 0 0 5
1
1
1
1
Furthermore, we have
3
2
;
1
0
4
;
1
0
h
i 666 ;8:198 ;10:198 ;8:198 ;1 ;8:198 777
~
~
:
;N M = 6 0
0
;3 ;1 0 75
4
1
1
1
0
0
Applying the frequency weighted balanced model reduction in Chapter 7 to the weighted
coprime factors, we obtain a rst order approximation
" ^ # 2 ;0:1814 1:0202 3
U =6
4 1:2244 0 75
V^
1
6:504
497
h
" # " #!
;N~n M~ n i U ; U^
= 2:0928 > 1;
V
V^
1
K^ = s +1:2491
6:8165 ; kTzw k2 = 52:29
2
3
1
:
0202
;
0
:
1814
14
:
5511
" ^ # 66
7
U = 6 ;0:5288 ;4:1434 ;1:2642 77
64 1:2244 29:5182 0 75
V^
6:504
which gives
;1:3084
h
" # " #!
;N~n M~ n i U ; U^
= 0:024 < 1;
V
V^
1
s + 1:1102) ; kT k = 39:14:
K^ = ; (s +3617:069(
zw 2
:3741)(s ; 4:7601)
Note that the rst order reduced controller does not have the same number of right
half plane poles as that of the full order controller. Moreover, the sucient stability
condition is not satised nevertheless the controller is a stabilizing controller. It is also
interesting to note that the unstable pole of the second order reduced controller is not
at the same location as that of the full order controller.
3
In this section, we consider H1 performance preserving controller order reduction problem. Again we consider the feedback system shown in Figure 19.1 with a generalized
plant realization given by
2 A B B 3
1
2
G(s) = 64 C1 D11 D12 75 :
C2 D21 D22
"
A ; j!I B2
C1
D12
498
(A4)
"
A ; j!I B1
C2
D21
It is shown in Chapter 17 that all stabilizing controllers satisfying kTzw k1 <
can be
parameterized as
K = F`(M1 ; Q); Q 2 RH1 ; kQk1 <
(19:9)
where M1 is of the form
M1 =
"
# 2 A^ B^1 B^2 3
M11 (s) M12 (s) = 6 ^ ^ ^ 7
4 C1 D11 D12 5
M21 (s) M22 (s)
C^ D^ D^
2
21
22
z
y
w
"
499
Proof. We shall assume N21 has full row rank. The case when N12 has full column
rank can be shown in the same way.
To show that kTzw k1 < 1, consider the closed-loop system at any frequency s = j!
with the signals xed as complex constant vectors. Let kQk1 =: < 1 and note that
Twy = N21+ (I ; N22 Q) where N21+ is a right inverse of N21 . Also let := kTwy k1 . Then
kwk2 kyk2, and kN k1 1 implies that kz k22 + kyk22 kwk22 + kuk22. Therefore,
kz k22 kwk22 + (2 ; 1)kyk22 [1 ; (1 ; 2 );2 ]kwk22
which implies kTzw k1 < 1 .
2
Consider the class of (reduced order) controllers that can be represented in the form
K^ = K0 + W2 W1 ;
where K0 may be interpreted as a nominal, higher order controller, is a stable perturbation, with stable, minimum phase, and invertible weighting functions W1 and W2 .
Suppose that kF` (G; K0 )k1 <
. A natural question is whether it is possible to obtain
a reduced order controller K^ in this class such that the H1 performance bound remains
valid when K^ is in place of K0 . Note that this is somewhat a special case of the above
general problem; the specic form of K^ restricts that K^ and K0 must possess the same
right half plane poles, thus to a certain degree limiting the set of attainable reduced
order controllers.
Suppose K^ is a suboptimal H1 controller, i.e., there is a Q 2 RH1 with kQk1 <
such that K^ = F` (M1 ; Q). It follows from simple algebra that
Q = F`(K a;1 ; K^ )
where
"
"
0 I
M1;1 0 I :
a
I 0
I 0
Furthermore, it follows from straightforward manipulations that
kQk1 <
()
F` (K a;1 ; K^ )
1 <
()
F` (K a;1 ; K0 + W2 W1 )
1 <
~ )
< 1
()
F` (R;
where
K ;1 :=
"
;1=2
R~ =
I 0
0
W1
#"
R11 R12
R21 R22
#"
;1=2 I
0
0
W2
500
and R is given by the star product
"
"
It is easy to see that R~12 and R~21 are both minimum phase and invertible, and
hence have full column and full row rank, respectively for all ! 2 R [ 1. Consequently,
by
invoking
Lemma 19.3, we conclude that if R~ is a contraction and kk1 < 1 then
of a Q such that kQk1 <
, or
F`(R;~ )
1 < 1. This guarantees the existence
equivalently, the existence of a K^ such that
F` (G; K^ )
<
. This observation leads
1
to the following theorem.
Theorem 19.4 Suppose W1 and W2 are stable, minimum phase and invertible transfer
matrices such that R~ is a contraction. Let K0 be a stabilizing controller
such
that
^
^
kF`(G; K0 )k1 <
. Then K is also a stabilizing controller such that
F` (G; K )
1 <
if
kk1 =
W2;1 (K^ ; K0 )W1;1
1 < 1:
Since R~ can always be made contractive for suciently small W1 and W2 , there are
innite many
;1W1 and W2 that
satisfy the conditions in the theorem. It is obvious that
;
1
^
to make
W2 (K ; K0 )W1
< 1 for some K^ , one would like to select the \largest"
1
W1 and W2 .
Lemma 19.5 Assume kR22 k1 <
and dene
2
R12
0 ;R11 0
"
#
6
6
R21
0
L = L1 L2 = F`(66 ;R11 0
L2 L3
4 0 R21 0 ;R22
R12
;R22
# "
3
77
77 ;
;1I ):
5
L1 L2 :
L2 L 3
An algorithm that maximizes det(W1 W1 ) det(W2 W2 ) has been developed by Goddard and Glover [1993]. The procedure below, devised directly from the above theorem,
can be used to generate a required
reduced order
controller which will preserve the
closed-loop H1 performance bound
F`(G; K^ )
<
.
1
501
The H1 controller reduction problem can also be considered in the coprime factor
framework. For that purpose, we need the following alternative representation of all
admissible H1 controllers.
Lemma 19.6 The family of all admissible controllers such that kTzw k1 < can also
be written as
"
~ ~
~ = ~ 11 ~ 12
21 22
;1 D^ 11
D^ 12
;1
D^ 12
2 A^ ; B^ D^ ;1C^ B^ D^ ;1 B^ ; B^ D^ ;1D^ 3
2 12
1
2 12 11
2 12 1
75
;1
;1 = 64
;D^ 12;1 C^1
D^ 12
;D^ 12;1 D^ 11
;1 C^1 D^ 22 D^ ;1 D^ 21 ; D^ 22 D^ ;1 D^ 11
C^2 ; D^ 22 D^ 12
12
12
502
"
M11 M12
M21 M22
where M1 =
Lemma 10.2, we have
"
11 12
=
21 22
"
# "
# "
~
~ = ~ 11
21
~ 12
~ 22
Theorem 19.7 Let K0 = 12;221 be a central H1 controller such that kF`(G; K0)k1 <
^ V^ 2 RH1 be such that
and let U;
" ;1 #
" # " #!
I 0 ;1 12 ; U^
< 1=p2:
(19:10)
0 I
V^
22
Proof. Note that by Lemma 19.6, K is an admissible controller such that kTzw k1 <
if and only if there exists a Q 2 RH1 with kQk1 <
such that
"
U
V
# "
11 Q + 12
:=
21 Q + 22
"
Q
=
I
and
(19:11)
K = UV ;1 :
Hence, to show that K^ = U^ V^ ;1 with U^ and V^ satisfying equation (19.10) is also
a stabilizing controller such that kF`(G; K^ )k1 <
, we need to show that there is
another coprime factorization for K^ = UV ;1 and a Q 2 RH1 with kQk1 <
such
that equation (19.11) is satised.
Dene
" ;1 #
I 0 ;1
:=
0 I
"
# "
^
12
; U^
22
V
#!
"
and partition as
Then
and
"
503
U^
V^
# "
"
U
:=
:
V
"
"
12
=
;
I 0 = ;
U
22
0 I
I ; V
"
U^ (I ; V );1 = ;
U (I ; V );1 :
V^ (I ; V );1
I
Dene U := U^ (I ; V );1 , V := V^ (I ; V );1 and Q := ;
U (I ; V );1 . Then UV ;1
is another coprime factorization for K^ . To show that K^ = UV ;
1 = U^ V^ ;1 is a stabilizing
controller such that
kF`(G; K^ )k1
<
, we need to show that
U (I ; V );1
1 <
,
or equivalently
U (I ; V );1
1 < 1. Now
U (I ; V );1 =
i h
i ;1
I 0 I; 0 I
1
02
h
i 3
0
p
I
0
= F` @4 p h
p i 5 ; 2A
I= 2 0 I= 2
h
i 3
2
0
I
0
4 p h
p i5
I= 2 0 I= 2
504
;1
(ii) Find a reduced order controller K^ = U^ V^ ;1 (or V^~ U^~ ) such that the following
frequency weighted H1 error
" ;1 #
" # " #!
I 0 ;1 12 ; U^
< 1=p2
0 I
V^
22
or
(iii) The closed-loop system with the reduced order controller K^ is stable and the
performance is maintained with the reduced order controller, i.e.,
Theorem 19.9 Let W1(s) 2 RH;1 and W2(s) 2 RH;1 with minimal state space realizations
"
"
"
A^ B^
and let G(s) 2 RH1 . Suppose that G^ 1 (s) = ^1 ^ 1 2 RH1 is an r-th order
C1 D1
optimal Hankel norm approximation of [W1 GW2 ]+ , i.e.,
[W GW ] ; Q
G^ 1 = arg deginf
1 2+
H
Qr
and assume
"
# "
A1w ; I B1w ;
C1w
D1w
A2w ; I B2w
C2w
D2w
505
have respectively full row rank and full column rank for all = i (A^1 ); i = 1; : : : ; r.
Then there exist matrices X; Y; Q, and Z such that
A1w X ; X A^1 + B1w Y = 0
(19.12)
^
C1w X + D1w Y = C1
(19.13)
^
QA2w ; A1 Q + ZC2w = 0
(19.14)
QB2w + ZD2w = B^1 :
(19.15)
#
" ^
A1 Z is the frequency weighted optimal Hankel norm approxFurthermore, Gr :=
Y 0
imation, i.e.,
inf^
W1 (G ; G^ )W2
= kW1 (G ; Gr )W2 kH = r+1 ([W1 GW2 ]+ ) :
H
degGr
Proof. We shall assume W2 = I for simplicity. The general case can be proven
similarly. Assume without loss of generality that A^1 has a diagonal form
A^1 = diag[1 ; 2 ; : : : ; r ]:
(The proof can be easily modied if A^1 has a general Jordan canonical form). Partition
X , Y , and C^1 as
X = [X1 ; X2 ; : : : ; Xr ]; Y = [Y1 ; Y2 ; : : : ; Yr ]; C^1 = [C^11 ; C^12 ; : : : ; C^1r ]:
Then the equations (19.12) and (19.13) can be rewritten as
"
A1w ; i I B1w
C1w
D1w
"
#"
# "
Xi = 0 ; i = 1; 2; : : :; r:
Yi
C^1i
A1w ; i I B1w
C1w
D1w
has full row rank for all i and thus the existence of X and Y is guaranteed. Let
"
#
^1
;
X
B
A
1
w
W^ 1 =
2 RH;1 :
C1w ;D^ 1
Then using equations (19.12) and (19.13), we get
2 A B Y 0 3 2 A A X ; X A^ + B Y ;X B^
1w 1w
1
1w
1
1w 1w
W1 Gr = 64 0
A^1 B^1 75 = 64 0
A^1
B^1
C1w D1w Y 0
C1w
C1w X + D1w Y
0
3
75
506
=
2 A 0 ;X B^ 3
64 01w A^1 B^1 1 75 = W^ 1 + G^1:
C1w C^1
kW1 (G ; Gr )kH =
[W1 G]+ + [W1 G]; ; W^ 1 ; G^ 1
H =
[W1 G]+ ; G^ 1
H
= r+1 ([W1 G]+ ) inf^
W1 (G ; G^ )
kW1 (G ; Gr )kH :
H
degGr
Let W1 and W2 be any antistable transfer matrices with all zeros in the right half
plane.
(i) Let
#
" ^
Z
A
1
G^ 1 =
Y 0
be a weighted optimal Hankel norm approximation of G.
(ii) Let the reduced order model G^ be parameterized as
" ^
#
" ^
#
A
B
A
Z
1
1
^
^
G() =
; or G() =
:
Y D
C D
507
Then we have
W1(G ; G^ )W2
1 =
W 1 (G ; G^)W 2
1 :
These factorizations can be easily done using Corollary 13.28 if W1 and W2 are stable and
W1 (1) and W"2 (1) have# respectively full column rank and full row rank. For example,
A B 2 RH with D full row rank and W (j!)W (j!) > 0 for all
1
2
2
C D
"
#
BW
2 RH1 such that M ;1 (s) 2 RH1
!. Then there is a M (s) = A
CW (DD )1=2
assume W2 =
and
where
and
PA + AP + BB = 0
XA + A X + (C ; BW X ) (DD );1 (C ; BW X ) = 0:
Finally, take W 2 (s) = M (s). Then W 2 (s) has all the poles and zeros in the right half
plane and
W1(G ; G^ )W2
1 =
W1 (G ; G^)W 2
1 :
In the case where W1 and W2 are not necessarily stable, the following procedures can
be applied to accomplish this task.
508
(ii) Let W1n = M1;1N1 and W2n = N2M2;1 be respectively the left and right coprime
factorizations such that M1 and M2 are inners. (This step can be done using
Theorem 13.34.)
(iii) Perform the following spectral factorizations
"
"
"
;1 C )
;BR;1B
X = Ric ;(A ; BR ;D
0
;C (I ; DR 1 D )C (A ; BR;1 D C )
W ;1 2 RH;1 and
where W;
W =
"
A + Y CC
B + Y CD :
R;1=2 (D C ; B X )(I + Y X );1
R1=2
!. Let
"
"
X = Ric ;A ;BB
0
19.4. An Example
509
"
~ ;1
;C R~ ;1 C
Y = Ric ;(A ; BD ~R;1 C )
0
;B (I ; D R D)B (A ; BD R~ ;1 C )
W ;1 2 RH;1 and
where W;
"
;1
; Y C )R~;1=2
:
W = A + BB X (I + Y X ) (BD
R~ 1=2
C + DB X
19.4 An Example
We consider a four-disk control system studied by Enns [1984]. We shall set up the
dynamical system in the standard linear fractional transformation form
x_ = Ax
" p+ B1w#+ B2"u #
q1 H x + 0 u
z =
0
I
h
i
y = C2 x + 0 I w
where q1 = 1 10;6; q2 = 1 and
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
3
77
77
77
77
77
77
5
0
0
0
0
0
h
i
h
B1 = pq2 B2 0 ; H = 0 0 0 0 0:55 11 1:32 18
213
66 0 77
66 0 77
66 77
B2 = 66 0 77
66 0 77
66 0 77
405
i
510
UWA
Ko ; K^
1
" # " ^
12 ; U
22
V^
#
1
h
~ 21
SWA
i h
~ 22 ; U^~ V^~
i
1
Wa (Ko ; K^ )
1
19.4. An Example
511
h
" # " #!
;N~n M~ n i 12 ; U^
22
V^
1
SWLCF
PWA
h
~ 12
i h
~ 22 ; U^~ V^~
i " ;Nn #
Mn
1
;1
W2 (Ko ; K^ )W1;1
1
" ;1 #
" # " #!
I 0 ;1 12 ; U^
0 I
22
V^
1
h
~ 12
B
H/O
i h
~ 22 ; U^~ V^~
i ~ ;1 "
;1I
#
I
1
0
Table 19.3 lists the performance weighted right coprime factor reduction errors and
their lower bounds obtained using Corollary 19.10. The in Table 19.3 is dened as
" ;1 #
" # " ^ #!
I
0 ;1
12
:=
; U^
:
0 I
22
V
1
"
U^
By Corollary 19.10, r+1 if the McMillan degree of ^
V
is no greater than r.
Similarly, Table 19.4 lists the stability weighted right coprime factor reduction errors
512
Order of K^
PWA
PWRCF
PWLCF
UWA
UWRCF
UWLCF
SWA
SWRCF
SWLCF
B
H/O
B
H/O
B
H/O
B
H/O
B
H/O
B
H/O
B
H/O
B
H/O
B
H/O
7
1.196
1.196
1.2
1.196
1.197
1.197
U
23.15
1.198
1.197
1.985
5.273
1.327
1.375
1.236
2.401
1.417
1.267
6
1.196
1.197
1.196
1.198
1.196
1.197
1.321
U
1.196
1.197
1.258
U
1.199
2.503
1.197
1.893
1.217
1.485
5
1.199
1.195
1.207
1.196
U
1.198
U
U
1.199
1.282
27.04
U
2.27
2.802
1.251
1.612
48.04
2.259
4
1.197
1.199
1.195
1.199
1.197
1.198
U
U
1.196
1.218
5.059
U
1.47
4.341
1.201
1.388
3.031
1.849
3
2
1
0
U
4.99
U U
2.73 1.94
U U
2.98 1.674
U U
2.036 1.981
U U
U
U
U U
1.586 2.975 3.501 U
U
U
U U
U
U
U U
U
U
U U
U
U
U U
U
U
U U
U
U
U U
23.5
U
U U
1.488 15.12 2.467 U
13.91 1.415
U U
2.622 3.527
U U
U
U
U U
4.184 27.965 3.251 U
Table 19.2: F`(G; K^ ) with reduced order controller: U{closed-loop system is unstable
and their lower bounds obtained using Corollary 19.10. The s and u in Table 19.3 are
dened as
h
i " 12 # " U^ #!
s :=
;N~n M~ n
; ^
22
V
1
" # " ^ #
u :=
12 ; U^
22
V 1
where u is obtained by taking the same U^ and V^ as in s , not from the unweighted
model reduction.
Table 19.2 shows that performance weighted controller reduction methods work very
well. In particular, the PWRCF and PWLCF are easy to use and eective and there
is in general no preference of using either the right coprime method or left coprime
method. Although some unweighted reduction methods and some stability weighted
reduction methods do give reasonable results in some cases, their performances are
very hard to predict. What is the worst is that the small approximation error for the
513
reduction criterion may have little relevance to the closed-loop H1 performance. For
example, Table 19.4 shows that the 7-th order weighted approximation error s using
the Hankel/Optimization method is very small, however, the H1 performance is very
far away from the desired level.
Although the unweighted right coprime factor reduction method gives very good
results for this example, one should not be led to conclude that the unweighted right
coprime factor method will work well in general. If this is true, then one can easily
conclude that the unweighted left coprime factor method will do equally well by considering a dual problem. Of course, Table 19.2 shows that this is not true because the
unweighted left coprime factor method does not give good results. The only reason for
the good performance of the right coprime factor method for this example is the special
data structure of this example, i.e., the relationship between the B1 matrix and B2
matrix:
h
i
B1 = pq2 B2 0 :
The interested reader may want to explore this special structure further.
Order of K^
r
7
6
5
4
3
2
1
0
Lower bounds r+1 0.295 0.303 0.385 0.405 0.635 0.668 0.687 0.702
B 1.009 0.626 4.645 0.750 71.8 6.59 127.2 2.029
H/O 0.295 0.323 0.389 0.658 0.960 1
1
1
Table 19.3: PWRCF: and lower bounds
514
Order of K^
r
7
6
5
4
;
6
;
6
;
6
Lower bounds of s r+1 1:1 10
1:2 10
1:9 10
1:9 10;6
s
B
0.8421
0.5048
2.5439
0.5473
H/O
0.0001
0.2092
0.3182
0.3755
u
B
254.28
9.7018
910.01
21.444
H/O
185.9
30.85
305.3
15.38
Order of K^
r
3
2
1
0
Lower bounds of s r+1 9 10;6 6:23 10;5 1:66 10;4 2:145 10;4
s
B
11.791
1.3164
9.1461
1.5341
H/O
0.5403
0.7642
1
1
u
B
2600.2
365.45
3000.6
383.277
H/O
397.9
288.1
384.3
384.3
Table 19.4: SWRCF: s and the corresponding u
20
516
We will be particularly interested in the case where the set S is described by a set
of functions, hi (x) = 0; i = 1; 2; : : : ; m and m < n or equivalently
2
66
rf (x) := 66
4
@f
@x1
@f
@x2
..
.
@f
@xn
3
77
77
5
2
h
i 666
rH (x) := rh1 (x) rh2 (x) rhm (x) = 6
4
@h1
@x1
@h1
@x2
@h2
@x1
@h2
@x2
@hm
@x1
@hm
@x2
@h1
@xn
@h2
@xn
@hm
@xn
..
.
..
.
..
.
3
77
77 :
5
Theorem 20.1 Suppose that x0 2 Rn is a local minimum of the f (x) subject to the
constraints H (x) = 0 and suppose further that x0 is a regular point of the constraints.
Then there exists a unique multiplier
2 3
66 12 77
= 66 .. 77 2 Rm
4 . 5
m
In the case where the regular point conditions are either not satised or hard to
verify, we have the following alternative.
517
Theorem 20.2 Suppose that x0 2 Rn is a local minimum of f (x) subject to the constraints H (x) = 0. Then there exists
"
0
2 Rm+1
Remark 20.1 Although some second order necessary and sucient conditions for local
minimality can be given, they are usually not very useful in our applications for the
reason that they are very hard to verify except in some very special cases. Furthermore,
even if some sucient conditions can be veried, it is still not clear whether the minima
is global. It is a common practice in many applications that the rst-order necessary
conditions are used to derive some necessary conditions for the existence of a minima.
Then nd a solution (or solutions) from these necessary conditions and check if the
solution(s) satises our objectives regardless of the solution(s) being a global minima
or not.
~
In most of the control applications, constraints are given by a symmetric matrix
function, and in this case, we have the following lemma.
Lemma 20.3 Let T (x) = T (x) 2 Rll be a symmetric matrix function and let x0 2 Rn
be such that T (x0 ) = 0. Then x0 is a regular point of the constraints T (x) = T (x) = 0
if, for P = P 2 Rll , r Trace(T (x0 )P ) = 0 has the unique solution P = 0.
Proof. Since T (x) = [tij (x)] is a symmetric matrix, tij = tji and the eective constraints for T (x) = 0 are given by the l(l + 1)=2 equations, tij (x) = 0 for i = 1; 2; : : : ; l
and i j l. By denition, x0 is a regular point for the eective constraints (tij (x) = 0
for i = 1; 2; : : : ; l and i j l) if the following equation has the unique solution pij = 0
for i = 1; 2; : : : ; l and i j l:
(x0 ) :=
Xl
i=1
rtii pii +
Xl Xl
i=1 j =i+1
2rtij pji = 0:
Now the result follows by dening pij := pji and by noting that (x0 ) can be written
as
0
1
Xl Xl
(x0 ) = r @
with P = P .
i=1 j =1
Corollary 20.4 Suppose that x0 2 Rn is a local minimum of f (x) subject to the constraints T (x) = 0 where T (x) = T (x) 2 Rll and suppose further that x0 is a regular
518
point of the constraints. Then there exists a unique multiplier P = P 2 Rll such that
if we set F (x) = f (x) + Trace(T (x)P ), then rF (x0 ) = 0, i.e.,
rF (x0 ) = rf (x0 ) + r Trace(T (x0 )P ) = 0:
In general, in the case where a local minimal point x0 is not necessarily a regular
point, we have the following corollary.
Corollary 20.5 Suppose that x0 2 Rn is a local minimum of f (x) subject to the constraints T (x) = 0 where T (x) = T (x) 2 Rll . Then there exist 0 6= (0 ; P ) 2 R Rll
with P = P such that
0 rf (x0 ) + r Trace(T (x0 )P ) = 0:
Remark 20.2 We shall also note that the variable x 2 Rn may be more conveniently
given in terms of a matrix X 2 Rkq , i.e., we have
2
66 x11..
66 .
66 xk1
66 x12
x = VecX := 66 ..
66 .
66 xk2
66 ..
4 .
Then
is equivalent to
2
6
rF (x) := 64
2
6
@F (x) := 66
66
@X
4
@F (x)
@x11
@F (x)
@x21
xkq
@F (x)
@x11
..
.
@F (x)
@xkq
@F (x)
@x12
@F (x)
@x22
3
77
77
77
77
77 :
77
77
77
5
3
77 = 0
5
@F (x)
@x1q
@F (x)
@x2q
3
77
77 = 0:
75
..
..
..
.
.
.
@F (x) @F (x) @F (x)
@xk1 @xk2
@xkq
This later expression will be used throughout in the sequel.
~
As an example, let us consider the following H2 norm minimization with constant
state feedback: the dynamic system is given by
x_ = Ax + B1 w + B2 u
z = C1 x + D12 u;
519
520
z
y
w
2 A B B 3
1
2
6
G(s) = 4 C1 0 D12 75 :
C2 D21
i h
C1 D12 = 0 I ;
(iii) D12
(iv)
"
" #
B1 D = 0 .
D21 21
I
"
K = Ac Bc
Cc 0
that internally stabilizes the system G and minimizes the H2 norm of the transfer matrix
Tzw . For technical reasons, we will further assume that the realization of the controller
is minimal, i.e., (Ac ; Bc) is controllable and (Cc ; Ac ) is observable.
Suppose a such controller exists, and then the closed loop transfer matrix Tzw can
be written as
2 A BC B 3 "
#
2 c
1
A~ B~
7
6
Tzw = 4 Bc C2 Ac Bc D21 5 =: ~
C 0
C1
D12 Cc
521
(20:5)
(20:6)
Theorem
20.6
# Suppose (Ac ; Bc; Cc) is a controllable and observable triple and K =
"
Ac Bc
Cc 0
internally stabilizes the system G and minimizes the norm Tzw . Then
there exist n n nonnegative denite matrices X , Y , X^ , and Y^ such that Ac , Bc, and
Cc are given by
(20.7)
(20.8)
(20.9)
Y^ X^ = M ;; ; = Inc
with M positive-semisimple1 and such that with := ; and ? := In ; the following
conditions are satised:
0
0
0
0
=
=
=
=
(20.10)
(20.11)
(20.12)
(20.13)
Proof. The problem can be viewed as a constrained minimization problem with the objective function given by equation (20.5) and with constraints given by equation (20.6).
Let Y~ = Y~ 2 R(n+nc )(n+nc ) and denote
Then
522
and partition X~ and Y~ as
"
"
0
0
0
0
0
0
=
=
=
=
=
=
(20.19)
(20.20)
(20.21)
(20.22)
(20.23)
(20.24)
523
:=
:=
:=
:=
X12 X22;1X12 0
Y12 Y22;1 Y12 0
X11 ; X12 X22;1 X12
Y11 ; Y12 Y22;1 Y12 :
(20.28)
(20.29)
(20.30)
(20.31)
524
Remark 20.3 It is interesting to note that if the full order controller nc = n is considered, then we have = I and ? = 0. In that case, equations (20.10) and (20.11)
become standard H2 Riccati equations:
0 = A X + XA ; XB2 B2 X + C1 C1
0 = AY + Y A ; Y C2 C2 Y + B1 B1
Moreover, there exist unique stabilizing solutions X 0 and Y 0 to these two Riccati
equations such that A ; B2 B2 X and A ; Y C2 C2 are stable. Using these facts, we get
that equations (20.12) and (20.13) have unique solutions:
Z1
X^ =
e(A;Y C2 C2 )t XB2B2 Xe(A;Y C2 C2 )t dt 0
Z01
Y^ =
e(A;B2 B2 X )t Y C2 C2 Y e(A;B2 B2X ) t dt 0:
0
"
Ko := A ; B2 B2 X ; Y C2 C2 Y C2
;B2 X
0
is controllable and observable. (Note that Ko is known to be the optimal H2 controller
from Chapter 14). Furthermore, if X^ and Y^ are nonsingular, we can indeed nd ; and
such that ; = In . In fact, in this case, ; and are both square and ; = ( );1 .
Hence, we have
"
# "
;1
K = Ac Bc = ;(A ; B2 B2 X ; Y;1C2 C2 ); ;Y C2
C 0
;B2 X ;
0
" c
#
A ; B2 B2 X ; Y C2 C2 Y C2 = K
=
o
;B2X
0
i.e., if X^ and Y^ are nonsingular or, equivalently, if optimal controller Ko is controllable
and observable as we assumed, then Theorem 20.6 generates the optimal controller.
However, in general, Ko is not necessarily minimal; hence Theorem 20.6 will not be
applicable.
It is possible to derive some similar results to Theorem 20.6 without assuming the
minimality of the optimal controller by using pseudo-inverse in the derivations, but that,
in general, is much more complicated. An alternative solution to this dilemma would be
simply by direct testing: if a given nc does not generate a controllable and observable
controller, then lower nc and try again.
~
Remark 20.4 We should also note that although we have the necessary conditions for
a reduced order optimal controller, it is generally hard to solve these coupled equations
although some ad hoc homotopy algorithm might be used to nd a local minima. ~
525
Remark 20.5 This method can also be used to derive the H1 results presented in the
previous chapters. The interested reader should consult the references for details. It
should be pointed out that this method suers a severe deciency: global results are
hard to nd. This is due to (a) only rst order necessary conditions can be relatively
easily derived; (b) the controller order must be xed; hence even if a xed-order optimal
controller can be found, it may not be optimal over all stabilizing controllers.
~
526
21
Lemma 21.1 The equation (21.1) has a unique solution if and only if i (A)j (B) 6= 1
for all i; j .
527
528
Remark 21.1 If i (A)j (B) = 1 for some i; j , then the equation (21.1) has either no
solution or more than one solution depending on the specic data given. If B = A and
Q = Q , then the equation is called the discrete Lyapunov equation.
~
The following results are analogous to the corresponding continuous time cases, so they
will be stated without proof.
Lemma 21.2 Let Q be a symmetric matrix and consider the following Lyapunov equation:
AXA ; X + Q = 0
1. Suppose that A is stable, and then the following statements hold:
P Ai Q(A )i and X 0 if Q 0.
(a) X = 1
i=0
(b) if Q 0, then (Q; A) is observable i X > 0.
2. Suppose that X is the solution of the Lyapunov equation; then
(a) ji (A)j 1 if X > 0 and Q 0.
(b) A is stable if X 0, Q 0 and (Q; A) is detectable.
"
"
S := S11 S12 ;
S21 S22
0 ;I
and let J =
2 R2n2n ; then S is called simplectic if J ;1 S J = S ;1 . A
I 0
simplectic matrix has no eigenvalues at the origin, and, furthermore, it is easy to see that
if is an eigenvalue of a simplectic matrix S , then ; 1=, and 1= are also eigenvalues
of S .
Let A, Q, and G be real n n matrices with Q and G symmetric and A nonsingular.
Dene a 2n 2n matrix:
"
#
A
+ G(A );1 Q ;G(A );1
S :=
:
;(A );1 Q
(A );1
529
Then S is a simplectic matrix. Assume that S has no eigenvalues on the unit circle.
Then it must have n eigenvalues in jz j < 1 and n in jz j > 1. Consider the two ndimensional spectral subspaces X; (S ) and X+ (S ): the former is the invariant subspace
corresponding to eigenvalues in jz j < 1, and the latter corresponds to eigenvalues in
jz j > 1. After nding a basis for X; (S ), stacking the basis vectors up to form a matrix,
and partitioning the matrix, we get
"
X; (S ) = Im T1
T2
" #
X; (S ); Im 0
I
(21:2)
Note that the discrete Riccati equation in (21.2) can also be written as
A (I + XG);1 XA ; X + Q = 0:
Remark 21.2 In the case that A is singular, all results presented in this chapter will
still be true if the eigenvalue problem of S is replaced by the following generalized
eigenvalue problem:
"
# "
#
I
G
A
0
;
0 A
;Q I
and X; (S ) is taken to be the subspace spanned by the generalized principal vectors
corresponding to those generalized eigenvalues in jz j < 1. Here the generalized principal
530
"
"
"
A 0 x = I G x
;Q I 1
0 A 1
# "
#!
"
#
A 0 ; I G
I
G
xi =
x ; i = 2; : : : ; k:
;Q I
0 A
0 A i;1
See Dooren [1981] and Arnold and Laub [1984] for details.
"
such that
"
X; (S ) = Im T1
T2
and T1 is invertible. Let X := T2 T1;1, then
"
"
"
X; (S ) = Im T1 = Im I T1 = Im I :
T2
X
X
Obviously, X is unique since
"
"
I = Im I
Im
X1
X2
XT1 = T2;
pre-multiply by T1 to get
(21:3)
S; (T JT )S; ; T JT = 0:
531
T JT = 0
i.e.,
T2T1 = T1T2 :
"
# "
I = I
X
X
Pre-multiply (21.5) by ;X I
T1S; T1;1:
(21:5)
to get
i "I #
;X I S X = 0:
Equivalently, we get
; XA + (I + XG)(A );1 (X ; Q) = 0:
(21:6)
532
"
~
Remark 21.3 Let X+(S ) = Im TT~1 and suppose that T~1 is nonsingular. Then the
2
Riccati equation has an anti-stabilizing solution X~ = T~2 T~1;1 such that (I + GX~ );1 A is
antistable.
"
ST = TS;;
(21:8)
Uk = TS;k ; k = 0; 1; : : ::
Then Uk+1 = SUk with U0 = T . Dening V =
Further dene
"
0 I
, we get T1T2 = U0 V U0 .
0 0
Yk := ;UkV Uk + U0 V U0
= ;
= ;
Now
"
I ;Q
0 I
kX
;1
i=0
kX
;1
i=0
#"
Ui (S V S ; V )Ui :
;Q
#"
0
I 0 0
;
1
;
1
0 ;A G(A )
Q I
since G and Q are assumed to be positive semi-denite. So Yk 0 for all k 0. Note
that Uk ! 0 as k ! 1 since S; has all eigenvalues inside the unit disc. Therefore
T1T2 = limk!1 Yk 0.
2
SV S ; V =
Lemma 21.5 Suppose that G and Q are positive semi-denite. Then S 2 dom(Ric) i
(A; G) is stabilizable and S has no eigenvalues on the unit circle.
533
Proof. The necessary part is obvious. We now show that the stabilizability of (A; G)
and S having no eigenvalues on the unit circle are, in fact, sucient. To show this, we
only need to show that T1 is nonsingular, i.e. Ker T1 = 0. First, it is claimed that
KerT1 is S; -invariant. To prove this, let x 2 Ker T1 . Rewrite (21.8) as
(A + G(A );1 Q)T1 ; G(A );1 T2 = T1 S;
(21.9)
;
1
;
1
;(A ) QT1 + (A ) T2 = T2 S; :
(21.10)
Substitute (21.10) into (21.9) to get
AT1 ; GT2 S; = T1 S; ;
(21:11)
and pre-multiply the above equation by x S; T2 and post-multiply x to get
This in turn implies that T1 S; x = 0 from (21.11). Hence Ker T1 is invariant under S; .
Now to prove that T1 is nonsingular, suppose on the contrary that Ker T1 6= 0. Then
S; jKer T1 has an eigenvalue, , and a corresponding eigenvector, x:
S;x = x
(21:12)
534
Proof. (() Suppose, on the contrary, that S has an eigenvalue ej . Then
" #
for some
" #
" #
S x = ej x
y
y
x 6= 0, i.e.,
y
(A + G(A );1 Q)x ; G(A );1 y = ej x
;(A );1 Qx + (A );1 y = ej y:
Multiplying the second equation by G and adding it to the rst one give
Ax ; ej Gy = ej x
(21.13)
j
;Qx + y = e A y:
(21.14)
Pre-multiplying equation (21.13) by e;j y and equation (21.14) by x yield
e;j y Ax = yGy + y x
;x Qx + x y = ej x A y:
Thus
It follows that
;yGy ; x Qx = 0:
y G = 0
Qx = 0:
Ax = ej x
ej A y = y:
Since x and y cannot be zero simultaneously, ej is either an unobservable mode of
(Q; A) or an uncontrollable mode of (A; G), a contradiction.
()): Suppose that S has no eigenvalue on the unit circle but ej is an unobservable
mode of (Q; A) and x is a corresponding eigenvector. Then it is easy to verify that
" #
" #
S x = ej x ;
so ej is an eigenvalue of S , again a contradiction. The case for (A; G) having uncontrollable mode on the unit circle can be proven similarly.
2
535
Proof. Let Q = C C for some matrix C . The rst half of the theorem follows from
Lemmas 21.5 and 21.6. Now rewrite the discrete Riccati equation as
A (I + XG);1 X (I + GX );1 A ; X + A X (I + GX );2 GXA + C C = 0 (21:15)
and note that by denition (I + GX );1 A is stable and A X (I + GX );2 GXA + C C 0.
Thus X 0 by Lyapunov theorem. To show that the kernel of X has the refereed
property, suppose x 2 KerX , pre-multiply (21.15) by x , and post-multiply by x to get
XAx = 0; Cx = 0:
(21:16)
This implies that KerX is an A-invariant subspace. If KerX 6= 0, then there is an
0 6= x 2 KerX , so Cx = 0, such that Ax = x. But for x 2 KerX , (I + GX );1 Ax =
(I + GX );1 x = x, so jj < 1 since (I + GX );1A is stable. Thus is an unobservable
stable mode of (Q; A).
On the other hand, suppose that jj < 1 is a stable unobservable mode of (Q; A).
Then there exists a x 2 C n such that Ax = x and Cx = 0; do the same pre- and postmultiplications on (21.15) as before to get
jj2 x (XG + I );1 Xx ; x Xx = 0:
This can be rewritten as
x X 1=2 [jj2 (I + X 1=2 GX 1=2 );1 ; I ]X 1=2 x = 0:
Now X 0; G 0, and jj < 1 imply that jj2 (I + X 1=2GX 1=2 );1 ; I < 0. Hence
Xx = 0, i.e., X is singular.
2
Lemma 21.8 Suppose that D has full column rank and let R = DD > 0; then the
following statements are equivalent:
"
536
"
A ; ej I B
C
D
#"
"
#" #
x = 0:
;R;1D C I
A ; ej I B
C
D
(21:17)
"
A ; ej I B
C
D
Now let
" # "
Then
" # "
and
u 6= 0 such that
v
#" #
u = 0:
v
u =
I
0
;
1
v
;R D C I
x =
I
0
;
1
y
R DC I
#" #
x :
y
#" #
u 6= 0
v
(A ; BR;1 D C ; ej I )x + By = 0
(I ; DR;1D )Cx + Dy = 0:
Pre-multiply (21.19) by D to get y = 0. Then we have
(A ; BR;1 D C )x = ej x; (I ; DR;1 D )Cx = 0
;
i.e., ej is an unobservable mode of (I ; DR;1 D )C; A ; BR;1 D C .
(21:18)
(21:19)
Corollary 21.9 Suppose that D has full column rank and denote R = DD > 0. Let
S have the form
"
;1
;1
S = E + G(E;)1 Q ;G(E ;)1
;(E ) Q
(E )
where E = A ; BR;1 D C , G = BR;1 B , Q = C (I ; DR;1D )"C , and E is assumed
#
j I B
A
;
e
to be invertible. Then S 2 dom(Ric) i (A; B ) is stabilizable and
has
C
D
full column rank for all 2 [0; 2]. Furthermore, X = Ric(S ) 0.
537
Note that the Riccati equation corresponding to the simplectic matrix in Corollary 21.9 is
E XE ; X ; E XG(I + XG);1 XE + Q = 0:
This equation can also be written as
A XA ; X ; (B XA + D C ) (D D + B XB );1 (B XA + D C ) + C C = 0:
"
M (z ) = A B
C D
M (z ) =
"
"
#
(A );1
;(A );1 C
:
B (A );1 D ; B (A );1 C
#
Lemma 21.10 Let M (z) = A B 2 RL1 and let S be a simplectic matrix dened
C 0
by
"
;1
;1
S := A ; BB (;A1 ) C C BB (A;1)
:
;(A ) C C
(A )
538
539
"
M1 (z ) = z ; 1= = 1= 2 RL1 :
1 1
z
"
A B
C D
"
540
(e) 9 X > 0 such that
"
A B
C D
# "
(g)
"
"
TAT ;1 TB
CT ;1 D
A B
C D
#!
:=
X 0
0 I
#"
# "
A B ; X 0 < 0;
C D
0 I
#!
"
#"
#"
#;1
T
0
A
B
T
0
< 1;
=
0 I C D 0 I
("
1 In 0
0
2
: 1 2 C ; 2 2 C mp
C (n+m)(n+p)
541
The last equality is obtained from substituting in Riccati equation. Now pre-multiply
the above equation by B (z ;1 I ; A);1 and post-multiply by (zI ; A);1 B to get
I ; M (z ;1 )M (z ) = W (z ;1 )W (z )
where
"
A
B
:
W (z ) =
;
1
=
2
(I ; B XB ) B XA ;(I ; B XB )1=2
"
I ; M (z ;1)M (z ) = W (z ;1 )(I ; B XB )W (z )
"
A
B :
W (z ) =
;
1
(I ; B XB ) B XA ;I
Remark 21.4 As in the continuous time case, the equivalence between (a) and (b) in
Theorem 21.12 can be used to compute the H1 norm of a discrete time transfer matrix.
~
542
"
3
2
A
0
B
64 (A );1 C C
(A );1
(A );1 C D 75
(D ; CA;1 B ) C ;B (A );1 (D ; CA;1 B ) D
2
3
A
0
B
66
XB + (A );1 C D 775 :
0
(A );1
4
DT C + B XA ;B (A );1 (D ; CA;1 B ) D
2
The following corollary is a special case of this lemma which gives the necessary and
sucient conditions of a discrete inner transfer matrix with the state-space representation.
"
A B
C D
2 RH1 is a controllable realization; then N (z ) is inner if and only if there exists a matrix X = X 0 such that
(a) A XA ; X + C C = 0
(b) D C + B XA = 0
(c) (D ; CA;1 B ) D = D D + B XB = I .
543
The following alternative characterization of the inner transfer matrix is often useful
and insightful.
"
A B
C D
"
;1
P = TAT;1 TB
CT
D
and P P = I:
Proof. Rewrite P P = I as
"
A C
B D
#"
T T
#"
I
# "
"
A B = T T
:
C D
I
A XA ; X + C C A XB + C D
= 0:
B XA + D C B XB + D D ; I
This is the desired equation, so N is inner. On the other hand, if the realization is
minimal, then X > 0. This implies that T exists.
In a similar manner, Corollary 21.15 can be used to derive the state-space representation of the complementary inner factor (CIF).
"
N? (z ) = A Y
C Z
where Y and Z satisfy
A XY + C Z = 0
B XY + D Z = 0
Z Z + Y XY = I:
(21:20)
(21:21)
(21:22)
544
i "A B Y #
is inner.
N N? =
C D Z
"Recall#that two transfer matrices M (z); N (z) 2 RH1 are said to be right coprime if
M
N
U (z )N (z ) + V (z )M (z ) = I:
The left coprime is dened analogously. A plant G(z ) 2 RL1 is said to have double
coprime factorization if 9 a right coprime factorization G = NM ;1 , a left coprime
~ V~ 2 RH1 such that
factorization G = M~ ;1 N~ , and U; V; U;
"
#"
#
V U
M ;U~ = I:
(21:23)
;N~ M~
N V~
The state space formulae for discrete time transfer matrix coprime factorization are the
same as for the continuous time. They are given by the following theorem.
"
A B
C D
alization. Choose F and L such that A + BF and A + LC are both stable. Let
~ V~ ; N; M; N~ , and M~ be given as follows
U; V; U;
"
# 2 A + BF
M := 6
4 F
N
C + DF
BZr
Zr
DZr
" ~ # 2 A + BF LZl;1
U := 6
F
0
4
V~
;(C + DF ) Zl;1
" ~# "
3
75
3
75
#
M := A + LC L B + LD
N~
Zl C
Zl Zl D
" # "
#
U := A + LC L ;(B + LD)
V
Zl;1 F
0 Zl;1
where Zr and Zl are any nonsingular matrices. Then G = NM ;1 = M~ ;1N~ are rcf and
lcf, respectively, and (21.23) is satised.
545
Some coprime factorizations are particularly interesting, for example, the coprime
factorization with inner numerator. This factorization in the case of G(z ) 2 RH1
yields an inner-outer factorization.
"
j
Theorem 21.19 Assume that (A; B) is stabilizable, A ;Ce I DB has full column
rank for all 2 [0; 2], and D has full column rank. Then there exists a right coprime
factorization G = NM ;1 such that N is inner. Furthermore, a particular realization is
given by
"
# 2 A + BF
M := 6
4 F
N
C + DF
where
BR;1=2
R;1=2
DR;1=2
3
75
R = D D + B XB
F = ;R;1 (B XA + D C )
Using Lemma 21.17, the complementary inner factor of N in Theorem 21.19 can be
obtained as follows
"
#
A
+ BF Y
N? =
C + DF Z
where Y and Z satisfy
A XY + C Z = 0
B XY + D Z = 0
Z Z + Y XY = I:
Note that Y and Z are only related to F implicitly through X .
Remark 21.5 If G(z) 2 RH1 , then the denominator matrix M in Theorem 21.19 is
an outer. Hence, the factorization G = N (M ;1) is an inner-outer factorization.
~
Suppose that the system G(z ) is not stable; then a coprime factorization with inner
denominator can also be obtained by solving a special Riccati equation.
546
"
Theorem 21.20 Assume that G(z) := A B 2 RL1 and that (A; B) is stabilizC D
able. Then there exists a right coprime factorization G = NM ;1 such that M is inner
if and only if G has no poles on the unit circle. A particular realization is
" # 2 A + BF BR;1=2 3
M := 6
R;1=2 75
4 F
N
C + DF DR;1=2
where
R = I + B XB
F = ;R;1B XA
"
M
i.e., if
N
is an inner.
Similarly,
i an lcf G = M~ ;1N~ is called a normalized left coprime factorization if
~
~
M N is a co-inner. Then the following results follow in the same way as for the
continuous time case.
Theorem 21.21 Let a realization of G be given by
"
G= A B
C D
and dene
547
Z = R + B XB
F = ;Z ;1(B XA + D C )
Z~ = R~ + CY C
L = ;(BD + AY C )Z~;1
"
M
(c) The controllability Gramian P and observability Gramian Q of
N
by
are given
P = (I + Y X );1Y; Q = X
h
while the controllability Gramian P~ and observability Gramian Q~ of M~ N~
are given by
P~ = Y; Q~ = (I + XY );1 X:
The following theorem gives a solution to a special class of spectral factorization problems.
"
Theorem 21.22 Assume G(z) := A B 2 RH1 and
> kG(z)k1. Then, there
C D
exists a transfer matrix M 2 RH1 such that M M =
2I ; G G and M ;1 2 RH1 .
A particular realization of M is
"
#
A
B
M (z ) =
;R1=2 F R1=2
548
where
RD =
2 I ; D D
R = RD ; B XB
F = (RD ; B XB );1 (B XA + D C )
where As := A + BRD;1 D C .
"
Theorem 21.23 Let G(z) := A B 2 RH1 with D full row rank and G(ej )G (ej ) >
C D
0 for all . Then, there exists a transfer matrix M 2 RH1 such that M M = GG . A
particular realization of M is
"
M (z ) = A BW
CW DW
where
BW = APC + BD
DW DW = DD
CW = DW (DD );1 (C ; BW XA)
and
APA ; P + BB = 0
A XA ; X + (C ; BW XA) (DD );1 (C ; BW XA) = 0:
z
y
w
549
2 A B B 3 "
#
1
2
A
B
7
6
:
G(z ) = 4 C1 D11 D12 5 =:
C D
C2 D21
(A2) D12 is full column rank with D12 D? unitary and D21 is full row rank with
"
#
D21 unitary;
D~ ?
(A3)
(A4)
"
"
A ; ej I B2
C1
D12
A ; ej I B1
C2
D21
kTzw k2 .
Denote
C1 ; Ay := A ; B1 D C2 :
Ax := A ; B2 D12
21
Note that the stabilizing solutions exist by the assumptions (A3) and (A4). Note also
that if Ax and Ay are nonsingular, the solutions can be obtained through the following
two simplectic matrices:
"
;1
x );1
H2 := Ax + B2 B2 ;(A1 x ) C1 D ?D? C1 ;B2 B2 (A
;(Ax ) C1 D? D? C1
(Ax );1
"
;1 1 D~ D~ ?B1 ;C2 C2 A;
1
y
?
J2 := Ay + C2;C12 Ay ~ B
:
;Ay B1 D? D~ ? B1
A;y 1
550
Dene
Rb
F2
F0
L2
L0
and
I + B2 X2 B2
;(I + B2 X2 B2 );1 (B2 X2 A + D12 C1 )
;(I + B2 X2 B2 );1 (B2 X2 B1 + D12 D11 )
;(AY2 C2 + B1 D21 )(I + C2 Y2 C2 );1
)(I + C2 Y2 C );1
(F2 Y2 C2 + F0 D21
2
:=
:=
:=
:=
:=
"
B1L2
:
Gf (z ) := 1=2 AL2
1
=
2
Rb (L0 C2 ; F2 ) Rb (L0 D21 ; F0 )
"
Remark 21.6 Note that for a discrete time transfer matrix G(z) = A B 2 RH2,
C D
its H2 norm can be computed as
kG(z )k22 = TracefD D + B LoB g = TracefDD + CLc C g
where Lc and Lo are the controllability and observability Gramians
ALcA ; Lc + BB = 0
A Lo A ; Lo + C C = 0:
Using the above formula, we can compute min kTzw k22 by noting that X2 and Y2 satisfy
the equations
551
For example,
kGc k22 = Trace f(D11 + D12 F0 ) (D11 + D12 F0 ) + (B1 + D2 F0 )X2 (B1 + D2 F0 )g
and
kGf k22 = Trace Rb f(L0 D21 ; F0 )(L0 D21 ; F0 ) + (L0 C2 ; F2 )Y2 (L0 C2 ; F2 ) g :
Proof. Let x denote the states of the system G. Then the system can be written as
x_ = Ax + B1 w + B2 u
z = C1 x + D11 w + D12 u
y = C2 x + D21 w:
(21.24)
(21.25)
(21.26)
"
z = AF2 B1 + B2 F0 B2
C1F2 D11 + D12 F0 D12
#" #
w = G w + UR1=2
c
b
"
where
;1=2
A
B
R
F
2
2
b
U (s) :=
:
C1F2 D12 Rb;1=2
It is easy to shown that U is an inner and that U Gc 2 RH?2 . Now denote the transfer
function from w to by Tw . Then
and
for any given stabilizing controller K . Hence if the states (x) and the disturbance (w)
are both available for feedback (i.e., full information control) and u = F2 x + F0 w, then
Tw = 0 and kTzw k2 = kGc k2 . Therefore, u = F2 x + F0 w is an optimal full information
control law. Note that
= Tw w; Tw = F` (G ; K )
y
G
w
2 A B B 3
1
2
G = 64 ;F2 ;F0 I 75
C2
D21
552
2
66 A + B2F2
K = F` (Mt ; KFC ); Mt = 64 ;F2
C2
0
0
hI ;B2i
h0 Ii
0 0
i3
77
75
h
i3
2
A
B
I
0
1
6
h
i7
G^ = 664 ;F2 ;F0 h 0 I i 775 :
C2
D21
0 0
G~ :=
=
"
2
66
64
#
I 0 0
Rb1=2 0 G^ 6
0 75
4
0 I
0 I
0 0 Rb;1=2
h
i3
A
B1
I
0
h
i7
;Rb1=2 F2 ;Rb1=2F0 h 0 I i 775
D21
C2
0 0
and
K~ FC :=
"
0
KFC :
0 Rb1=2
~ ~
min kRb1=2F` (G^ ; KFC )k2 = min
~ kF`(G ; KFC )k2 :
KFC
KFC
553
"
L2
since the transpose (or
1
Rb =2 L0
F`(G~ ;
and
"
L2 ) = G
f
1
Rb =2 L0
min
kTzw k22 = kGck22 + kGf k22 :
K
"
# "
K = F` (Mt ; L2 ) = A + B2 F2 + L2 C2 ; B2 L0 C2 L2 ; B2 L0 :
L0
;F2 + L0 C2
L0
The proof of uniqueness is similar to the continuous time case, and hence omitted. 2
It should be noted that in contrast with the continuous time the full information
optimal control problem in the discrete time is not a state feedback even when D11 = 0.
The discrete time H1 control problem is much more involved and it is probably
more eective to obtain the discrete solution by using a bilinear transformation.
"
G(s) = A B 2 RH1 :
C D
Let P and Q be two positive semi-denite symmetric matrices such that
APA ; P + BB 0
A QA ; Q + C C 0:
"
P = Q = 1 0
0 2
(21:27)
(21:28)
554
with
2A A B 3
11 12 1
6
G(s) = 4 A21 A22 B2 75 :
C1
C2 D
(21:29)
(21:30)
A21 x = 0;
C1 x = 0:
"
A11 A12
A21 A22
#" #
" #
x = ej x
"
555
(21:31)
(21:32)
A21 = 0;
Hence we have
"
B2 = 0:
A = A11 A12 ;
0 A22
The subsystem with A11 still satises inequalities (21.27) and (21.28) with 1 > 0. This
proves that we can assume without loss of generality that 2 > 0.
2
Remark 21.7 It is important to note that the realization for the truncated subsystem
"
Gr = A11 B1
C1 D
is still balanced in some sense1 since the system parameters satisfy the following equations:
A11 1 A11 ; 1 + A12 2 A12 + B1 B1 0
A11 1 A11 ; 1 + A21 2 A21 + C1 C1 0:
But these equations imply that
hold.
"
n
X
Theorem 21.26 Suppose Gr = A11 B1 . Then kG ; Gr k1 2
i .
C1 D
n
X
i=r+1
i .
i=1
1 Balanced in the sense that the same inequalities as (21.27) and (21.28) are satised.
556
n I = I , we have
kG ; Gr k1 2; r = n ; 1:
Then the theorem follows immediately by scaling and recursively applying this result
since the reduced system Gr is still balanced.
It will be seen that it is more convenient to set = 11=2 . The proof of the theorem
will follow from the following two lemmas and the bounded real lemma which establishes
the relationship between the H1 norm of a transfer matrix and its realizations. (Note
that in the following, a constant matrix X is said to be contractive or a contraction if
kX k 1 and strictly contractive if kX k < 1).
The lemma below shows that for any stable system there is a realization such
" that#
h
i
A
A B is a contraction, and, similarly, there is another realization such that
C
is a contraction.
Lemma 21.27 Suppose that a realization of the transfer matrix G satises P = Q =
diagf2 ; I g; then
" ;1
#
A12 ;1A11 ;1 B1
A22
A21
B2
and
2 A ;1 A 3
64 A2111;1 A2212 75
C1 ;1 C2
are contractive.
"
2
Proof. Since P = 0 I0
A B
i" P
0 I
#"
A P
B
h
Hence
i " ;1A12
"
A22
#2 0 I
64 I 0
0
0 75 :
0 0 I
557
"
12
Lemma 21.28 Suppose that X = XZ11 X
X22
2
6
M = 64
"
Y Z
and Y = 11
Y21 Y22
2 0
66 p1 Y11
Md = 66 2
4 Y21
p12 Y11
p12 X11
p12 Y22
X12
p12 X22
0
; p12 X22
2
are contrac-
3
77
5
3
0 777
:
; p12 Y22 75
p12 X11
0
;Z
Considering X and Y are contractive, we can easily verify that Md Md I , i.e, Md is a
contraction.
2
We can now prove the theorem.
Proof of Theorem 21.26. Note that
"
# 2 A11 0 B1 3
Gr = A11 B1 = 64 0 0 0 75 :
C1 D
C1 0 D
Hence
3
2
p12 B1
A11 0 0
0
7
66 0 0 0
0
0 77
6
1 (G ; G ) = 6 0
0 A11 A12 p12 B1 77 :
6
r
6
2
64 0 0 A21 A22 p12 B2 775
; p12 C1 0 p12 C1 p12 C2 0
Now apply the similarity transformation
2
3
;
0 0
66
77
T = 66 0;1 ;I 0;1 I 77
4 0 05
0 I
0 I
558
"
G= A B
C D
559
and assume that the realization is stabilizable and detectable. Recall from Theorem 21.21
" that
# there exists a normalized right coprime factorization G = NM ;1 such
M is inner.
that
p
Lemma 21.29 Let i = i (Y X ).
given by
"
M
N
are
i = p i 2 < 1:
1 + i
)
i2 = i (PQ) = 1 +i(Y(X
Y X)
i
and i (Y X ) 0.
It is known that there exists a transformation such that X and Y are balanced:
"
#
1 0
X =Y ==
0 2
with 1 = diag[1 Is1 ; : : : ; r Isr ] > 0.
Now partitioning the system G and matrix F accordingly,
2A A B 3
11 12 1
h
i
6
G = 4 A21 A22 B2 75 F = F1 F2 :
C1
C2 D
"
2 A + B F B Z ;1=2 3
#
1
M^ := 6 11 1 1
;1=2 7
F
Z
4
5 2 RH1
1
N^
C1 + DF1
DZ ;1=2
560
Remark 21.8 It should be understood that the reduced model can be obtained by
directly computing X and P and by obtaining a balanced model without solving the
Riccati equation for Y .
~
This reduced coprime factors combined with the robust or H1 controller design
methods can be used to design lower order controllers so that the system is robustly
stable and some specied performance criteria are satised. We will leave the readers
to explore the utility of this model reduction method. However, we would like to point
out that unlike the continuous time case, the reduced coprime factors in discrete time
may not be normalized. In fact, we can prove a more general result.
"
N (z ) = A B
C D
"
A XA ; X + C C A XB + C D
= 0:
B XA + D C B XB + D D ; I
N=
is also balanced with
and
"
# 2 A11 A12 B1 3
A B =6
A21 A22 B2 75
4
C D
C1
C2 D
"
X = = 1 0
0 2
AA ; + BB = 0
"
Nr = A11 B1
C1 D
is stable and contractive, i.e., Nr Nr I .
(21:35)
561
"h
I 0
"
or
"
"
# "
This gives
"
A11 B1
C1 D
# "
1 0
0 I
#"
A11 B1
C1 D
# "
#
; 1 0 0:
0 I
"
#
TA11T ;1 TB1
1
C1 T ;1 D
" #
562
Bibliography
[1] Adamjan, V.M., D.Z. Arov, and M.G. Krein (1971). \Analytic properties of
Schmidt pairs for a Hankel operator and the generalized Schur-Takagi problem,"
Math. USSR Sbornik, Vol. 15, pp. 31-73.
[2] Adamjan, V.M., D.Z. Arov, and M.G. Krein (1978). \Innite block Hankel matrices and related extension problems," AMS Transl., vol. 111, pp. 133-156.
[3] Al-Saggaf, U. M. and G. F. Franklin (1987). \An error bound for a discrete reduced
order model of a linear multivariable system," IEEE Trans. Automat. Contr.,
Vol.AC-32, pp. 815-819.
[4] Al-Saggaf, U. M. and G. F. Franklin (1988). \Model reduction via balanced realizations: an extension and frequency weighting techniques," IEEE Trans. Automat.
Contr., Vol.AC-33, No. 7, pp. 687-692.
[5] Anderson, B. D. O. (1967). \A system theory criterion for positive real matrices,"
SIAM J. Contr. Optimiz., Vol. 6, No. 2, pp. 171-192.
[6] Anderson, B.D.O. (1967). \An algebraic solution to the spectral factorization
problem," IEEE Trans. Auto. Control, Vol. AC-12, pp. 410-414.
[7] Anderson, B. D. O. (1993). \Bode Prize Lecture (Control design: moving from
theory to practice)," IEEE Control Systems, Vol. 13, No.4, pp. 16-25.
[8] Anderson, B.D.O., P.Agathoklis, E.I.Jury and M.Mansour (1986). \Stability
and the Matrix Lyapunov Equation for Discrete 2-Dimensional Systems," IEEE
Trans., Vol.CAS-33, pp.261 ; 266.
[9] Anderson, B. D. O. and Y. Liu (1989). \Controller reduction: concepts and approaches, " IEEE Trans. Automat. Contr. , Vol. AC-34, No. 8, pp. 802-812.
[10] Anderson, B. D. O. and J. B. Moore (1979). Optimal Filtering. Prentice-Hall,
Englewood Clis, New Jersey.
[11] Anderson, B. D. O. and J. B. Moore (1989). Optimal Control: Linear Quadratic
Methods. Prentice-Hall, Englewood Clis, New Jersey.
563
564
BIBLIOGRAPHY
BIBLIOGRAPHY
565
[26] Boyd, S. and C.A. Desoer (1985). \Subharmonic functions and performance
bounds in linear time-invariant feedback systems," IMA J. Math. Contr. and
Info., vol. 2, pp. 153-170.
[27] Boyd, S. and Q. Yang (1989). \Structured and simultaneous Lyapunov functions
for system stability problems," Int. J. Control, vol. 49, No. 6, pp. 2215-2240.
[28] Brasch, F. M. and J. B. Pearson (1970). \Pole placement using dynamical compensators," IEEE Trans. Auto. Contr., Vol. 15, pp. 34-43.
[29] Brogan, W. L. (1991). Modern Control Theory, 3rd Ed., Prentice Hall, Englewood
Clis, New Jersey.
[30] Bruinsma, N. A., and M. Steinbuch (1990). \A fast algorithm to compute the
H1 -norm of a transfer function matrix", Systems and Control letters, Vol. 14, pp.
287-293.
[31] Bryson Jr., A. E. and Y-C. Ho (1975). Applied Optimal Control, Hemisphere
Publishing Corporation.
[32] Chen, C. T. (1984). Linear System Theory and Design. Holt, Rinehart and Winston.
[33] Chen, J. (1992a). \Sensitivity integral relations and design tradeos in linear
multivariable feedback systems," submitted to IEEE Trans. Auto. Contr..
[34] Chen, J. (1992b). \Multivariable gain-phase and sensitivity integral relations and
design tradeos," submitted to IEEE Trans. Auto. Contr..
[35] Chen, J. (1995). Personal communication.
[36] Chen, M. J. and C.A. Desoer (1982). \Necessary and sucient condition for robust
stability of linear distributed feedback systems," Int. J. Control, vol. 35, no. 2,
pp. 255-267.
[37] Chen, T. and B. A. Francis (1992). \ H2 -optimal sampled-data control," IEEE
Trans. Auto. Contr., Vol. 36, No. 4, pp. 387-397.
[38] Chu, C. C. (1985). H1 optimization and robust multivariable control. PhD Thesis,
University of Minnesota.
[39] Chu, C. C. (1988). \On discrete inner-outer and spectral factorization," Proc.
Amer. Contr. Conf..
[40] Clements, D. J. (1993). \Rational spectral factorization using state-space methods," Systems and Control Letters, Vol. 20, pp. 335-343.
[41] Clements, D. J. and K. Glover (1989). \Spectral factorization by Hermitian pencils," Linear Algebra and its Applications, Vol. 122-124, pp. 797-846.
566
BIBLIOGRAPHY
[42] Copple, W. A. (1974). \Matrix quadratic equations," Bull. Austral. Math. Soc.,
Vol. 10, pp. 377 401.
[43] Dahleh, M. A. and J. B. Pearson (1987). \`1 optimal feedback controllers for
MIMO discrete time systems," IEEE Trans. Auto. Contr., Vol. AC-32, No. 4, pp.
314 322.
[44] Daniel, R. W., B. Kouvaritakis, and H. Latchman (1986). \Principal direction
alignment: a geometric framework for the complete solution to the problem,"
IEE Proceedings, vol. 133, Part D, No. 2, pp. 45-56.
[45] Davis, C., W. M. Kahan, and H. F. Weinberger (1982). \Norm-preserving dilations
and their applications to optimal error bounds," SIAM J. Numer. Anal., vol 19,
pp.445 469.
[46] Davison, E. J. (1968). \On pole assignment in multivariable linear systems," IEEE
Trans. Auto. Contr., Vol. 13, No. 6, pp. 747-748.
[47] Delsarte, P., Y. Genin, and Y. Kamp (1979). \The Navanlinna-Pick problem for
matrix valued functions," SIAM J. Applied Math., Vol. 36, pp. 47-61.
[48] Desai, U. B. and D. Pal (1984). \A transformation approach to stochastic model
reduction," IEEE Trans. Automat. Contr., Vol 29, NO. 12, pp. 1097-1100.
[49] Desoer, C. A., R. W. Liu, J. Murray, and R. Saeks (1980). \Feedback system
design: the fractional representation approach to analysis and synthesis," IEEE
Trans. Auto. Contr., Vol. AC-25, No. 6, pp. 399 412.
[50] Desoer, C. A. and M. Vidyasagar (1975). Feedback Systems: Input-Output Properties. Academic Press, New York.
[51] Dooren, P. V. (1981). \A generalized eigenvalue approach for solving Riccati equations," SIAM J. Sci. Stat. Comput., Vol. 2, No. 2, pp. 121-135.
[52] Doyle, J. C. (1978). \Guaranteed Margins for LQG Regulators," IEEE Trans.
Auto. Contr., Vol. AC-23, No. 4, pp. 756 757.
[53] Doyle, J.C. (1982). \Analysis of feedback systems with structured uncertainties,"
IEE Proceedings, Part D, Vol.133, pp.45 ; 56.
[54] Doyle, J. C. (1985). \Structured uncertainty in control system design", IEEE
CDC, Ft. Lauderdale.
[55] Doyle, J. C. (1984). \Lecture notes in advances in multivariable control,"
ONR/Honeywell Workshop, Minneapolis.
[56] Doyle, J. C., B. Francis, and A. Tannenbaum (1992). Feedback Control Theory,
Macmillan Publishing Company.
BIBLIOGRAPHY
567
[57] Doyle, J.C., K. Glover, P.P. Khargonekar, B.A. Francis (1989). \ State-space solutions to standard H2 and H1 control problems," IEEE Trans. Auto. Control, vol.
AC-34, No. 8, pp. 831-847. Also see 1988 American Control Conference, Atlanta,
June, 1988.
[58] Doyle, J., K. Lenz, and A. Packard (1986). \Design Examples using synthesis:
Space Shuttle Lateral Axis FCS during reentry," IEEE CDC, December 1986, pp.
2218-2223.
[59] Doyle, J.C., A.Packard and K. Zhou (1991). \Review of LFTs, LMIs and ," Proc.
30th IEEE Conf. Dec. Contr., England, pp. 1227-1232
[60] Doyle, J.C., R.S. Smith, and D.F. Enns (1987). \Control of Plants with Input
Saturation Nonlinearities," American Control Conf., pp 1034-1039.
[61] Doyle, J. C. and G. Stein (1981). \Multivariable feedback design: Concepts for
a classical/modern synthesis," IEEE Trans. Auto. Control, vol. AC-26, pp. 4-16,
Feb. 1981.
[62] Doyle, J. C., J. Wall and G. Stein (1982). \Performance and robustness analysis
for structured uncertainty," in Proc. 21st IEEE Conf. Decision Contr., pp. 629636.
[63] Doyle, J. C., K. Zhou, and B. Bodenheimer (1989). \Optimal Control with mixed
H2 and H1 performance objective," Proc. of American Control Conference, Pittsburgh, Pennsylvania, pp. 2065-2070.
[64] Doyle, J. C., K. Zhou, K. Glover, and B. Bodenheimer (1994). \Mixed H2 and H1
performance objectives II: optimal control," IEEE Trans. on Automat. Contr., vol.
39, no. 8, pp. 1575-1587.
[65] El-Sakkary, A. (1985). \The gap metric: Robustness of stabilization of feedback
systems," IEEE Trans. Automat. Contr., Vol. 30, pp. 240-247.
[66] Enns, D. (1984a). Model Reduction for Control System Design, Ph.D. Dissertation, Department of Aeronautics and Astronautics, Stanford University, Stanford,
California.
[67] Enns, D. (1984b) \Model reduction with balanced realizations: An error bound
and a frequency weighted generalization," Proc. 23rd Conf. Dec. Contr. , Las
Vegas, NV.
[68] Fan, M. K. H. and A.L. Tits (1986). \Characterization and ecient computation
of the structured singular value," IEEE Trans. Auto. Control, vol. AC-31, no. 8,
pp. 734-743.
[69] Fan, M. K. H. and A.L. Tits, and J. C. Doyle (1991). \Robustness in the presence
of mixed parametric uncertainty and unmodeled dynamics," IEEE Trans. Auto.
Control, vol. AC-36, no. 1, pp. 25-38.
568
BIBLIOGRAPHY
[70] Foias, C. and A. Tannenbaum (1988). \On the four-block problem, I," Operator
Theory: Advances and Applications, vol. 32 (1988), pp. 93-112 and "On the four
block problem, II: the singular system," Integral Equations and Operator Theory,
vol. 11 (1988), pp. 726-767.
[71] Francis, B. A. (1987). A course in H1 control theory, Lecture Notes in Control
and Information Sciences, vol. 88.
[72] Francis, B.A. and J.C. Doyle (1987). \Linear control theory with an H1 optimality
criterion," SIAM J. Control Opt., vol. 25, pp. 815-844.
[73] Freudenberg, J.S. (1989). \Analysis and design for ill-conditioned plants, part 2:
directionally uniform weightings and an example," Int. J. Contr., vol. 49, no. 3,
pp. 873-903.
[74] Freudenberg, J.S. and D.P. Looze (1985). \Right half plane zeros and poles and
design tradeos in feedback systems," IEEE Trans. Auto. Contr., vol. AC-30, no.
6, pp. 555-565.
[75] Freudenberg, J.S. and D.P. Looze (1988). Frequency Domain Properties of Scalar
and Multivariable Feedback Systems, Lecture Notes in Contr. and Info. Science,
vol. 104, Berlin: Springer-Verlag, 1988.
[76] Freudenberg, J. S., D.P. Looze and J.B. Cruz (1982). \Robustness analysis using
singular value sensitivities", Int. J. Control, vol. 35, no. 1, pp. 95-116.
[77] Garnett, J. B. (1981). Bounded Analytic Functions, Academic Press.
[78] Georgiou, T. T. (1988). \On the computation of the gap metric," Syst. Contr.
Lett., Vol. 11, pp. 253-257.
[79] Georgiou, T. T. and M. C. Smith (1990). \Optimal robustness in the gap metric,"
IEEE Trans. Automat. Contr., Vol.AC-35, No. 6, pp. 673-686.
[80] Glover, K. (1984). \All optimal Hankel-norm approximations of linear multivariable systems and their L1 -error bounds," Int. J. Control, vol. 39, pp. 1115-1193,
1984.
[81] Glover, K. (1986). \Robust stabilization of linear multivariable systems: Relations
to approximation," Int. J. Contr., Vol. 43, No. 3, pp. 741-766.
[82] Glover, K. (1989). \A tutorial on Hankel-norm Approximation," in Data to Model,
J. C. Willems, Ed. New York: Springer-Verlag.
[83] Glover, K. (1986). \Multiplicative approximation of linear multivariable systems
with L1 error bounds, " Proc. Amer. Contr. Conf., Seattle, WA, pp. 1705-1709.
BIBLIOGRAPHY
569
[84] Glover, K. and J. Doyle (1988). \State-space formulae for all stabilizing controllers
that satisfy an H1 norm bound and relations to risk sensitivity," Systems and
Control Letters, vol. 11, pp. 167-172.
[85] Glover, K. and J. C. Doyle (1989). \A state space approach to H1 optimal control," in Three Decades of Mathematical Systems Theory: A Collection of Surveys
at the Occasion of the 50th Birthday of Jan C. Willems, H. Nijmeijer and J. M.
Schumacher (Eds.), Springer-Verlag, Lecture Notes in Control and Information
Sciences, vol. 135, 1989.
[86] Glover, K., D.J.N. Limebeer, J.C. Doyle, E.M. Kasenally, and M.G. Safonov
(1991). \A characterization of all solutions to the four block general distance
problem," SIAM J. Optimization and Control, Vol. 29, No. 2, pp. 283 324.
[87] Glover, K., D. J. N. Limebeer, and Y. S. Hung (1992). \A structured approximation problem with applications to frequency weighted model reduction, " IEEE
Trans. Automat. Contr., vol.AC-37, No. 4, pp. 447-465.
[88] Glover, K. and D. McFarlane (1989). \Robust stabilization of normalized coprime
factor plant descriptions with H1 bounded uncertainty," IEEE Trans. Automat.
Contr., vol. 34, no. 8, pp.821-830.
[89] Glover, K. and D. Mustafa (1989). \Derivation of the maximum entropy H1
controller and a state space formula for its entropy," Int. J. Control, vol. 50, pp.
899.
[90] Goddard, P. J. and K. Glover (1993). \Controller reduction: weights for stability
and performance preservation," Proceedings of the 32nd Conference on Decision
and Control, San Antonio, Texas, December 1993, pp. 2903-2908.
[91] Goddard, P. J. and K. Glover (1994). \Performance preserving frequency weighted
controller approximation: a coprime factorization approach," Proceedings of the
33nd Conference on Decision and Control, Orlando, Florida, December 1994.
[92] Gohberg, I., and S. Goldberg (1981). Basic Operator Theory. Birkhauser, Boston.
[93] Gohberg, I., P. Lancaster, and L. Rodman (1986). \On hermitian solutions of the
symmetric algebraic Riccati equation," SIAM J. Control and Optimization, Vol.
24, No. 6, November 1986, pp. 1323-1334.
[94] Golub, G. H., and C. F. Van Loan (1983). Matrix Computations. Baltimore, Md.:
Johns Hopkind Univ. Press.
[95] Green, M. (1988a). \A relative error bound for balanced stochastic truncation,"
IEEE Trans. Automat. Contr., Vol.AC-33, No. 10, pp. 961-965.
[96] Green, M (1988b). \Balanced stochastic realizations," Journal of Linear Algebra
and its Applications, Vol. 98, pp. 211-247.
570
BIBLIOGRAPHY
[97] Green, M., K. Glover, D.J.N. Limebeer and J.C. Doyle (1988), \A J-spectral
factorization approach to H1 control," SIAM J. Control and Optimization, Vol.
28, pp. 1350-1371.
[98] Hautus, M. L. J. and L. M. Silverman (1983). \System structure and singular
control," Linear Algebra and its Applications, Vol. 50, pp. 369-402.
[99] Helton, J. W. (1981). \Broadband gain equalization directly from data," IEEE
Trans. Circuits Syst., Vol. CAS-28, pp. 1125-1137.
[100] Hestenes, M. (1975). Optimization Theory: the nite dimensional case, John Wiley & Sons.
[101] Heymann, M. (1968). \Comments \On pole assignment in multi-input controllable
linear systems," " IEEE Trans. Auto. Contr., Vol. 13, No. 6, pp. 748-749.
[102] Hinrichsen, D., and A. J. Pritchard (1990). \An improved error estimate for
reduced-order models of discrete time system," IEEE Transactions on Automatic
Control, AC-35, pp. 317-320.
[103] Homan, K. (1962). Banach Spaces of Analytic Functions, Prentice-hall, Inc.,
Englewood Clis, N.J.
[104] Horn, R. A. and C. R. Johnson (1990). Matrix Analysis, Cambridge University
Press, First Published in 1985.
[105] Horn, R. A. and C. R. Johnson (1991). Topics in Matrix Analysis, Cambridge
University Press.
[106] Horowitz, I. M. (1963). Synthesis of Feedback Systems, Academic Press, London.
[107] Hung, Y. S. (1989).\H1 -optimal control- Part I: model matching, - Part II:
solution for controllers," Int. J. Control, Vol. 49, pp. 1291-1359.
[108] Hung, Y. S. and K. Glover (1986). \Optimal Hankel-norm approximation of stable
systems with rst order stable weighting functions," Syst. Contr. Lett., Vol. 7 pp.
165-172.
[109] Hung, Y. S. and A. G. J. MacFarlane (1982). Multivariable feedback: A quasiclassical approach, Springer-Verlag.
[110] Hyde, R. A. and K. Glover (1993). \The Application of Scheduled H1 Controllers
to a VSTOL Aircraft," IEEE Trans. Automatic Control, Vol. 38, No. 7, pp. 10211039.
[111] Hyland, D. C. and D. S. Bernstein (1984). \The optimal projection equations for
xed-order dynamic compensation," IEEE Trans. Auto. Contr., Vol. AC-29, N0.
11, pp. 1034 1037.
BIBLIOGRAPHY
571
572
BIBLIOGRAPHY
[127] Kung, S. K., B. Levy, M. Morf and T. Kailath (1977). \New results in 2-D Systems
Theory, 2-D State-Space Models { Realization and the Notions of Controllability,
Observability and Minimality," Proc. IEEE, pp. 945 961.
[128] Kung, S. K. and D. W. Lin (1981). \Optimal Hankel norm model reduction:
Multivariable systems." IEEE Trans. Automat. Contr., Vol. AC-26, pp. 832-852.
[129] Kwakernaak, H. (1986). \A polynomial approach to minimax frequency domain
optimization of multivariable feedback systems," Int. J. Control, pp. 117-156.
[130] Kwakernaak, H. and R. Sivan (1972). Linear Optimal Control Systems, WileyInterscience, New York.
[131] Kucera, V. (1972). \A contribution to matrix quadratic equations," IEEE Trans.
Auto. Control, AC-17, No. 3, 344-347.
[132] Kucera, V. (1979). Discrete Linear Control, John Wiley and Sons, New York.
[133] Lancaster, P. and M. Tismenetsky (1985). The Theory of Matrices: with applications, 2nd Ed., Academic Press.
[134] Latham, G. A. and B. D. O. Anderson (1986). \Frequency weighted optimal
Hankel-norm approximation of stable transfer function," Syst. Contr. Lett., Vol.
5 pp. 229-236.
[135] Laub, S. J. (1980). \Computation of 'balancing' transformations," Proc. 1980
JACC, Session FA8-E.
[136] Lenz, K. E., P. P. Khargonekar, and John C. Doyle (1987). \Controller order
reduction with guaranteed stability and performance, " American Control Conference, pp. 1697-1698.
[137] Limebeer, D. J. N., B. D. O. Anderson, P. P. Khargonekar, and M. Green (1992).
\A game theoretic approach to H1 control for time varying systems," SIAM J.
Control and Optimization, Vol. 30, No. 2, pp. 262-283.
[138] Limebeer, D. J. N., M. Green, and D. Walker (1989). \Discrete time H1 control,"
Proc. IEEE Conf. Dec. Contr., Tampa, Florida, pp. 392-396.
[139] Limebeer, D.J.N. and G.D. Halikias (1988). \A controller degree bound for H1 optimal control problems of the second kind," SIAM J. Control Opt., vol. 26, no.
3, pp. 646-677.
[140] Limebeer, D.J.N. and Y.S. Hung (1987). \An analysis of pole-zero cancelations
in H1 -optimal control problems of the rst kind," SIAM J. Control Opt., vol.25,
pp. 1457-1493.
[141] Liu, K. Z., T. Mita, and R. Kawtani (1990). \Parameterization of state feedback
H1 controllers," Int. J. Contr., Vol.51, No. 3, pp. 535-551.
BIBLIOGRAPHY
573
[142] Liu, K. Z. and T. Mita (1992). \Generalized H1 control theory," Proc. American
Control Conference.
[143] Liu, Y. and B. D. O. Anderson (1986). \Controller reduction via stable factorization and balancing," Int. J. Control, Vol. 44, pp. 507-531.
[144] Liu, Y. and B. D. O. Anderson (1989). \Singular perturbation approximation of
balanced systems," Int. J. Control, Vol. 50, No. 4, pp. 1379-1405.
[145] Liu, Y. and B. D. O. Anderson (1990). \Frequency weighted controller reduction
methods and loop transfer recovery," Automatica, Vol. 26, No. 3, pp. 487-497.
[146] Liu, Y., B. D. O. Anderson, and U. Ly (1990). \Coprime factorization controller
reduction with Bezout identity induced frequency weighting," Automatica, Vol.
26, No. 2, pp. 233-249, 1990.
[147] Lu, W. M., K. Zhou, and J. C. Doyle (1991). \Stabilization of LFT systems,"
Proc. 30th IEEE Conf. Dec. Contr., Brighton, England, pp. 1239-1244.
[148] Luenberger, D. G. (1969). Optimization by Vector Space Methods, John Wiley &
Sons, Inc.
[149] Luenberger, D. G. (1971). \An introduction tp observers," IEEE Trans. Automat.
Contr., Vol. 16, No. 6, pp. 596-602.
[150] MacFarlane, A.G.J. and J.S. Hung (1983). \Analytic Properties of the singular
values of a rational matrix," Int. J. Contr., vol. 37, no. 2, pp. 221-234.
[151] Maciejowski, J. M. (1989). Multivariable Feedback Design, Wokingham: AddisonWesley.
[152] Mageirou, E. F. and Y. C. Ho (1977). \Decentralized stabilization via game theoretic methods," Automatica, Vol. 13, pp. 393-399.
[153] Mariton, M. and R. Bertrand (1985). \A homotopy algorithm for solving coupled
Riccati equations,"Optimal. Contr. Appl. Meth., vol. 6, pp. 351-357.
[154] Martensson, K. (1971). \On the matrix Riccati equation," Information Sciences,
Vol. 3, pp. 17-49.
[155] McFarlane, D. C., K. Glover, and M. Vidyasagar (1990). \Reduced-order controller design using coprime factor model reduction," IEEE Trans. Automat.
Contr., Vol. 35, No. 3, pp. 369-373.
[156] McFarlane, D. C. and K. Glover (1990). Robust Controller Design Using Normalized Coprime Factor Plant Descriptions, vol. 138, Lecture Notes in Control and
Information Sciences. Springer-Verlag.
574
BIBLIOGRAPHY
[157] McFarlane, D. C. and K. Glover (1992). \A loop shaping design procedure using
H1 synthesis," IEEE Trans. Automat. Contr., Vol. 37, No. 6, pp. 759-769.
[158] Meyer, D. G. (1990). \The construction of special coprime factorizations in discrete time," IEEE Trans. Automat. Contr., Vol. 35, No. 5, pp. 588-590.
[159] Meyer, D. G. (1990). \Fractional balanced reduction: model reduction via fractional representation," IEEE Trans. Automat. Contr., Vol. 35, No. 12, pp. 13411345.
[160] Meyer, D. G. and G. Franklin (1987). \A connection between normalized coprime factorizations and linear quadratic regulator theory," IEEE Trans. Automat. Contr., Vol. 32, pp. 227-228.
[161] Molinari, B. P. (1975). \The stabilizing solution of the discrete algebraic Riccati
equation," IEEE Trans. Auto. Contr., June 1975, pp. 396 399.
[162] Moore, B. C. (1981). \Principal component analysis in linear systems: controllability, observability, and model reduction," IEEE Trans. Auto. Contr., Vol. 26,
No.2.
[163] Moore, J. B., K. Glover, and A. Telford (1990). \All stabilizing controllers as
frequency shaped state estimate feedback," IEEE Trans. Auto. Contr., Vol. AC35, pp. 203 208.
[164] Mullis, C. T. and R. A. Roberts (1976). \Synthesis of minimum roundo noise
xed point digital lters," IEEE Trans. Circuits and Systems, Vol. 23, pp. 551-562.
[165] Mustafa, D. (1989). \Relations between maximum entropy / H1 control and
combined H1 /LQG control," Systems and Control Letters, vol. 12, no. 3, p. 193.
[166] Mustafa, D. and K. Glover (1990). Minimum Entropy H1 Control, Lecture notes
in control and information sciences, Springer-Verlag.
[167] Mustafa, D. and K. Glover (1991). \Controller Reduction by H1 - Balanced Truncation", IEEE Trans. Automat. Contr., vol.AC-36, No. 6, pp. 669-682.
[168] Naylor, A. W., and G. R. Sell (1982). Linear Operator Theory in Engineering and
Science, Springer-Verlag.
[169] Nagpal, K. M. and P. P. Khargonekar (1991). \Filtering and smoothing in an H1
setting," IEEE Trans. Automat. Contr., Vol.AC-36, No. 2, pp. 152-166.
[170] Nehari, Z. (1957). \On bounded bilinear forms," Annals of Mathematics, Vol. 15,
No. 1, pp. 153-162.
[171] Nesterov, Y. and A. Nemirovski (1994). Interior-point polynomial algorithms in
convex programming, SIAM.
BIBLIOGRAPHY
575
[172] Nett, C.N., C.A.Jacobson and N.J.Balas (1984). \A Connection between StateSpace and Doubly Coprime Fractional Representations," IEEE Trans., Vol.AC-29,
pp.831 ; 832.
[173] Osborne, E.E. (1960). \On Preconditioning of Matrices," J. Assoc. Comp. Mach.,
7, pp. 338-345.
[174] Packard, A. (1991). Notes on . Unpublished lecture notes.
[175] Packard, A. and J. C. Doyle (1988). \Structured singular value with repeated
scalar blocks," 1988 ACC, Atlanta.
[176] Packard, A. and J. C. Doyle (1988). Robust Control of Multivariable and Large
Scale Systems, Final Technical Report for Air Force Oce of Scientic Research.
[177] Packard, A. and J. C. Doyle (1993), \The complex structured singular value,"
Automatica, Vol. 29, pp. 71-109.
[178] Packard, A. and P. Pandey (1993). Continuity properties of the real/complex
structured singular value," IEEE Trans. Automat. Contr., Vol. 38, No. 3, pp.
415-428.
[179] Packard, A., K. Zhou, P. Pandey, J. Leonhardson, G. Balas (1992). \Optimal,
Constant I/O similarity scaling for full information and state feedback control
problems," Syst. Contr. Lett., Vol. 19, No. 4, pp. 271-280.
[180] Parrott, S. (1978). \On a quotient norm and the Sz-Nagy Foias lifting theorem,"
Journal of Functional Analysis, vol. 30, pp. 311 328.
[181] Pernebo, L. and L. M. Silverman (1982). "Model Reduction via Balanced State
Space Representation," IEEE Trans. Automat. Contr., vol.AC-27, No. 2, pp. 382387.
[182] Petersen, I.R. (1987). \Disturbance attenuation and H1 optimization: a design
method based on the algebraic Riccati equation," IEEE Trans. Auto. Contr., vol.
AC-32, pp. 427-429.
[183] Postlethwaite, I., J. M. Edmunds, and A. G. J. MacFarlane (1981). \Principal
gains and Principal phases in the analysis of linear multivariable feedback systems," IEEE Trans. Auto. Contr., Vol. AC-26, No. 1, pp. 32-46.
[184] Postlethwaite, I. and A. G. J. MacFarlane (1979). A complex variable approach to
the analysis of linear multivariable feedback systems. Berlin: Springer.
[185] Power, S. C. (1982). Hankel operators on Hilbert space. Pitman Advanced Publishing Program.
576
BIBLIOGRAPHY
BIBLIOGRAPHY
577
[201] Safonov, M.G. and D.J.N. Limebeer, and R. Y. Chiang (1990). \Simplifying the
H1 Theory via Loop Shifting, matrix pencil and descriptor concepts," Int. J.
contr., Vol 50, pp. 2467 2488.
[202] Sarason, D. (1967). \Generalized interpolation in H1 ," Trans. AMS., vol. 127,
pp. 179-203.
[203] Scherer, C. (1992). \H1 control for plants with zeros on the imaginary axis,"
SIAM J. Contr. Optimiz., Vol. 30, No. 1, pp. 123-142.
[204] Scherer, C. (1992). \H1 optimization without assumptions on nite or innite
zeros," SIAM J. Contr. Optimiz., Vol. 30, No. 1, pp. 143-166.
[205] Silverman, L. M. (1969). \Inversion of multivariable linear systems," IEEE Trans.
Automat. Contr., Vol. AC-14, pp. 270-276.
[206] Skelton, R. E. (1988). Dynamic System Control, John Willey & Sons.
[207] Stein, G. and J. C. Doyle (1991). \Beyond singular values and loop shapes," AIAA
Journal of Guidance and Control, vol. 14, no.1.
[208] Stein, G., G. L. Hartmann, and R. C. Hendrick (1977). \Adaptive control laws
for F-8c
ight test," IEEE Trans. Auto. Contr.
[209] Stoorvogel, A. A. (1992). The H1 Control Problem: A State Space Approach,
Prentice-Hall, Englewood Clis.
[210] Tadmor, G. (1990). \Worst-case design in the time domain: the maximum principle and the standard H1 problem," Mathematics of Control, Systems and Signal
Processing, pp. 301-324.
[211] Tits, A. L. (1994). \The small theorem with non-rational uncertainty," preprint.
[212] Tits, A. L. and M. K. H. Fan (1994). \On the small theorem," preprint.
[213] Vidyasagar, M. (1984). \The graph metric for unstable plants and robustness
estimates for feedback stability," IEEE Trans. Automatic Control, Vol. 29, pp.403417.
[214] Vidyasagar, M. (1985). Control System Synthesis: A Factorization Approach. MIT
Press, Cambridge, MA.
[215] Vidyadagar, M. (1988). \Normalized coprime factorizations for non-strictly proper
systems," IEEE Trans. Automatic Control, Vol. 33, No. 3, pp. 300-301.
[216] Vidyadagar, M. and H. Kimura (1986). \Robust controllers for uncertain linear
multivariable systems," Automatica, Vol. 22, pp. 85-94.
[217] Vinnicombe, G. (1993). \Frequency domain uncertainty and the graph topology,"
IEEE Trans. Automatic Control, Vol. 38, No. 9, pp. 1371-1383.
578
BIBLIOGRAPHY
[218] Walker, D. J. (1990). \Robust stabilizability of discrete-time systems with normalized stable factor perturbation," Int. J. Control, Vol. 52, No.2, pp.441-455.
[219] Wall, J.E., Jr., J.C. Doyle, and C.A. Harvey (1980). \Tradeos in the design of
multivariable feedback systems," Proc. 18th Allerton Conf., pp. 715-725.
[220] Wang, W., J.C.Doyle and K.Glover (1991). \Stability and Model Reduction of
Linear Fractional Transformation System," 1991 IEEE CDC, England.
[221] Wang, W. and M. G. Safonov (1990). \A tighter relative-error bound for balanced
stochastic truncation," Syst. Contr. Lett., Vol 14, pp. 307-317.
[222] Wang, W. and M. G. Safonov (1992). \Multiplicative-error bound for balanced
stochastic truncation model reduction," IEEE Trans. Automat. Contr., vol.AC37, No. 8, pp. 1265-1267.
[223] Whittle, P. (1981). \Risk-sensitive LQG control," Adv. Appl. Prob., vol. 13, pp.
764-777.
[224] Whittle, P. (1986). \A risk-sensitive certainty equivalence principle," in Essays
in Time Series and Allied Processes (Applied Probability Trust, London), pp
383-388.
[225] Willems, J.C. (1971). \Least Squares Stationary Optimal Control and the Algebraic Riccati Equation", IEEE Trans. Auto. Control, vol. AC-16, no. 6, pp.
621-634.
[226] Willems, J. C. (1981). \Almost invariant subspaces: an approach to high gain
feedback design { Part I: almost controlled invariant subspaces," IEEE Trans.
Auto. Control, vol. AC-26, pp. 235-252.
[227] Willems, J. C., A. Kitapci and L. M. Silverman (1986). \Singular optimal control:
a geometric approach," SIAM J. Control Optimiz., Vol. 24, pp. 323-337.
[228] Wimmer, H. K. (1985). \Monotonicity of Maximal solutions of algebraic Riccati
equations," Systems and Control Letters, Vol. 5, April 1985, pp. 317-319.
[229] Wonham, W. M. (1968). \On a matrix Riccati equation of stochastic control,"
SIAM J. Control, Vol. 6, No. 4, pp. 681-697.
[230] Wonham, W.M. (1985). Linear Multivariable Control: A Geometric Approach,
third edition, Springer-Verlag, New York.
[231] Youla, D.C. and J.J. Bongiorno, Jr. (1985). \A feedback theory of two-degree-offreedom optimal Wiener-Hopf design," IEEE Trans. Auto. Control, vol. AC-30,
No. 7, pp. 652-665.
[232] Youla, D.C., H.A. Jabr, and J.J. Bongiorno (1976). \Modern Wiener-Hopf design
of optimal controllers: part I," IEEE Trans. Auto. Control, vol. AC-21, pp. 3-13.
BIBLIOGRAPHY
579
[233] Youla, D.C., H.A. Jabr, and J.J. Bongiorno (1976). \Modern Wiener-Hopf design
of optimal controllers: part II," IEEE Trans. Auto. Control, vol. AC-21, pp. 319338.
[234] Youla, D.C., H.A. Jabr, and C. N. Lu (1974). \Single-loop feedback stabilization
of linear multivariable dynamical plants," Automatica, Vol. 10, pp. 159-173.
[235] Youla, D. C. and M. Saito (1967). \Interpolation with positive-real functions",
Journal of The Franklin Institute, Vol. 284, No. 2, pp. 77-108.
[236] Young, P. M. (1993). Robustness with Parametric and Dynamic Uncertainty, PhD
Thesis, California Institute of Technology.
[237] Yousu, A. and R. E. Skelton (1984). \A note on balanced controller reduction,"
IEEE Trans. Automat. Contr., vol.AC-29, No. 3, pp. 254-257.
[238] Zames, G. (1966). \On the input-output stability of nonlinear time-varying feedback systems, parts I and II," IEEE Trans. Auto. Control, vol. AC-11, pp. 228
and 465.
[239] Zames, G. (1981). \Feedback and optimal sensitivity: model reference transformations, multiplicative seminorms, and approximate inverses," IEEE Trans. Auto.
Control, vol. AC-26, pp. 301-320.
[240] Zames, G. and A. K. El-Sakkary (1980). \Unstable systems and feedback: The
gap metric," Proc. Allerton Conf., pp. 380-385.
[241] Zhou, K. (1992). \Comparison between H2 and H1 controllers," IEEE Trans. on
Automatic Control, Vol. 37, No. 8, pp. 1261-1265.
[242] Zhou, K. (1992). \On the parameterization of H1 controllers," IEEE Trans. on
Automatic Control, Vol. 37, No. 9, pp. 1442-1445.
[243] Zhou, K. (1993). \Frequency weighted model reduction with L1 error bounds,"
Syst. Contr. Lett., Vol. 21, 115-125, 1993.
[244] Zhou, K. (1993). \Frequency weighted L1 norm and optimal Hankel norm model
reduction," submitted to IEEE Trans. on Automatic Control .
[245] Zhou, K. and P. P. Khargonekar (1988). \An algebraic Riccati equation approach
to H1 optimization," Systems and Control Letters, vol. 11, pp. 85-92, 1988.
[246] Zhou, K., J. C. Doyle, K. Glover, and B. Bodenheimer (1990). \Mixed H2 and H1
control," Proc. of the 1990 American Control Conference, San Diego, California,
pp. 2502-2507.
[247] Zhou, K., K. Glover, B. Bodenheimer, and J. Doyle (1994). \Mixed H2 and H1
performance objectives I: robust performance analysis," IEEE Trans. on Automat.
Contr., vol. 39, no. 8, pp. 1564-1574.
580
BIBLIOGRAPHY
[248] Zhu, S. Q. (1989). \Graph topology and gap topology for unstable systems," IEEE
Trans. Automat. Contr., Vol. 34, NO. 8, pp. 848-855.
Index
Bode's gain and phase relation, 155
bounded real function, 353
discrete time, 532
bounded real lemma, 353
discrete time, 532
Cauchy sequence, 90
Cauchy-Schwartz inequality, 92
causal function, 175
causality, 175
Cayley-Hamilton theorem, 21, 48, 51,
54
central solution
H1 control, 411
matrix dilation problem, 41
classical control, 232
co-inner function, 357
compact operator, 178
complementary inner function, 358
complementary sensitivity, 128, 142
conjugate system, 66
contraction, 408
controllability, 45
Gramian, 71
matrix, 47
operator, 175
controllable canonical form, 60
controller parameterization, 293
controller reduction, 485
additive, 486
coprime factors, 489
H1 performance, 493
coprime factor uncertainty, 222, 467
coprime factorization, 123, 316, 358, 467
582
discrete time, 539
normalized, 470
cyclic matrix, 21, 61
design limitation, 153
design model, 211
design tradeo, 141
destabilizing perturbation, 268
detectability, 45, 51
dierential game, 459
direct sum, 94
discrete algebraic Riccati equation, 523
discrete time, 523
coprime factorization, 539
H1 control, 548
H2 control, 543
inner-outer factorization, 540
Lyapunov equation, 523
model reduction, 553
normalized coprime factorization, 541
disturbance feedforward (DF), 10, 294,
298
H1 control, 423
H2 control, 386
D ; K iteration, 289
dom(Ric), 327
double coprime factorization, 123
dual system, 66
duality, 297
eigenvalue, 20
eigenvector, 20, 324
generalized, 21
lower rank generalized eigenvector,
21
entropy, 435
equivalence class, 305
equivalence relation, 305
equivalent control, 305
F` , 242
Fu , 242
INDEX
feedback, 116
ltering
H1 performance, 452
xed order controller, 516
Fourier transform, 96
frequency weighting, 135
frequency-weighted balanced realization,
164
Frobenius norm, 29
full control (FC), 10, 294, 298
H1 control, 422
H2 control, 387, 546
full information (FI), 10, 294, 297
H1 control, 416
H2 control, 385, 546
full rank, 18
column, 18
row, 18
gain, 155
gain margin, 232
gap metric, 484
generalized eigenvector, 26, 27, 324
generalized inverse, 35
generalized principal vector, 525
Gilbert's realization, 69
Gramian
controllability, 71, 74
observability, 71, 74
graph metric, 484
graph topology, 484
Hamiltonian matrix, 321
Hankel norm, 173
Hankel norm approximation, 173, 188
Hankel operator, 173, 393
mixed Hankel-Toeplitz operator, 397
Hankel singular value, 75, 162, 173
Hardy spaces, 95
harmonic function, 142
Hermitian matrix, 21
H1 control, 405
INDEX
discrete time, 548
loop shaping, 467
singular problem, 448
state feedback, 459
H1 ltering, 452
H1 optimal controller, 430, 437
H1 performance, 135, 138
H1 space, 89, 97
H1; space, 97
hidden modes, 77
Hilbert space, 91
homotopy algorithm, 520
H2 optimal control 365
discrete time, 543
H2 performance, 135
H2 space, 89, 95
H2 stability margin, 389
H2? space, 96
H2 (@ D ) space, 199
H2? (@ D ) space, 199
Hurwitz, 49
image, 18
induced norm, 28, 101
inertia, 179
inner function, 357
discrete time, 537
inner product, 92
inner-outer factorization, 143, 358
input sensitivity, 128
input-output stability, 406
integral control, 448
H1 control, 448
H2 control, 448
internal model principle, 448
internal stability, 119, 406
invariant subspace, 26, 322
invariant zero, 84, 333
inverse
of a transfer function, 66
isometric isomorphism, 91, 174
Jordan canonical form, 20
583
Kalman canonical decomposition, 52
kernel, 18
Kronecker product, 25
Kronecker sum, 25
Lagrange multiplier, 511
left coprime factorization, 123
linear combination, 17
linear fractional transformation (LFT),
241
linear operator, 91
L1 space, 96
loop gain, 130
loop shaping, 132
H1 approach, 467
normalized coprime factorization, 475
loop transfer matrix (function), 128
LQG stability margin, 389
LQR problem, 367
LQR stability margin, 373
l2 (;1; 1) space, 93
L2 space, 95
L2 (;1; 1) space, 93
L2 (@ D ) space, 199
Lyapunov equation, 26, 70
discrete time, 523
main loop theorem, 275
matrix
compression, 38
dilation, 38, 445
factorization, 343, 537
Hermitian, 21
inequality, 335
inertia, 179
inversion formulas, 22
norm, 27
square root of a, 36
maximum modulus theorem, 94
max-min problem, 400
McMillan degree, 81
McMillan form, 80
INDEX
584
minimal realization, 67, 73
minimax problem, 400
minimum entropy controller, 435
mixed Hankel-Toeplitz operator, 397
modal controllability, 51
modal observability, 51
model reduction
additive, 157
multiplicative, 157
relative, 157
model uncertainty, 116, 211
, 263
lower bound, 273
synthesis, 288
upper bound, 273
multidimensional system, 249
multiplication operator, 98
multiplicative approximation, 157, 165
multiplicative uncertainty, 213, 220
Nehari's Theorem, 203
nominal performance (NP), 215
nominal stability (NS), 215
non-minimum phase zero, 141
norm, 27
Frobenius, 29
induced, 28
semi-norm, 28
normal rank, 79
normalized coprime factorization, 362
loop shaping, 475
null space, 18
Nyquist stability theorem, 123
observability, 45
Gramian, 71
operator, 175
observable canonical form, 61
observable mode, 51
observer, 62
observer-based controller, 62
operator
extension, 91
restriction, 91
optimal Hankel norm approximation, 188
optimality of H1 controller, 430
optimization method, 511
orthogonal complement, 18, 94
orthogonal direct sum, 94
orthogonal matrix, 18
orthogonal projection theorem, 94
outer function, 358
output estimation (OE), 10, 294, 298
H1 control, 425
H2 control, 387
output injection, 298
H2 control, 386
output sensitivity, 128
Parrott's theorem, 40
Parseval's relations, 96
PBH (Popov-Belevitch-Hautus) tests, 51
performance limitation, 141
phase, 155
phase margin, 232
Plancherel theorem, 96
plant condition number, 230
Poisson integral, 141
pole, 77, 79
pole direction, 143
pole placement, 57
pole-zero cancelation, 77
positive real, 354
positive (semi-)denite matrix, 36
power signal, 100
pseudo-inverse, 35
quadratic control, 365
quadratic performance, 365, 393
Rp (s) , 80
range, 18
realization, 67
balanced, 75
INDEX
input normal, 77
minimal, 67
output normal, 77
Redheer star-product, 259
regular point, 512
regulator problem, 365
relative approximation, 157, 165
return dierence, 128
RH1
space, 97
RH;1 space, 97
RH2 space, 95
RH?2 space, 96
Riccati equation, 321
Riccati operator, 327
right coprime factorization, 123
risk sensitive, 436
robust performance (RP), 215
H1 performance, 226, 275
H2 performance, 226
structured uncertainty, 279
robust stability (RS), 215
structured uncertainty, 278
robust stabilization, 467
roll-o rate, 144
Schmidt pair, 178
Schur complement, 23
self-adjoint operator, 94, 178
sensitivity function, 128, 142
bounds, 153
integral, 144
separation theory, 310
H1 control, 426
H2 control, 388
simplectic matrix, 525
singular H1 problem, 448
singular value, 32
singular value decomposition (SVD), 31
singular vector, 32
skewed performance specication, 230
small gain theorem, 215
Smith form, 79
span, 17
585
spectral factorization, 343
spectral radius, 20
spectral signal, 100
stability, 45, 49
internal, 119
stability margin
LQG, H2 , 389
stabilizability, 45, 49
stabilization, 293
stabilizing controller, 293
stable invariant subspace, 27, 327, 525
star-product, 259
state feedback, 298
H1 control, 459
H2 control, 382
state space realization, 60
strictly positive real, 354
structured singular value, 263
lower bound, 273
upper bound, 273
structured uncertainty, 263
subharmonic function, 142
supremum norm, 89
Sylvester equation, 26
Sylvester's inequality, 19, 68
Toeplitz operator, 197, 395
trace, 19
tradeo, 141
transition matrix, 46
transmission zero, 81
uncertainty, 116, 211
state space, 252
unimodular matrix, 79
unitary matrix, 18
unstructured uncertainty, 211
weighted model reduction, 164, 500
weighting, 135
well-posedness, 117, 243
winding number, 409
586
Youla parameterization, 303, 316, 454
zero, 77, 79
blocking, 81
direction, 142
invariant, 84
transmission, 81
INDEX