122,99 €
A bridge between the application of subspace-based methods for parameter estimation in signal processing and subspace-based system identification in control systems Model-Based Processing: An Applied Subspace Identification Approach provides expert insight on developing models for designing model-based signal processors (MBSP) employing subspace identification techniques to achieve model-based identification (MBID) and enables readers to evaluate overall performance using validation and statistical analysis methods. Focusing on subspace approaches to system identification problems, this book teaches readers to identify models quickly and incorporate them into various processing problems including state estimation, tracking, detection, classification, controls, communications, and other applications that require reliable models that can be adapted to dynamic environments. The extraction of a model from data is vital to numerous applications, from the detection of submarines to determining the epicenter of an earthquake to controlling an autonomous vehicles--all requiring a fundamental understanding of their underlying processes and measurement instrumentation. Emphasizing real-world solutions to a variety of model development problems, this text demonstrates how model-based subspace identification system identification enables the extraction of a model from measured data sequences from simple time series polynomials to complex constructs of parametrically adaptive, nonlinear distributed systems. In addition, this resource features: * Kalman filtering for linear, linearized, and nonlinear systems; modern unscented Kalman filters; as well as Bayesian particle filters * Practical processor designs including comprehensive methods of performance analysis * Provides a link between model development and practical applications in model-based signal processing * Offers in-depth examination of the subspace approach that applies subspace algorithms to synthesized examples and actual applications * Enables readers to bridge the gap from statistical signal processing to subspace identification * Includes appendices, problem sets, case studies, examples, and notes for MATLAB Model-Based Processing: An Applied Subspace Identification Approach is essential reading for advanced undergraduate and graduate students of engineering and science as well as engineers working in industry and academia.
Sie lesen das E-Book in den Legimi-Apps auf:
Seitenzahl: 639
Veröffentlichungsjahr: 2019
Cover
Preface
References
Acknowledgements
Glossary
1 Introduction
1.1 Background
1.2 Signal Estimation
1.3 Model‐Based Processing
1.4 Model‐Based Identification
1.5 Subspace Identification
1.6 Notation and Terminology
1.7 Summary
MATLAB Notes
References
Problems
2 Random Signals and Systems
2.1 Introduction
2.2 Discrete Random Signals
2.3 Spectral Representation of Random Signals
2.4 Discrete Systems with Random Inputs
2.5 Spectral Estimation
2.6 Case Study: Spectral Estimation of Bandpass Sinusoids
2.7 Summary
Matlab Notes
References
Problems
3 State‐Space Models for Identification
3.1 Introduction
3.2 Continuous‐Time State‐Space Models
3.3 Sampled‐Data State‐Space Models
3.4 Discrete‐Time State‐Space Models
3.5 Gauss–Markov State‐Space Models
3.6 Innovations Model
3.7 State‐Space Model Structures
3.8 Nonlinear (Approximate) Gauss–Markov State‐Space Models
3.9 Summary
MATLAB Notes
References
Problems
4 Model‐Based Processors
4.1 Introduction
4.2 Linear Model‐Based Processor: Kalman Filter
4.3 Nonlinear State‐Space Model‐Based Processors
4.4 Case Study: 2D‐Tracking Problem
4.5 Summary
MATLAB Notes
References
Problems
5 Parametrically Adaptive Processors
5.1 Introduction
5.2 Parametrically Adaptive Processors: Bayesian Approach
5.3 Parametrically Adaptive Processors: Nonlinear Kalman Filters
5.4 Parametrically Adaptive Processors: Particle Filter
5.5 Parametrically Adaptive Processors: Linear Kalman Filter
5.6 Case Study: Random Target Tracking
5.7 Summary
MATLAB Notes
References
Problems
6 Deterministic Subspace Identification
6.1 Introduction
6.2 Deterministic Realization Problem
6.3 Classical Realization
6.4 Deterministic Subspace Realization: Orthogonal Projections
6.5 Deterministic Subspace Realization: Oblique Projections
6.6 Model Order Estimation and Validation
6.7 Case Study: Structural Vibration Response
6.8 Summary
MATLAB Notes
References
Problems
7 Stochastic Subspace Identification
7.1 Introduction
7.2 Stochastic Realization Problem
7.3 Classical Stochastic Realization via the Riccati Equation
7.4 Classical Stochastic Realization via Kalman Filter
7.5 Stochastic Subspace Realization: Orthogonal Projections
7.6 Stochastic Subspace Realization: Oblique Projections
7.7 Model Order Estimation and Validation
7.8 Case Study: Vibration Response of a Cylinder: Identification and Tracking
7.9 Summary
MATLAB NOTES
References
Problems
8 Subspace Processors for Physics‐Based Application
8.1 Subspace Identification of a Structural Device
8.2 MBID for Scintillator System Characterization
8.3 Parametrically Adaptive Detection of Fission Processes
8.4 Parametrically Adaptive Processing for Shallow Ocean Application
8.5 MBID for Chirp Signal Extraction
References
Appendix A: Probability and Statistics Overview
A.1 Probability Theory
A.2 Gaussian Random Vectors
A.3 Uncorrelated Transformation: Gaussian Random Vectors
A.4 Toeplitz Correlation Matrices
A.5 Important Processes
References
Appendix B: Projection Theory
B.1 Projections: Deterministic Spaces
B.2 Projections: Random Spaces
B.3 Projection: Operators
References
Appendix C: Matrix Decompositions
C.1 Singular‐Value Decomposition
C.2 QR‐Decomposition
C.3 LQ‐Decomposition
References
Appendix D: Output‐Only Subspace Identification
References
Index
End User License Agreement
Chapter 2
Table 2.1 Properties of covariance and spectral functions.
Table 2.2 Linear system with random inputs: covariance/spectrum relationships.
Chapter 3
Table 3.1 Gauss–Markov representation.
Table 3.2 ARMAX representation.
Table 3.3 Approximate nonlinear Gauss–Markov model.
Chapter 4
Table 4.1 Linear Kalman filter algorithm (predictor–corrector form).
Table 4.2 Heuristic notions for Kalman filter tuning.
Table 4.3 Statistical tuning tests for Kalman filter.
Table 4.4 Steady‐state Kalman filter algorithm.
Table 4.5 Linearized Kalman filter (LZKF) algorithm.
Table 4.6 Extended Kalman filter (EKF) algorithm.
Table 4.7 Iterated–extended Kalman filter (IEKF) algorithm.
Table 4.8 Discrete unscented Kalman filter (UKF) algorithm.
Table 4.9 Bootstrap particle filter (PF) SIR algorithm.
Chapter 5
Table 5.1 Augmented extended Kalman filter algorithm.
Table 5.2 Parametrically adaptive innovations model (RPE) algorithm.
Table 5.3 PF Performance towed array problem.
Chapter 6
Table 6.1 Ho–Kalman (SVD) realization algorithm
Table 6.2 Canonical realization algorithm.
Table 6.3 Multivariable output error state‐space algorithm (MOESP).
Table 6.4 Numerical algorithm for subspace state‐space system identification (N4...
Table 6.5 Deterministic subspace identification: three‐mass mechanical system.
Table 6.6 Deterministic subspace identification: eight‐mass mechanical system.
Chapter 7
Table 7.1 Iterative KSP algorithm (Riccati equation).
Table 7.2 Stochastic realization algorithm: Riccati equation approach.
Table 7.3 Stochastic realization algorithm: Kalman filter approach.
Table 7.4 Iterative KSP algorithm (Kalman filter).
Table 7.5 Stochastic realization algorithm: PAST OUTPUT‐MOESP (PO‐MOESP) approac...
Table 7.6 Stochastic subspace identification: six (6) sinusoidal system.
Table 7.7 Numerical algorithm for (stochastic) subspace state‐space system ident...
Table 7.8 Stochastic realization: statistical tests.
Table 7.9 Stochastic subspace identification: 3‐mass mechanical system.
Table 7.10 Stochastic subspace identification: six (6) modal frequencies.
Chapter 8
Table 8.1 Modal identification ensemble statistics.
2
Table B.1 Matrix Projections.
Chapter 1
Figure 1.1 Processing of a
deterministic
signal. (a) Raw data and spectrum with...
Figure 1.2 Processing of a
random
signal and noise. (a) Raw data and spectrum w...
Figure 1.3 Model‐based signal processing “model staircase”: Step
– simple mod...
Figure 1.4 Model‐based signal processing of a noisy measurement conceptually il...
Figure 1.5 Model‐based signal processing “model” for an RC‐circuit.
Figure 1.6 Model‐based processor for RC‐circuit: raw data (solid), true measure...
Figure 1.7 System identification problem: input (deterministic/random), disturb...
Figure 1.8 Model‐based identification of RC‐circuit parameters:
,
,
and
.
Figure 1.9 Identification of a sinusoidal model from noise‐free data: spectral ...
Chapter 2
Figure 2.1 Random signal comparison: random signal and spectrum and the determi...
Figure 2.2 Random signal processing: (a) Data and spectrum. (b) Covariance and ...
Figure 2.3 Ensemble of random walk realizations.
Figure 2.4 Signal and noise subspaces with two harmonic signal vectors and one ...
Figure 2.5 Bandpass sinusoidal problem (
,
,
,
,
Hz) spectral estimation: (a) ...
Chapter 3
Figure 3.1 Gauss–Markov model of a discrete process.
Figure 3.2 Gauss–Markov simulation of first‐order process: (a) state/measuremen...
Figure 3.3 Linearization of a deterministic system using the reference trajecto...
Chapter 4
Figure 4.1 Predictor–corrector form of the Kalman filter: Timing diagram.
Figure 4.2 LKF design for
‐circuit problem. (a) Estimated state (voltage) and ...
Figure 4.3 Nonlinear trajectory estimation: linearized Kalman filter (LZKF). (a...
Figure 4.4 Nonlinear trajectory estimation: extended Kalman filter (EKF). (a) E...
Figure 4.5 Nonlinear trajectory estimation: Iterated–extended Kalman filter (IE...
Figure 4.6 Nonlinear trajectory estimation: unscented Kalman filter (UKF). (a) ...
Figure 4.7 Particle filter representation of “empirical” posterior probability ...
Figure 4.8 PF surface (
vs
vs
) representation of posterior probability dis...
Figure 4.9 Nonlinear trajectory Bayesian estimation: Particle filter (PF) desig...
Figure 4.10 PMF data: Kullback–Leibler divergence and Hellinger distance calcul...
Figure 4.11 Tanker ground track geometry for the harbor docking application: (a...
Figure 4.12 EKF design for harbor docking problem (input known). (a)
‐position...
Figure 4.13 LZKF design for harbor docking problem (input known). (a)
‐positio...
Figure 4.14 UKF design for harbor docking problem (input unknown). (a)
‐positi...
Chapter 5
Figure 5.1 Parametrically adaptive processor structure illustrating the couplin...
Figure 5.2 Nonlinear (uncertain) trajectory estimation: parametrically adaptive...
Figure 5.3 Nonlinear (uncertain) trajectory estimation: Parametrically adaptive...
Figure 5.4 Nonlinear (uncertain) trajectory estimation: PF algorithm. (a) Estim...
Figure 5.5 PF posterior distribution estimation. (a) Estimated state posterior....
Figure 5.6 PF ensemble estimation for a
‐member realizations. (a) Ensemble sta...
Figure 5.7 Parametrically adaptive innovations state‐space processor: RC‐circui...
Figure 5.8 Synthetic aperture sonar tracking problem: simulated target motion f...
Figure 5.9 Particle filter bearing estimates for four targets in random motion:...
Figure 5.10 Particle filter predicted measurement estimates for four channel hy...
Figure 5.11 Particle filter classical “sanity” performance metrics: zero‐mean/w...
Figure 5.12 Particle filter instantaneous posterior bearing estimates: (a)
ta...
Chapter 6
Figure 6.1 Deterministic realization problem: classical approach and subspace a...
Figure 6.2 SVD realization. (a) Estimated singular values of the Hankel matrix....
Figure 6.3 The three‐output mechanical system under investigation.
Figure 6.4 Impulse response and spectrum of three‐output mechanical system. (a)...
Figure 6.5 SVD realization of three‐output mechanical system. (a) Singular valu...
Figure 6.6 Orthogonal projection operator: orthogonal complement operation.
Figure 6.7 Oblique projections: Projection
of future outputs (
) onto past da...
Figure 6.8 Extraction of future state and shifted state sequences: (a) Oblique ...
Figure 6.9 Pseudo‐random response of three‐output mechanical system. (a) Synthe...
Figure 6.10 SVD realization of three‐output mechanical system. (a) Power spectr...
Figure 6.11 SVD realization of three‐output mechanical system. (a) Raw and esti...
Figure 6.12 Structural vibrations of eight‐mode mechanical system: (a) accelero...
Figure 6.13 Structural vibration subspace identification eight‐mode mechanical ...
Figure 6.14 Subspace identification of eight‐mode mechanical system. (a) Raw da...
Chapter 7
Figure 7.1 Stochastic realization problem: (a) classical and (b) Subspace appro...
Figure 7.2 SVD order estimation of an unknown system from the Hankel covariance...
Figure 7.3 Simulated sinusoids in noise (
). (a) Noisy measurement. (b) Spectra...
Figure 7.4 Subspace identification of sinusoids in noise (
). (a) Spectral esti...
Figure 7.5 Comparison of identified sinusoids in noise (
). (a) Impulse respons...
Figure 7.6 Projections: orthogonal projection of
onto the joint subspaces
f...
Figure 7.7 Response of
‐output mechanical system to white noise excitation (
)...
Figure 7.8 N4SIDsubspace identification of
‐output mechanical system excited b...
Figure 7.9 N4SID‐subspace validation of
‐output mechanical system: signal esti...
Figure 7.10 N4SID‐subspace validation of
‐output mechanical system. (a) Channe...
Figure 7.11 Cylindrical object (copper pipe) vibrational response experiment: p...
Figure 7.12 Analysis of cylindrical object (pipe) single‐channel data: Fourier ...
Figure 7.13 Multichannel cylindrical object data: (a) Preprocessed triaxial cha...
Figure 7.14 Cylindrical object multichannel identification results: (a) Identif...
Figure 7.15 Stochastic realization of cylindrical object (pipe) power spectral:...
Figure 7.16 Subspace identification of cylindrical object,
‐mode model: (a) Es...
Figure 7.17 Postprocessing of modal frequency tracking (ensemble) statistics: (...
Figure 7.18 Postprocessed modal frequency tracking of cylindrical object data f...
Chapter 8
Figure 8.1 Structural device experimental setup: Motor‐driven stinger random vi...
Figure 8.2 Subspace identification of
‐mode model: (a) Estimated transfer func...
Figure 8.3 Postprocessing modal frequency tracking ensemble statistics: (a) Mod...
Figure 8.4 Postprocessed modal frequency tracking of (a)
‐mode, (b)
‐mode, (c...
Figure 8.5 Scintillator/photomultiplier system with equivalent circuit model.
Figure 8.6 Typical scintillator/photomultiplier response: (a) Pulse. (b) Bode p...
Figure 8.7 Cascaded scintillator/photomultiplier system model: (a) Diagram. (b)...
Figure 8.8 Deterministic state‐space model output for cascaded scintillator/pho...
Figure 8.9 Stochastic state‐space model output for cascaded scintillator/photom...
Figure 8.10 Subspace‐identified scintillator model: (a) Estimated and raw respo...
Figure 8.11 Parameter estimation for scintillator voltage response.
Figure 8.12 Estimated outputs for cascaded scintillator/photomultiplier system ...
Figure 8.13 Performance metrics for Kalman filter: (a) Innovation (
out). (b) ...
Figure 8.14 Kalman filter estimation for scintillator voltage response: (a) Pho...
Figure 8.15 Kalman filter performance for scintillator voltage response: (a) In...
Figure 8.16 As each individual interarrival is extracted, it is discriminated, ...
Figure 8.17 Fission detection/estimation processing: Count data, fission parame...
Figure 8.18 Sequential Bayesian (log‐likelihood) threat detection for
)=(
,
)...
Figure 8.19 Family of sequential Bayesian detection ROC curves for various SNRs...
Figure 8.20 Sequential Bayesian parameter estimates (arrows) and absolute RMS e...
Figure 8.21 Posterior PMF of various slices (
interarrivals) illustrating the ...
Figure 8.23 Modal function tracking with adaptive modal coefficient estimation:...
Figure 8.24 Adaptive modal coefficient parameter estimation data (MODEL) from t...
Figure 8.22 Raw/enhanced pressure‐field (DATA) data from the Hudson Canyon expe...
Figure 8.25 Raw pressure‐field data/enhanced data (DATA) from the Hudson Canyon...
Figure 8.26 Modal function tracking for adaptive wavenumber estimation: Hudson ...
Figure 8.27 Probability mass function (PMF) posterior estimation (mode
) surfa...
Figure 8.28 Probability mass function (PMF) posterior estimation (wavenumber
)...
Figure 8.29 True chirp and FSK spectra (20–30 kHz sweep): (a) Chirp spectrum. (...
Figure 8.30 Chirp data: (a) True chirp signal and disturbance with spectrum. (b...
Figure 8.31 Model‐based CHIRP identification: (a) Estimated chirp. (b) Linear c...
Figure 8.32 FSK data: (a) True FSK signal and disturbance with spectrum. (b) No...
Figure 8.33 Model‐based identification of FSK signal with disturbances: (a) Tru...
Cover
Table of Contents
Begin Reading
xiii
xiv
xv
xvi
xvii
xviii
20
21
xxi
xxiii
xxiv
xxv
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
185
186
187
188
189
190
191
192
193
194
195
197
196
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
286
285
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
362
361
363
364
365
366
367
368
369
370
371
372
373
374
375
376
378
377
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
477
478
479
480
481
482
483
484
485
486
487
488
489
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
James V. Candy
Lawrence Livermore National Laboratory University of California, Santa Barbara
This edition first published 2019
© 2019 John Wiley & Sons, Inc.
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by law. Advice on how to obtain permission to reuse material from this title is available at http://www.wiley.com/go/permissions.
The right of James V. Candy are to be identified as the author of this work has been asserted in accordance with law.
Registered Office
John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, USA
Editorial Office
111 River Street, Hoboken, NJ 07030, USA
For details of our global editorial offices, customer services, and more information about Wiley products visit us at www.wiley.com.
Wiley also publishes its books in a variety of electronic formats and by print‐on‐demand. Some content that appears in standard print versions of this book may not be available in other formats.
Limit of Liability/Disclaimer of Warranty
MATLAB® is a trademark of The MathWorks, Inc. and is used with permission. The MathWorks does not warrant the accuracy of the text or exercises in this book. This work's use or discussion of MATLAB® software or related products does not constitute endorsement or sponsorship by The MathWorks of a particular pedagogical approach or particular use of the MATLAB® software. While the publisher and authors have used their best efforts in preparing this work, they make no representations or warranties with respect to the accuracy or completeness of the contents of this work and specifically disclaim all warranties, including without limitation any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives, written sales materials or promotional statements for this work. The fact that an organization, website, or product is referred to in this work as a citation and/or potential source of further information does not mean that the publisher and authors endorse the information or services the organization, website, or product may provide or recommendations it may make. This work is sold with the understanding that the publisher is not engaged in rendering professional services. The advice and strategies contained herein may not be suitable for your situation. You should consult with a specialist where appropriate. Further, readers should be aware that websites listed in this work may have changed or disappeared between when this work was written and when it is read. Neither the publisher nor authors shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages.
Library of Congress Cataloging‐in‐Publication Data applied for
ISBN: 9781119457763
Cover design by Wiley
Cover image: Courtesy of Patricia Candy
Jesus replied, “I am the way, the truth and the life. No one comes to the Father except through me.” (John: 14:6)
This text encompasses the basic idea of the model‐based approach to signal processing by incorporating the often overlooked, but necessary, requirement of obtaining a model initially in order to perform the processing in the first place. Here we are focused on presenting the development of models for the design of model‐based signal processors (MBSP) using subspace identification techniques to achieve a model‐based identification (MBID) as well as incorporating validation and statistical analysis methods to evaluate their overall performance 1. It presents a different approach that incorporates the solution to the system identification problem as the integral part of the model‐based signal processor (Kalman filter) that can be applied to a large number of applications, but with little success unless a reliable model is available or can be adapted to a changing environment 2. Here, using subspace approaches, it is possible to identify the model very rapidly and incorporate it into a variety of processing problems such as state estimation, tracking, detection, classification, controls and communications to mention a few 3,4. Models for the processor evolve in a variety of ways, either from first principles accompanied by estimating its inherent uncertain parameters as in parametrically adaptive schemes 5 or by extracting constrained model sets employing direct optimization methodologies 6, or by simply fitting a black‐box structure to noisy data 7,8. Once the model is extracted from controlled experimental data, or a vast amount of measured data, or even synthesized from a highly complex truth model, the long‐term processor can be developed for direct application 1 . Since many real‐world applications seek a real‐time solution, we concentrate primarily on the development of fast, reliable identification methods that enable such an implementation 9–11. Model extraction/development must be followed by validation and testing to ensure that the model reliably represents the underlying phenomenology – a bad model can only lead to failure!
System identification 6 provides solutions to the problem of extracting a model from measured data sequences either time series, frequency data or simply an ordered set of indexed values. Models can be of many varieties ranging from simple polynomials to highly complex constructs evolving from nonlinear distributed systems. The extraction of a model from data is critical for a large number of applications evolving from the detection of submarines in a varying ocean, to tumor localization in breast tissue, to pinpointing the epicenter of a highly destructive earthquake, or to simply monitoring the condition of a motor as it drives a critical system component 1 . Each of these applications require an aspect of modeling and fundamental understanding (when possible) of the underlying phenomenology governing the process as well as the measurement instrumentation extracting the data along with the accompanying uncertainties. Some of these problems can be solved simply with a “black‐box” representation that faithfully reproduces the data in some manner without the need to capture the underlying dynamics (e.g. common check book entries) or a “gray‐box” model that has been extracted, but has parameters of great interest (e.g. unknown mass of a toxic material). However, when the true need exists to obtain an accurate representation of the underlying phenomenology like the structural dynamics of an aircraft wing or the untimely vibrations of a turbine in a nuclear power plant, then more sophisticated representations of the system and uncertainties are clearly required. In cases such as these, models that capture the dynamics must be developed and “fit” to the data in order to perform applications such as condition monitoring of the structure or failure detection/prediction of a rotating machine. Here models can evolve from lumped characterizations governed by sets of ordinary differential equations, linear or nonlinear, or distributed representations evolved from sets of partial differential equations. All of these representations have one thing in common, when the need to perform a critical task is at hand – they are represented by a mathematical model that captures their underlying phenomenology that must somehow be extracted from noisy measurements. This is the fundamental problem that we address in this text, but we must restrict our attention to a more manageable set of representations, since many monographs have addressed problem sets targeting specific applications 12,13.
In fact, this concept of specialty solutions leads us to the generic state‐space model of systems theory and controls. Here the basic idea is that all of the theoretical properties of a system are characterized by this fundamental set of models that enables the theory to be developed and then applied to any system that can be represented in the state‐space. Many models naturally evolve in the state‐space, since it is essentially the representation of a set of th‐order differential equations (ordinary or partial, linear or nonlinear, time (space) invariant or time (space) varying, scalar or multivariable) that are converted into a set of first‐order equations, each of which is a state. For example, a simple mechanical system consisting of a single mass, spring, damper construct is characterized by a set of second‐order, linear, time‐invariant, differential equations that can simply be represented in state‐space form by a set of two first‐order equations, each one representing a state: one for displacement and one for velocity 12 . We employ the state‐space representation throughout this text and provide sufficient background in Chapters 2 and 3.
System identification is broad in the sense that it does not limit the problem to various classes of models directly. For instance, for an unknown system, a model set is selected with some perception that it is capable of representing the underlying phenomenology adequately, then this set is identified directly from the data and validated for its accuracy. There is clearly a well‐defined procedure that captures this approach to solve the identification problem 6 –15. In some cases, the class structure of the model may be known a priori, but the order or equivalently the number of independent equations to capture its evolution is not (e.g. number of oceanic modes). Here, techniques to perform order estimation precede the fitting of model parameters first, then are followed by the parameter estimation to extract the desired model 14. In other cases, the order is known from prior information and parameter estimation follows directly (e.g. a designed mechanical structure). In any case, these constraints govern the approach to solving the identification problem and extracting the model for application. Many applications exist, where it is desired to monitor a process and track a variety of parameters as they evolve in time, (e.g. radiation detection), but in order to accomplish this on‐line, the model‐based processor must update the model parameters sequentially in order to accomplish its designated task. We develop these processors for both linear and nonlinear models in Chapters 4 and 5.
Although this proposed text is designed primarily as a graduate text, it will prove useful to practicing signal processing professionals and scientists, since a wide variety of case studies are included to demonstrate the applicability of the model‐based subspace identification approach to real‐world problems. The prerequisite for such a text is a melding of undergraduate work in linear algebra (especially matrix decomposition methods), random processes, linear systems, and some basic digital signal processing. It is somewhat unique in the sense that many texts cover some of its topics in piecemeal fashion. The underlying model‐based approach of this text is the thread that is embedded throughout in the algorithms, examples, applications, and case studies. It is the model‐based theme, together with the developed hierarchy of physics‐based models, that contributes to its uniqueness coupled with the new robust, subspace model identification methods that even enable potential real‐time methods to become a reality. This text has evolved from four previous texts, 1 , 5 and has been broadened by a wealth of practical applications to real‐world, model‐based problems. The introduction of robust subspace methods for model‐building that have been available in the literature for quite a while, but require more of a systems theoretical background to comprehend. We introduce this approach to identification by first developing model‐based processors that are the prime users of models evolving to the parametrically adaptive processors that jointly estimate the signals along with the embedded model parameters 1 , 5 . Next, we introduce the underlying theory evolving from systems theoretic realizations of state‐space models along with unique representations (canonical forms) for multivariable structures 16,17. Subspace identification is introduced for these deterministic systems. With the theory and algorithms for these systems in hand, the algorithms are extended to the stochastic case, culminating with a combined solution for both model sets, that is, deterministic and stochastic.
In terms of the system identification area, this text provides the link between model development and practical applications in model‐based signal processing filling this critical gap, since many identification texts dive into the details of the algorithms without completing the final signal processing application. Many use the model results to construct model‐based control systems, but do not focus on the processing aspects. Again the gap is filled in the signal processing community, by essentially introducing the notions and practicalities of subspace identification techniques applied to a variety of basic signal processing applications. For example, spectral estimation, communications, and primarily physics‐based problems, which this text will demonstrate in the final chapters. It is especially applicable for signal processors because they are currently faced with multichannel applications, which the state‐space formulations in this text handle quite easily, thereby opening the door for novel processing approaches. The current texts are excellent, but highly theoretical, attempting to provide signal processors with the underlying theory for the subspace approach 9 – 12 . Unfortunately, the authors are not able to achieve this, in my opinion, because the learning curve is too steep and more suitable for control system specialists with a strong systems theoretical background. It is difficult for signal processors to easily comprehend, but by incorporating the model‐based signal processing approach, which is becoming more and more known and utilized by the signal processing community as the connection will enable the readers to gently “bridge the gap” from statistical signal processing to subspace identification for subsequent processing especially in multichannel applications. This is especially true with readers familiar with our previous texts in model‐based processing 1 , 6 . It will also have an impact in the structural dynamics area due to our case studies and applications introducing structural/test engineers to the model‐based identifiers/processors 16 . They already apply many of these identification techniques to their problem sets.
The approach we take is to introduce the concept of subspace identification by first discussing the ideas of signal estimation, identification to model‐based signal processing (MBSP) leading to the concept of model‐based identification (MBID) 1 , 5 . Here the model set is defined, and a variety of techniques ranging from the black‐box approach to well‐defined structural models employing parameter estimation techniques are developed. After introducing these concepts in the first chapter, random signals and systems are briefly discussed leading to the concept of spectral estimation, which provides an underlying cornerstone of the original identification problem.
Next, state‐space models are introduced in detail evolving from continuous‐time, sampled‐data to discrete‐time systems leading to the stochastic innovations model linking the classical Wiener filter to the well‐known Kalman filter 2 . With this in hand, multivariable (multiple‐input/multiple‐output) systems are developed simply as time series, to sophisticated canonical forms, leading to the matrix fraction transfer function descriptions. Chapter 3 is concluded with approximate nonlinear Gauss–Markov representations in state‐space form.
Model‐based processors are highlighted in the next two chapters 4 and 5 ranging from developments of the linear representations leading to the optimal Kalman filter 2 . Next the suite of nonlinear processors is developed initiated by the linearized processor leading to the special cases of the extended and unscented Kalman filters and culminating with the novel particle filter evolving from the Bayesian approach 5 . These techniques are extended to the joint signal/parameter estimation problem to create the parametric adaptive processors. Throughout these chapters, examples and case studies are introduced to solidify these fundamental ideas.
Next, we introduce the foundations for the heart of the text – subspace identification first constrained to deterministic systems. Here we develop the fundamental realization problem that provides the basis of subspace identification using Hankel matrices. Many of the underlying systems theoretical results introduced by Kalman 18 in the 1960s are captured by properties of the Hankel matrix. The problem is extended to the deterministic identification problem by incorporating input/output sequences 15 . Perhaps one of the most important contributions to realization theory is the concept of a balanced realization enabling the evolution of robust algorithms. All of these concepts are carefully developed in this chapter. Canonical realizations, that is, the identification of models in unique canonical forms is an important concept in identification 16 – 18 . Here, much of this effort has been ignored over the years primarily because the concept of a unique representation can lead to large errors when identifying the model. However, it is possible to show that they can also be considered a viable approach, since they transform the Hankel array to the so‐called structural matrix enabling both the order and parameters to be identified simultaneously, leading to an invariant system description 17 ,19. Finally, we introduce the ideas of projection theory showing how orthogonal/oblique projections lead to popular deterministic identification techniques 9 – 11 . Chapter 6 is concluded with a detailed case study on the identification application for a mechanical/structural system.
Chapter 7 is the extension of the deterministic identification problem to the stochastic case. Here, in the realization context, covariance matrices replace impulse response matrices, while deterministic input/output sequences are replaced with noisy multichannel sequences – the real‐world problem. As in Chapter 6, we develop stochastic realization theory starting with the “indirect” realization approach 4 based on covariance matrices for infinite and finite sequences to develop the basis of stochastic realization theory. The main ideas evolve from the work of Akaike 20,21 and the development of predictor spaces leading to the fundamental results from the systems theoretic viewpoint. The optimal solution to this problem proceeds from classical spectral factorization techniques leading to the steady‐state Kalman filter and the fundamental innovations model that is an integral part of subspace realizations 1 20 –29. Next, subspace methods are reintroduced for random vector spaces and provided as solutions to the stochastic realization problem followed by the so‐called combined subspace technique extracting both deterministic and stochastic models simultaneously 9 – 13 . This chapter concludes with a case study discussing the design of a processor to detect modal anomalies in an unknown cylindrical structure.
The text concludes with a chapter describing sets of real‐world applications of these techniques. The applications range from failure detection, to the threat detection of fission sources, to the identification of chirp signals for radar/sonar application, to the parametrically adaptive processor design for localization and tracking in the ocean environment, and to the design of an MBP chirp‐based signals as well as a critical radiation system – the scintillator.
Appendices are included for critical review as well as problem sets and notes for the MATLAB software used in the signal processing/controls/identification areas at the end of each chapter.
1 Candy, J. (2006).
Model‐Based Signal Processing
. Hoboken, NJ: Wiley/IEEE Press.
2 Kalman, R. (1960). A new approach to linear filtering and prediction problems.
Trans. ASME J. Basic Eng.
82: 34–45.
3 Van Der Veen, A., Deprettere, E., and Swindlehurst, A. (1993). Subspace‐based methods for the identification of linear time‐invariant systems.
Proc. IEEE
81 (9): 1277–1308.
4 Viberg, M. (1995). Subspace based signal analysis using singular value decomposition.
Automatica
31 (12): 1835–1851.
5 Candy, J. (2016).
Bayesian Signal Processing: Classical, Modern and Particle Filtering Methods
, 2e. Hoboken, NJ: Wiley/IEEE Press.
6 Ljung, L. (1999).
System Identification: Theory for the User
, 2e. Englewood Cliffs, NJ: Prentice‐Hall.
7 Ljung, L. and Soderstrom, T. (1983).
Theory and Practice of Recursive Identification
. Cambridge: MIT Press.
8 Soderstrom, T. and Stoica, P. (1989).
System Identification
. New York: Academic Press.
9 van Overschee, P. and De Moor, B. (1996).
Subspace Identification for Linear Systems: Theory, Implementation, Applications
. Boston, MA: Kluwer Academic Publishers.
10 Katayama, T. (2005).
Subspace Methods for System Identification
. London: Springer.
11 Verhaegen, M. and Verdult, V. (2007).
Filtering and System Identification: A Least‐Squares Approach
. Cambridge: Cambridge University Press.
12 Juang, J. (1994).
Applied System Identification
. Upper Saddle River, NJ: Prentice‐Hall PTR.
13 Aoki, M. (1990).
State Space Modeling of Time Series
, 2e. London: Springer.
14 Norton, J. (1986).
An Introduction to Identification
. New York: Academic Press.
15 Ho, B. and Kalman, R. (1966). Effective reconstruction of linear state variable models from input/output data.
Regelungstechnik
14: 545–548.
16 Luenberger, D. (1967). Canonical forms for linear multivariable systems.
IEEE Trans. Autom. Control
AC‐12: 290–293.
17 Candy, J., Warren, M., and Bullock, T. (1977). Realization of an invariant system description from Markov sequences.
IEEE Trans. Autom. Control
AC‐23 (12): 93–96.
18 Chen, C. (1984).
Linear System Theory and Design
. New York: Holt, Rinehart & Winston.
19 Guidorzi, R. (1975). Canonical structures in the identification of multivariable systems.
Automatica
11: 361–374.
20 Akaike, H. (1974). Stochastic theory of minimal realization.
IEEE Trans. Autom. Control
19: 667–674.
21 Akaike, H. (1975). Markovian representation of stochastic processes by canonical variables.
SIAM J. Control
13 (1): 162–173.
22 Faurre, P. (1976). Stochastic realization algorithms. In:
System Identification: Advances and Case Studies
(ed. R. Mehra and D. Lainiotis), 1–23. New York: Academic Press.
23 Larimore, W. (1990). Canonical variate analysis in identification, filtering and adaptive control. In:
Proceedings of the 29th Conference on Decision and Control
, Hawaii, USA, 596–604.
24 Tse, E. and Wiennert, H. (1975). Structure determination and parameter identification for Multivariable stochastic linear systems.
IEEE Trans. Autom. Control
20: 603–613.
25 Glover, K. and Willems, J. (1974). Parameterizations of linear dynamical systems: canonical forms and identifiability.
IEEE Trans. Autom. Control
19: 640–646.
26 Denham, M. (1974). Canonical forms for identification of multivariable linear systems.
IEEE Trans. Autom. Control
19: 646–656.
27 Candy, J., Bullock, T., and Warren, M. (1979). Invariant system description of the stochastic realization.
Automatica
15: 493–495.
28 Candy, J., Warren, M., and Bullock, T. (1978). Partial realization of invariant system descriptions.
Int. J. Control
28 (1): 113–127.
29 Sullivan, E. (2015).
Model‐Based Processing for Underwater Acoustic Arrays
. New York: Springer.
James V. Candy
Danville, CA
The support and encouragement of my wife, Patricia, is the major motivational element needed to undertake this endeavor. My family, extended family, and friends having endured the many regrets, but still offer encouragement in spite of all of my excuses. Of course, the constant support of my great colleagues and friends, especially Drs. S. Lehman, I. Lopez, E. Sullivan, and Mr. B. Beauchamp, who carefully reviewed the manuscript and suggested many improvements, cannot go without a hearty acknowledgment.
ADC
analog‐to‐digital conversion
AIC
Akaike information criterion
AR
autoregressive (model)
ARMA
autoregressive moving average (model)
ARMAX
autoregressive moving average exogenous input (model)
ARX
autoregressive exogenous input (model)
AUC
area‐under‐curve (ROC curve)
BSP
Bayesian signal processing
BW
bandwidth
CD
central difference
CDF
cumulative distribution
CM
conditional mean
CRLB
Cramer–Rao lower bound
C‐Sq
Chi‐squared (distribution or test)
CT
continuous‐time
CTD
concentration–temperature–density (measurement)
CVA
canonical variate analysis
EKF
extended Kalman filter
EM
expectation–maximization
FPE
final prediction error
GLRT
generalized likelihood ratio test
G‐M
Gaussian mixture
GM
Gauss–Markov
G‐S
Gaussian sum
HD
Hellinger distance
HPR
high probability region
IEKF
iterated–extended Kalman filter
i.i.d.
independent‐identically distributed (samples)
KD
Kullback divergence
KL
Kullback–Leibler
KLD
Kullback–Leibler divergence
KSP
Kalman–Szego–Popov (equations)
LD
lower diagonal (matrix) decomposition
LE
Lyapunov equation
LKF
linear Kalman filter
LMS
least mean square
LS
least‐squares
LTI
linear, time‐invariant (system)
LZKF
linearized Kalman filter
MA
moving average (model)
MAICE
minimum Akaike information criterion
MAP
maximum a posteriori
MATLAB
mathematical software package
MBID
model‐based identification
MBP
model‐based processor
MBSP
model‐based signal processing
MC
Monte Carlo
MDL
minimum description length
MIMO
multiple‐input/multiple‐output (system)
MinE
minimum probability of error
ML
maximum likelihood
MOESP
multivariable output error state‐space algorithm
MMSE
minimum mean‐squared error
MSE
mean‐squared error
MV
minimum variance
N4SID
numerical algorithm for subspace state‐space system identification
NMSE
normalized mean‐squared error
N‐P
Neyman–Pearson (detector)
ODP
optimal decision (threshold) point
probability density function (continuous)
P‐E
probability‐of‐error (detector)
PEM
prediction error method
PF
particle filter
PI‐MOESP
past‐input multivariable output error state‐space algorithm
PMF
probability mass function (discrete)
PO‐MOESP
past‐output multivariable output error state‐space algorithm
PSD
power spectral density
RC
resistor capacitor (circuit)
REBEL
recursive Bayesian estimation library
RLC
resistor–inductor–capacitor (circuit)
RLS
recursive least‐squares
RMS
root mean‐squared
RMSE
root minimum mean‐squared error
ROC
receiver operating characteristic (curve)
RPE
recursive prediction error
RPEM
recursive prediction error method
SID
subspace identification
SIR
sequential importance sampling‐resampling
SIS
sequential importance sampling
SMC
sequential Markov chain
SNR
signal‐to‐noise ratio
SPRT
sequential probability ratio test
SPT
sigma‐point transformation
SSIS
sequential sampling importance sampling
SSP
state‐space processor
SSQE
sum‐squared error
SVD
singular‐value (matrix) decomposition
UD
upper diagonal matrix decomposition
UKF
unscented Kalman filter
UT
unscented transform
WSSR
weighted sum‐squared residual statistical test
W‐test
whiteness test
Z
Z‐transform
Z‐M
In this chapter, we introduce the idea of model‐based identification, starting with the basic notions of signal processing and estimation. Once defined, we introduce the concepts of model‐based signal processing, that lead to the development and application of subspace identification. Next, we show that the essential ingredient of the model‐based processor is the “model” that must be available either through the underlying science (first principles) or through the core of this text – model‐based identification.
The development of processors capable of extracting information from noisy sensor measurement data is essential in a wide variety of applications, whether it be locating a hostile target using radar or sonar systems or locating a tumor in breast tissue or even locating a seismic source in the case of an earthquake. The nondestructive evaluation (NDE) of a wing or hull of a ship provides a challenging medium even in the simplest of arrangements requiring sophisticated processing especially if the medium is heterogeneous. Designing a controller for a smart car or a drone or for that matter a delicate robotic surgical instrument also depends on providing enhanced signals for feedback and error corrections. Robots replacing humans in assembly lines or providing assistance in mundane tasks must sense their surroundings to function in a such a noisy environment. Most “hi‐tech” applications require the incorporation of “smart” processors capable of sensing their operational environment, enhancing noisy measurements and extracting critical information in order to perform a preassigned task such as detecting a hostile target and launching a weapon or detecting a tumor and extracting it. In order to design a processor with the required capability, it is necessary to utilize as much available a priori information as possible. The design may incorporate a variety of disciplines to achieve the desired results. For instance, the processor must be able to sense the operational environment, whether it be highly cluttered electromagnetic propagation at an airport or a noisy ocean acoustic environment in a busy harbor. Array radiation measurements in the case of an active radar system targeting signals of great interest can detect incoming threats, while passive listening provided by an acoustic array aids in the detection of submarines or similarly tumors in the human body for ultrasonics as well. The ability of the processor to operate effectively in such harsh environments requires more and more sophistication, rather than just simple filtering techniques. It is here that we address not only the need, but also the a priori requirements for a design. For instance, the detection and localization of a quiet diesel submarine cannot be achieved without some representation of the noisy, varying ocean incorporated in the processing scheme. How does such information get embedded? This is the question for not only the signal processor, but also the ocean acoustician and sensor designer to ponder. The solution boils down to the melding of this information enabling the development of a processor capable of performing well. So we see that except in an exceptional case, the knowledge of the underlying phenomenology that governs just how a signal propagates in an uncertain medium or environment coupled with that of how a sensor can make a reasonable measurement to provide the desired information and a processor capable of extracting that information defines a “team” consisting of a phenomenologist, sensor designer and signal processor that can enable a solution to the problem at hand. In this text, we discuss such an approach that incorporates all of these capabilities. We start with the basic processor and then progress to a scheme capable of incorporating the underlying phenomenology, measurement systems, and uncertainties into the processor. In order to do so, we start with defining signal processing and signal estimation, followed by the fundamental model‐based signal processor and then approaches to obtain the required model from experimental as well as application data sets.
Signal processing is based on one fundamental concept – extracting critical information from uncertain measurement data [1,2]. Processing problems can lead to some complex and intricate paradigms to perform this extraction especially from noisy, sometimes inadequate measurements. Whether the data are created using a seismic geophone sensor from a monitoring network or an array of hydrophone transducers located on the hull of an ocean‐going vessel, the basic processing problem remains the same – extract the useful information. Techniques in signal processing (e.g. filtering, Fourier transforms, time–frequency and wavelet transforms) are effective; however, as the underlying process generating the measurements becomes more complex, the resulting processor may require more and more information about the process phenomenology to extract the desired information. The challenge is to formulate a meaningful strategy that is aimed at performing the processing required, even in the face of these high uncertainties. This strategy can be as simple as a transformation of the measured data to another domain for analysis or as complex as embedding a full‐scale propagation model into the processor [3]. For example, think of trying to extract a set of resonances (damped sinusoids) from accelerometer time series. It is nearly impossible to calculate zero‐crossings from the time series, but it is a simple matter to transform the data to the spectral domain using a Fourier transform and then applying the property that sinusoids are impulse‐like in Fourier space facilitating their extraction through peak detection. Finding a sinusoidal source propagating in the ocean is another matter that is quite complex due to the attenuation and dispersion characteristics of this harsh, variable environment. Here, a complex propagation model must be developed and applied to “unravel” the highly distorted data to reveal the source – a simple Fourier transform will no longer work. The aims of both approaches are the same – to extract the desired information and reject the extraneous and, therefore, develop a processing scheme to achieve this goal. The underlying signal processing philosophy is a “bottoms‐up” perspective enabling the problem to dictate the solution, rather than vice versa.
More specifically, signal processing forms the basic nucleus of many applications. It is a specialty area that many researchers/practitioners apply in their daily technical regimen with great success such as the simplicity in Fourier analysis of resonance data or in the complexity of analyzing the time–frequency response of dolphin sounds. Applications abound with unique signal processing approaches offering solutions to the underlying problem. For instance, the localization of a target in the hostile underwater ocean acoustic environment not only challenges the phenomenologist but also taxes the core of signal processing basics, thereby requiring that more sophistication and a priori knowledge be incorporated into the processor. This particular application has led to many advances both in underwater signal processing and in the development of a wide variety of so‐called model‐based or physics‐based processors. A prime example of this technology is the advent of the model‐based, matched‐field processor [ 3 –5] that has led not only to a solution of the target localization problem, but also to many applications in other areas such as nondestructive evaluation and biomedical imaging. Therefore, the conclusion remains the same, signal processing is a necessary ingredient as a working tool that must be mastered by phenomenologists to extract the useful information from uncertain measurements. In fact, we define signal processing as a set of techniques “to extract the desired information and reject the extraneous from uncertain measurement data.”
Signal processing relies on any prior knowledge of the phenomenology generating the underlying measurements. Characterizing this phenomenology and propagation physics along with the accompanying measurement instrumentation and noise are the preliminaries that all phenomenologists must tackle to solve such a processing problem. In many cases, this is much easier said than done. The first step is to determine what the desired information is and typically this is not the task of the signal processor, but that of the phenomenologist performing the study. In our case, we assume that the investigation is to extract information stemming from signals emanating from a source, whether it be an autonomous unmanned vehicle (AUV) on the highway or passively operating in the deep ocean, or a vibrating structure responding to ground motion. Applications can be very complex especially in the case of ultrasound propagating through complex media such as tissue in biomedical applications or through heterogeneous materials of critical parts in nondestructive evaluation (NDE) investigations or photons emanating from a radiation source [6]. In any case, the processing usually involves manipulating the measured data to extract the desired information, such as location and tracking of the AUV, failure detection for the structure, or tumor/flaw detection, and localization in both biomedical and NDE [3] .
If a measured signal is free from extraneous variations and is repeatable from measurement to measurement, then it is defined as a deterministic signal (Chapter 2). However, if it varies extraneously and is no longer repeatable, then it is defined as random signal. This text is concerned with the development of processing techniques to extract pertinent information from random signals utilizing any a priori information available. We call these techniques signal estimation or signal enhancement, and we call a particular algorithm a signal estimator or just estimator. Symbolically, we use the “caret” (^) notation to annotate an estimate (e.g. s → ŝ). Sometimes, estimators are called filters (e.g. Wiener filter) because they perform the same function as a deterministic (signal) filter except for the fact that the signals are random; that is, they remove unwanted disturbances. Noisy measurements are processed by the estimator to produce “filtered” data. To solidify these concepts, consider the following examples.
A deterministic signal composed of two sinusoids: the information at 10 Hz and the disturbance at 20 Hz, with its corresponding Fourier spectrum shown in Figure 1.1a. From a priori information, it is known that the desired signal has no frequencies above 15 Hz; however, the raw spectrum reveals the disturbance at 20 Hz. Since the data are deterministic, a low‐pass filter with a cutoff frequency of 12.5 Hz is designed to extract the desired information (10 Hz signal) and reject the extraneous (20 Hz disturbance). The filtered data are shown in Figure 1.1 b, where we can see the filtered signal and the resulting spectrum.
Figure 1.1 Processing of a deterministic signal. (a) Raw data and spectrum with signal at Hz and disturbance at Hz. (b) Processed data extracting the Hz signal (desired) and rejecting the extraneous ( Hz disturbance).
Consider the output of the estimation filter designed to eliminate random noise from a transient signal; that is, the estimator is a function of the signal and noise
Consider the following example to illustrate this processor.
A random pulse‐like signal contaminated by noise is shown in Figure 1.2a. Here, we see the measured data along with its Fourier spectrum. We design a signal estimator to extract the desired signal and remove the noise. The processed data are shown in Figure 1.2 b, where we observe the results of the estimation filter and the corresponding enhanced spectrum. Here, we see how the filtered response has eliminated the random noise. We discuss the concepts of signal estimation using the modern parametric design methods in Chapter 4.
If the estimation filter or more commonly the estimator employs a model of phenomenology or process under investigation, then it is considered a model‐based processor. For example, suppose we use a so‐called Gauss–Markov model (see Chapter 3) in our estimator design, that is, if we use
then the resulting signal estimator,
is called a model‐based signal processor and in this case a Kalman filter. Model‐based signal processing is discussed in Chapter 4. We shall see that random signals can be characterized by stochastic processes (Chapter 2), transformed to equivalent deterministic representations (covariance and power spectrum) and processed (model‐based processors, Chapters 4 and 5) much the same as a deterministic signal.
Figure 1.2 Processing of a random signal and noise. (a) Raw data and spectrum with noise. (b) Processed data extracting the signal (estimate) and rejecting the extraneous (noise).
Estimation can be thought of as a procedure made up of three primary parts:
Criterion function
Models
Algorithm.
The criterion function can take many forms and can also be classified as deterministic or stochastic. Models represent a broad class of information, formalizing the a priori knowledge about the process generating the signal, measurement instrumentation, noise characterization, and underlying probabilistic structure. Finally, the algorithm or technique chosen to minimize (or maximize) the criterion can take many different forms depending on (i) the models, (ii) the criterion, and (iii) the choice of solution. For example, one may choose to solve the well‐known least‐squares problem recursively or with a numerical‐optimization algorithm. Another important aspect of most estimation algorithms is that they provide a “measure of quality” of the estimator. Usually, what this means is that the estimator also predicts vital statistical information about how well it is performing.
Intuitively, we can think of the estimation procedure as follows:
The specification of a criterion
The selection of models from a priori knowledge
The development and implementation of an algorithm.
Criterion functions are usually selected on the basis of information that is meaningful about the process or the ease with which an estimator can be developed. Criterion functions that are useful in estimation can be classified as deterministic and probabilistic. Some typical functions are as follows:
Deterministic:
Squared error
Absolute error
Integral absolute error
Integral squared error
Probabilistic:
Maximum likelihood
Maximum a posteriori (Bayesian)
Maximum entropy
Minimum (error) variance.
Models can also be deterministic as well as probabilistic; however, here we prefer to limit their basis to knowledge of the process phenomenology (physics) and the underlying probability density functions as well as the necessary statistics to describe the functions. Phenomenological models fall into the usual classes defined by the type of underlying mathematical equations and their structure, namely, linear or nonlinear, differential or difference, ordinary or partial, time invariant or varying. Usually, these models evolve to a stochastic model by the inclusion of uncertainty or noise processes.
Finally, the estimation algorithm can evolve from various influences. A preconceived notion of the structure of the estimator heavily influences the resulting algorithm. We may choose, based on computational considerations, to calculate an estimate recursively, rather than as a result of a batch process because we require an online, pseudo‐real‐time estimate. Also each algorithm must provide a measure of estimation quality, usually in terms of the expected estimation error. This measure provides a means for comparing estimators. Thus, the estimation procedure is a combination of these three major ingredients: criterion, models, and algorithm.
Another view of the underlying processing problem is to decompose it into a set of steps that capture the strategic essence of the processing scheme. Inherently, we believe that the more “a priori” knowledge about the measurement and its underlying phenomenology we can incorporate into the processor, the better we can expect the processor to perform – as long as the information that is included is correct! One strategy called the model‐based approach provides the essence of model‐based signal processing (MBSP) [3] . Some believe that all signal processing schemes can be cast into this generic framework. Simply, the model‐based approach is “incorporating mathematical models of both physical phenomenology and the measurement process (including noise) into the processor to extract the desired information.” This approach provides a mechanism to incorporate knowledge of the underlying physics or dynamics in the form of mathematical process models along with measurement system models and accompanying noise as well as model uncertainties directly into the resulting processor. In this way, the model‐based processor (MBP) enables the interpretation of results directly in terms of the problem physics. It is actually a modeler's tool enabling the incorporation of any a priori information about the problem to extract the desired information. This approach of selecting the appropriate model is depicted in the signal processing staircase of Figure 1.3. We note that as we progress up the “modeling” steps to increase the (SNR), the complexity of the model increases to achieve the desired results. This is our roadmap [3]
