Share Email Print

Spie Press Book • new

Optimal Signal Processing Under Uncertainty
Author(s): Edward R. Dougherty
Format Member Price Non-Member Price

Book Description

In the classical approach to optimal filtering, it is assumed that the stochastic model of the physical process is fully known. For instance, in Wiener filtering it is assumed that the power spectra are known with certainty. The implicit assumption is that the parameters of the model can be accurately estimated. When models are complex or parameter estimation is difficult (or expensive), this assumption is unwarranted. With uncertain models, the natural solution is to optimize over both the original objective and the model uncertainty, thereby arriving at optimal robust operators, the topic of this book. The book also addresses the correlated problem of optimal experimental design: determining the experiment to perform in order to maximally reduce the uncertainty impacting the operational objective. Model uncertainty impacts a wide spectrum of disciplines: engineering, physics, biology, medicine, and economics. This book aims to provide the reader with a solid theoretical background to the state-of-the art in treating a problem that is only going to grow as our desire to control and make decisions regarding complex systems grows, and to do so by considering a broad set of topics: filtering, control, structural intervention, compression, classification, and clustering.


Book Details

Date Published: 3 July 2018
Pages: 308
ISBN: 9781510619296
Volume: PM287

Table of Contents
SHOW Table of Contents | HIDE Table of Contents

Table of Contents

Preface
Acknowledgments

1 Random Functions
1.1 Moments
1.2 Calculus
1.3 Three Fundamental Processes
      1.3.1 Poisson process
      1.3.2 White noise
      1.3.3 Wiener process
1.4 Stationarity
1.5 Linear Systems

2 Canonical Expansions
2.1 Fourier Representation and Projections
2.2 Constructing Canonical Expansions
2.3 Orthonormal Coordinate Functions
      2.3.1 Canonical expansions and the Parseval identity
      2.3.2 Karhunen–Loève expansion
2.4 Derivation from a Covariance Expansion
2.5 Integral Canonical Expansions
2.6 Expansions of WS Stationary Processes
      2.6.1 Power spectral density and linear operators
      2.6.2 Integral representation of a WS stationary process

3 Optimal Filtering
3.1 Optimal Mean-Square-Error Filters
3.2 Optimal Finite-Observation Linear Filters
      3.2.1 Orthogonality principle
      3.2.2 Optimal filter for linearly independent observations
      3.2.3 Optimal filter for linearly dependent observations
3.3 Optimal Linear Filters for Random Vectors
3.4 Recursive Linear Filters
      3.4.1 Recursive generation of direct sums
      3.4.2 Static recursive optimal linear filtering
      3.4.3 Dynamic recursive optimal linear filtering
3.5 Optimal Infinite-Observation Linear Filters
      3.5.1 Wiener–Hopf equation
      3.5.2 Wiener filter
3.6 Optimal Filtering via Canonical Expansions
      3.6.1 Integral decomposition into white noise
      3.6.2 Extended Wiener–Hopf equation
      3.6.3 Solution via discrete canonical expansions
3.7 Discussion
      3.7.1 Granulometries
      3.7.2 Optimal granulometric bandpass filters
      3.7.3 Reconstructive granulometric bandpass filters
3.8 General Schema for Optimal Design

4 Optimal Robust Filtering
4.1 Intrinsically Bayesian Robust Filters
      4.1.1 Effective characteristics
      4.1.2 IBR linear filtering
      4.1.3 IBR granulometric bandpass filters
      4.1.4 The general schema under uncertainty
4.2 Optimal Bayesian Filters
4.3 Model-Constrained Bayesian Robust Filters
4.4 Robustness via Integral Canonical Expansions
      4.4.1 Robustness in the blurred-signal-plus-noise model
      4.4.2 Robustness relative to the power spectral density
      4.4.3 Global filters
4.5 Minimax Robust Filters
      4.5.1 Minimax robust granulometric bandpass filters
4.6 IBR Kalman Filtering
      4.6.1 IBR Kalman predictor
      4.6.2 IBR Kalman filter
      4.6.3 Robustness of IBR Kalman filtering
4.7 IBR Kalman–Bucy Filtering

5 Optimal Experimental Design
5.1 Mean Objective Cost of Uncertainty
      5.1.1 Utility
5.2 Experimental Design for IBR Linear Filtering
      5.2.1 Wiener filtering
      5.2.2 WS stationary blurring-plus-noise model
5.3 IBR Karhunen–Loève Compression
      5.3.1 IBR compression
      5.3.2 Experimental design for Karhunen–Loève compression
5.4 Markovian Regulatory Networks
      5.4.1 Markov chain peturbations
      5.4.2 Application to Boolean networks
      5.4.3 Dynamic intervention via external control
5.5 Complexity Reduction
      5.5.1 Optimal network reduction
      5.5.2 Preliminary node elimination
      5.5.3 Performance
5.6 Sequential Experimental Design
      5.6.1 Experimental design using dynamic programming
5.7 Design with Inexact Measurements
      5.7.1 Design with measurement error
      5.7.2 Design via surrogate measurements
5.8 General MOCU-based Experimental Design

6 Optimal Classification
6.1 Bayes Classifier
      6.1.1 Discriminants
6.2 Optimal Bayesian Classifier
      6.2.1 Bayesian MMSE error estimator
      6.2.2 Effective class-conditional densities
6.3 Classification Rules
      6.3.1Discrete classification
      6.3.2 Quadratic discriminant analysis
6.4 OBC in the Discrete and Gaussian Models
      6.4.1 Discrete model
      6.4.2 Gaussian model
6.5 Consistency
6.6 Optimal Sampling via Experimental Design
6.7 Prior Construction
      6.7.1 General framework
      6.7.2 From prior knowledge to constraints
      6.7.3 Dirichlet prior distribution
6.8 Epistemology
      6.8.1 RMS bounds
      6.8.2 Validity in the context of a prior distribution

7 Optimal Clustering
7.1 Clustering
7.2 Bayes Clusterer
      7.2.1 Clustering error
      7.2.2 Bayes clustering error
7.3 Separable Point Processes
7.4 Intrinsically Bayesian Robust Clusterer
      7.4.1 Effective random labeled point processes

References

Index

Preface

Whereas modern science concerns the mathematical modeling of phenomena, essentially a passive activity, modern engineering involves determining operations to actively alter phenomena to effect desired changes of behavior. It begins with a scientific (mathematical) model and applies mathematical methods to derive a suitable intervention for the given objective. Since one would prefer the best possible intervention, engineering inevitably becomes optimization and, since all but very simple systems must account for randomness, modern engineering might be defined as the study of optimal operators on random processes. The seminal work in the birth of modern engineering is the Wiener–Kolmogorov theory of optimal linear filtering on stochastic processes developed in the 1930s. As Newton’s laws constitute the gateway into modern science, the Wiener–Kolmogorov theory is the gateway into modern engineering.

The design of optimal operators takes different forms depending on the random process constituting the scientific model and the operator class of interest. The operators might be linear filters, morphological filters, controllers, classifiers, or cluster operators, each having numerous domains of application. The underlying random process might be a random signal/image for filtering, a Markov process for control, a feature-label distribution for classification, or a random point set for clustering. In all cases, operator class and random process must be united in a criterion (cost function) that characterizes the operational objective and, relative to the criterion, an optimal operator found. For the classical Wiener filter, the model is a pair of jointly distributed wide-sense stationary random signals, the objective is to estimate a desired signal from an observed signal via a linear filter, and the cost function to be minimized is the mean-square error between the filtered observation and the desired signal.

Besides the mathematical and computational issues that arise with classical operator optimization, especially with nonlinear operators, nonstationary processes, and high dimensions, a more profound issue is uncertainty in the scientific model. For instance, a long recognized problem with linear filtering is incomplete knowledge regarding the covariance functions (or power spectra in the case of Wiener filtering). Not only must optimization be relative to the original cost function, be it mean-square error in linear filtering or classification/clustering error in pattern recognition, optimization must also take into account uncertainty in the underlying random process. Optimization is no longer relative to a single random process but instead relative to an uncertainty class of random processes. This means the postulation of a new cost function integrating the original cost function with the model uncertainty. If there is a prior distribution (or posterior distribution if data are employed) governing likelihood in the uncertainty class, then one can choose an operator from some class of operators that minimizes the expected cost over the uncertainty class. In the absence of a prior distribution, one might take a minimax approach and choose an operator that minimizes the maximum cost over the uncertainty class.

A prior (or posterior) distribution places the problem in a Bayesian framework. A critical point, and one that will be emphasized in the text, is that the prior distribution is not on the parameters of the operator model, but on the unknown parameters of the scientific model. This is natural. If the model were known with certainty, then one would optimize with respect to the known model; if the model is uncertain, then the optimization is naturally extended to include model uncertainty and the prior distribution should be on that uncertainty. For instance, in the case of linear filtering the covariance function might be uncertain, meaning that some of its parameters are unknown, in which case the prior distribution characterizes uncertainty relative to the unknown parameters.

A basic principle embodied in the book is to express an optimal operator under the joint probability space formed from the joint internal and external uncertainty in the same form as an optimal operator for a known model by replacing the mathematical structures characterizing the standard optimal operator with corresponding structures, called effective structures, that incorporate model uncertainty. For instance, in Wiener filtering the power spectra are replaced by effective power spectra, in Kalman filtering the Kalman gain matrix is replaced by the effective Kalman gain matrix, and in classification the class-conditional distributions are replaced by effective classconditional distributions.

The first three chapters of the book review those aspects of random processes that are necessary for developing optimal operators under uncertainty. Chapter 1 covers random functions, including the moments and the calculus of random functions. Chapter 2 treats canonical expansions for random functions, a topic often left uncovered in basic courses on stochastic processes. It treats discrete expansions within the context of Hilbert space theory for random functions, in particular, the equivalence of canonical expansions of the random function and the covariance function entailed by Parseval's identity. It then goes on to treat integral canonical expansions in the framework of generalized functions. Chapter 3 covers the basics of classical optimal filtering: optimal finite-observation linear filtering, optimal infinite-observation linear filtering via the Wiener–Hopf integral equation, Wiener filtering for wide-stationary processes, recursive (Kalman) filtering via direct-sum decomposition of the evolving observation space, and optimal morphological filtering via granulometric bandpass filters.

For the most part, although not entirely, the first three chapters are a compression of Chapters 2 through 4 of my book Random Processes for Image and Signal Processing, aimed directly at providing a tight background for optimal signal processing under uncertainty, the goal being to make a one-semester course for Ph.D. students. Indeed, the book has been developed from precisely such a course, attended by Ph.D. students, post-doctoral students, and faculty.

Chapter 4 covers optimal robust filtering. The first section lays out the basic definitions for intrinsically Bayesian robust (IBR) filtering, the fundamental principle being filter optimization with respect to both internal model stochasticity and external model uncertainty, the latter characterized by a prior distribution over an uncertainty class of random-process models. The first section introduces the concepts of effective process and effective characteristic, whereby the structure of the classical solutions is retained with characteristics such as the power spectra and the Wiener–Hopf equation generalized to effective power spectra and the effective Wiener–Hopf equation, which are relative to the uncertainty class. Section 4.2 covers optimal Bayesian filters, which are analogous to IBR filters except that new observations are employed to update the prior distribution to a posterior distribution. Section 4.3 treats model-constrained Bayesian robust (MCBR) filters, for which optimization is restricted to filters that are optimal for some model in the uncertainty class. In Section 4.4 the term "robustness" is defined quantitatively via the loss of performance and is characterized for linear filters in the context of integral canonical expansions, where random process representation is now parameterized via the uncertainty. Section 4.5 reviews classical minimax filtering and applies it to minimax morphological filtering. Sections 4.6 and 4.7 extend the classical Kalman (discrete time) and Kalman– Bucy (continuous time) recursive predictors and filters to the IBR framework, where classical concepts such as the Kalman gain matrix get extended to their effective counterparts (effective Kalman gain matrix).

When there is model uncertainty, a salient issue is the design of experiments to reduce uncertainty; in particular, which unknown parameter should be determined to optimally reduce uncertainty. To this end, Section 5.1 introduces the mean objective cost of uncertainty (MOCU), which is the expected cost increase relative to the objective resulting from the uncertainty, expectation being taken with respect to the prior (posterior) distribution. Whereas entropy is a global measure of uncertainty not related to any particular operational objective, MOCU is based directly on the engineering objective. Section 5.2 analyzes optimal MOCU-based experimental design for IBR linear filtering. Section 5.3 revisits Karhunen–Loève optimal compression when there is model uncertainty, and therefore uncertainty as to the Karhunen–Loève expansion. The IBR compression is found and optimal experimental design is analyzed relative to unknown elements of the covariance matrix. Section 5.4 discusses optimal intervention in regulatory systems modeled by Markov chains when the transition probability matrix is uncertain and derives the experiment that optimally reduces model uncertainty relative to the objective of minimizing undesirable steady-state mass. The solution is computationally troublesome, and the next section discusses complexity reduction. Section 5.6 examines sequential experimental design, both greedy and dynamic-programming approaches, and compares MOCU-based and entropy-based sequential design. To this point, the chapter assumes that parameters can be determined exactly. Section 5.7 addresses the issue of inexact measurements owing to either experimental error or the use of surrogate measurements in place of the actually desired measurements, which are practically unattainable. The chapter closes with a section on a generalized notion of MOCU-based experimental design, a particular case being the knowledge gradient.

The optimal Bayesian filter paradigm was first introduced in classification with the design of optimal Bayesian classifiers (OBCs). Classically (Section 6.1), if the feature-label distribution is known, then an optimal classifier, one that minimizes classification error, is found as the Bayes classifier. As discussed in Section 6.2, when there are unknown parameters in the featurelabel distribution, then there is an uncertainty class of feature-label distributions, and an optimal Bayesian classifier minimizes the expected error across the uncertainty class relative to the posterior distribution derived from the prior and the sample data. In order to compare optimal Bayesian classification with classical methods, Section 6.3 reviews the methodology of classification rules based solely on data. Section 6.4 derives the OBC in the discrete and Gaussian models. Section 6.5 examines consistency, that is, convergence of the OBC as the sample size goes to infinity. Rather than sample randomly or separately (randomly given the class sizes), sampling can be done in a nonrandom fashion by iteratively deciding which class to sample from prior to the selection of each point or by deciding which feature vector to observe. Optimal sequential sampling in these paradigms is discussed in Section 6.7 using MOCU-based experimental design. Section 6.8 provides a general framework for constructing prior distributions via optimization of an objective function subject to knowledge-based constraints. Epistemological issues regarding classification are briefly discussed in Section 6.9.

Clustering shares some commonality with classification in that both involve operations on points, classification on single points and clustering on point sets, and both have performances measured via a natural definition of error, classification error for the former and cluster (partition) error for the latter. But they also differ fundamentally in that the underlying random process for classification is a feature-label distribution, and for clustering it is a random point set. Section 7.1 describes some classical clustering algorithms. Section 7.2 discusses the probabilistic foundation of clustering and optimal clustering (Bayes cluster operator) when the underlying random point set is known. Section 7.3 describes a special class of random point sets that can be used for modeling in practical situations. Finally, Section 7.4 discusses IBR clustering, which involves optimization over both the random set and its uncertainty when the random point set is unknown and belongs to an uncertainty class of random point sets. Whereas with linear and morphological filtering the structure of the IBR or optimal Bayesian filter essentially results from replacing the original characteristics with effective characteristics, optimal clustering cannot be so conveniently represented. Thus, the entire random point process must be replaced by an effective random point process.

Let me close this preface by noting that this book views science and engineering teleologically: a system within Nature is modeled for a purpose; an operator is designed for a purpose; and an optimal operator is obtained relative to a cost function quantifying the achievement of that purpose. Right at the outset, with model formation, purpose plays a critical role. As stated by Erwin Schrödinger (Schrödinger, 1957), "A selection has been made on which the present structure of science is built. That selection must have been influenced by circumstances that are other than purely scientific…. The origin of science [is] without any doubt the very anthropomorphic necessity of man's struggle for life." Norbert Wiener, whose thinking is the genesis behind this book, states (Rosenblueth and Wiener, 1945) this fundamental insight from the engineering perspective: "The intention and the result of a scientific inquiry is to obtain an understanding and a control of some part of the universe.” It is not serendipity that leads us inexorably from optimal operator representation to optimal experimental design. This move, too, is teleological, as Wiener makes perfectly clear (Rosenblueth and Wiener, 1945): "An experiment is a question. A precise answer is seldom obtained if the question is not precise; indeed, foolish answers—i.e., inconsistent, discrepant or irrelevant experimental results—are usually indicative of a foolish question." Only in the context of optimization can one know the most relevant questions to ask Nature.

Edward R. Dougherty
College Station, Texas
June 2018


© SPIE. Terms of Use
Back to Top