Table of Contents:

      Developing Models for Kalman Filters

 

Chapter Title

Section 1

Chapter 1 Why Model?

What is a model? What is a linear model? How is this related to Kalman Filters? For what purpose?

Chapter 2 Linearity

How do strict and conventional notions of linearity differ? How is this related to Kalman Filter models?

Chapter 3 State Models and Kalman Filters

Why are the linear models discussed so far all unsuitable for purposes of Kalman Filters — what is missing?

Chapter 4 State Transition Dynamic Models

Detailing the special form your model must have to be suited for classic Kalman Filtering — and why.

Chapter 5 Feedback in State Models

Some adjustments you will need in your model if you choose to — or must — employ feedback stabiliization as your system operates.

Chapter 6 Linear Least Squares Review

An essential tool: how to formulate and solve a Linear Least Squares problem to evaluate "best fit" model parameters.

Chapter 7 Practical Linear Least Squares Calculations

Introducing the Octave package, and showing how simple it is to perform Linear Least Squares calculations in practice.

Chapter 8 Least Squares Dynamic Fit

A lot of ideas come together in an attempt to build a dynamic mode using Least Squares methods.

Chapter 9 Adapting ARMA Models

Another approach: fabricate a suitable state transition model from an ARMA model obtained by regression analysis methods.

Chapter 10 Verifying the Model

How to numerically simulate and evaluate your proposed model using actual system I/O data.

Chapter 11 Updating and Sequential Least Squares

It is not necessary to collect huge data sets and grind them down all at once for Least Squares fitting — you can consolidate as you go.

Chapter 12 Scaling, Weighting, and Least Squares

An important feature... a serious hazard... how you present data to a least squares problem affects the solution you will get.

Chapter 13 Fading Memory in Least Squares Problems

Critical adjustments are required to allow Least Squares updates to systematically prefer new data over old.

Chapter 14 Introducing Recursive Least Squares (RLS)

Begin the search for efficient Least Squares solutions when frequent updates of parameter values are needed.

Chapter 15 Efficient RLS Computation

Inverse updates provide the missing piece for the RLS method.

Chapter 16 Adaptive "Recursive Least Squares" Applications

Avoiding disaster when using RLS methods for system models that change "adaptively" over time.

Chapter 17 Total Least Squares Methods

An alternative to Linear Least Squares when all system inputs and outputs are noisy.

Chapter 18 LMS: Let the Model Self-tune

Applying the LMS method to let a state transition model incrementally improve itself, based on test data — and patience.

Chapter 19 Restructuring State Models, Part 1

State transition models are not unique! Introducing transformations that can produce equivalent models.

Chapter 20 Restructuring State Models, Part 2

Introducing Householder transformations, for sparse and efficient state model systems.

Chapter 21 Restructuring State Models, Part 3

Introducing Eigenvector transformations, for producing state model systems with minimal state interaction.

Chapter 22 Model Order

Discussing the importance of representing the correct number of internal states in the model.

Chapter 23 Randomized Test Signals

How to effectively collect system test data for determining system order.

Chapter 24 Autocorrelation in Test Signals

Distinguishing artificial side effects of testing in correlation data.

Chapter 25 Measuring System Correlation

How to perform a correlation analysis on system I/O data — it's easy.

Chapter 26 Recognizing State Effects on Correlation

Patterns in correlation data that indicate the presence of internal states.

Chapter 27 Test Case: Identifying States in Correlation

A numerical example of counting internal states using correlation methods.

Chapter 28 Transition Matrix for Response Modes

Exploratory validation of correlation analysis results by constructing a model.

Chapter 29 Finishing Correlation-Based Model

Least Squares methods complete a correlation-based model to see if results are credible.

Chapter 30 LMS: Experiments in Tuning New States

LMS methods are used to splice an additional behavior onto an existing model.

Chapter 31 Reducing Overspecified Models

Reducing model order: removing redundant and undesirable elements from an existing model.

Chapter 32 Compacting the Reduced Model

Numerical cleanup after model order reduction, to obtain a compacted equivalent model.

Chapter 33 Adjusting Time Steps for Discrete Models

Transforming a state transition model to operate at a different time interval than the original model.

Chapter 34 State Observers: Theory

Introducing the other approach: trying to adjust the model's internal state variables rather than the model itself.

Chapter 35 State Observers: Design

Exploring how to set the observer parameters — its gains — to tune observer performance.

Chapter 36 State Observerse with a Weak Model

Benefits and hazards of observers: making good models work better, hiding the deficiency of a bad model.

Chapter 37 Reformulating the State Observer

A mathematical reformulation that combines the functions of state transition prediction and observer.

Chapter 38 Minimalist Observer: the Alpha Beta Filter

Observers with a model so weak that it barely qualifies as a model — yet, sometimes completely sufficient.

Chapter 39 Quantifying Variation

How variance is used in Kalman Filters to represent the properties of random noise.

Chapter 40 Initial Variance

How an interpretation of variance is employed to represent highly uncertain initial system conditions.

Chapter 41 Variance Propagation

How initial uncertainty and new random noise sources interact to affect progression of state uncertainty over time.

Chapter 42 Generating Correlated Random Vectors

How to produce pseudo-random noise with specified correlation properties, so that Kalman Filters can be simulated.

Chapter 43 Simulating the Noisy Observer

Experiments testing observer response to correlated noise.

Chapter 44 Observer Optimization

How to determine the "Kalman Gains" that achieve Wiener optimal tracking of the system state by the observer.

Chapter 45 The Steady State Kalman Filter

For fixed transition and noise models, how the complicated run-time variance updates can be eliminated.

Chapter 46 Consistent Covariances

An accurate noise model might not be possible — but it can at least be consistent with the actual system.

Chapter 47 The Dreaded Kalman Divergence

What happens when bad models produce seriously bad results, and why this doesn't need to happen to you.

Chapter 48 Considerations for Data Smoothing

How Kalman Filters can be tweaked to produce optimal estimates at past and future times, not just the next step.

 


 

Contents of the "Developing Models for Kalman Filters" section of the ridgerat-tech.us website, including this page, are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License unless otherwise specifically noted. For complete information about the terms of this license, see http://creativecommons.org/licenses/by-sa/4.0/. The license allows usage and derivative works for individual or commercial purposes under certain restrictions. For permissions beyond the scope of this license, see the contact information page for this site.