1 / 28

ASEN 5070: Statistical Orbit Determination I Fall 2013 Professor Brandon A. Jones

ASEN 5070: Statistical Orbit Determination I Fall 2013 Professor Brandon A. Jones Professor George H. Born Lecture 19: Minimum Variance Estimation and Introduction to Sequential Processing. Announcements. Exam 1 Plan to return and review on Monday, Oct. 21 Homework 6 Posted.

taniel
Download Presentation

ASEN 5070: Statistical Orbit Determination I Fall 2013 Professor Brandon A. Jones

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. ASEN 5070: Statistical Orbit Determination I Fall 2013 Professor Brandon A. Jones Professor George H. Born Lecture 19: Minimum Variance Estimation and Introduction to Sequential Processing

  2. Announcements • Exam 1 • Plan to return and review on Monday, Oct. 21 • Homework 6 Posted

  3. Minimum Variance

  4. Problem Statement • With the least squares solution, we minimized the square of the residuals • Instead, what if we want the estimate that gives us the highest confidence in the solution: • What is the linear, unbiased, minimum variance estimate of the state x?

  5. Problem Statement • What is the linear, unbiased, minimum variance estimate of the state x ? • This encompasses three elements • Linear • Unbiased, and • Minimum Variance • We consider each of these for formulate a solution

  6. Linear Estimator • To be linear, the estimated state is a linear combination of the observations: • What is the matrix M? • This mysterious M matrix gives us the solution to the minimum variance estimator

  7. Unbiased Estimator • To be unbiased, then Solution Constraint!

  8. Minimum Variance Estimator • Must satisfy previous requirements:

  9. What does it mean to have a minimum P ? • Put into the context of scalars:

  10. Covariance Cost Function

  11. Statement of Optimization Problem • We seek to minimize: • Subject to the equality constraint: • Using the method of Lagrange Multipliers, we seek to minimize: Term added to keep Q symmetric

  12. Solution Derivation • Using calculus of variations, we need the first variation to vanish to achieve a minimum:

  13. Solution Derivation • In order for the above to be satisfied: • We will focus on the first

  14. Solution Derivation • We now have two constraints, which will give us a solution:

  15. What about P non-negative definite? • Showed that P satisfies the constraints, but do we have a “minimum” • Must show that, for any other solution, • See book, p 186-187 for proof

  16. Minimum Variance Estimator • Turns out, we get the weighted, linear least squares! • Hence, the linear least squares gives us the minimum variance solution • Of course, this is predicated on all of our statistical/linearization assumptions

  17. Minimum Variance w/ a priori • Falls from similar derivations previously discussed:

  18. Minimum Variance and Sequential Processing

  19. Batch vs. Sequential Processing • Batch – process all observations in a single run of the filter • Sequential – process each observations individually (usually as they become available over time) X*

  20. Mapping of Filter State and Uncertainty • Recall how to map the state deviation and covariance matrix (previous lecture) • Can we leverage this information to sequentially process measurements in the minimum variance / least squares algorithm?

  21. Minimum Variance as a Sequential Processor • Given from a previous filter run: • We have new a observation and mapping matrix: • We can update the solution via:

  22. Sequential Estimator Updates • Two principle phases in any sequential estimator • Time Update • Map previous state deviation and covariance matrix to the current time of interest • Measurement Update • Update the state deviation and covariance matrix given the new observations at the time of interest • Jargon can change with communities • Forecast and analysis • Prediction and fusion • others…

  23. Sequential Estimation in Context of Bayesian Inference

  24. Notes on the Sequential Minimum Variance/Least Squares • No assumptions on the number of observations at tk. • Wait, but what if we have fewer observations than unknowns at tk? • Do we have an underdetermined system?

  25. Sequential Minimum Variance Measurement Update • The a priori may be based on independent analysis or a previous estimation • Independent analysis could be a product of: • Expected launch vehicle performance • Previous analysis of system (a priori gravity field) • Initial orbit determination solution

  26. Sequential Minimum Variance Measurement Update • We still have to invert a n × n matrix • Can be computationally expensive for large n • Gravity field estimation: ~n2+2n-3 coefficients! • May become sensitive to numeric issues

  27. Sequential Minimum Variance Measurement Update • Is there a better sequential processing algorithm? • YES! – This equations above may be manipulated to yield the Kalman filter

More Related