820 likes | 1.03k Views
SLAM Summer School 2004. An Introduction to SLAM – Using an EKF Paul Newman Oxford University Robotics Research Group. A note to students.
E N D
SLAM Summer School 2004 An Introduction to SLAM – Using an EKF Paul Newman Oxford University Robotics Research Group
A note to students The lecture I give will not include all these slides. Some of then and some of the notes I have supplied are more detailed than required and would take too long to deliver. I have included them for completeness and background – for example the derivation of the Kalman filter from Bayes Rule. I have included in the package a working matlab implementation of EKF based SLAM. You should be able to see all the properties of SLAM at work and be able to modify at your leisure. (without having to worry about the awkwardness of a real system to start with). I cannot cover all I would like to in the time available – where applicable, to fill gaps, I forward reference other talks that will be given during the week. I hope the talk, the slides and the notes will whet you appetite regarding what I reckon is great area of research. Above all, please please ask me to explain stuff that is unclear – this school is about you learning, not us lecturing. regards Paul Newman
Overview • Kalman Filter was the first tool employed in SLAM – Smith Self and Cheeseman. • Linear KFs implement Bayes rule. No hokie-ness • We can analyse KF properties easily and learn interesting things about Bayesian SLAM • The vanilla, monolithic, KF-SLAM formulation is a fine tool for small local areas • But we can do better for large areas – as other speakers will mention
Estimation is ….. Estimation Engine Data Estimate Prior Beliefs
Minimum Mean Squared Error Estimation Choose x so argument is minimised Expectation operator (“average”)
Evaluating…. From probability theory Very Important Thing
Recursive Bayesian Estimation Key idea: “one mans posterior is another’s prior” ;-) Sequence of data (measurements) We want conditional mean (mmse) of x given Zk Can we iteratively calculate this – ie every time a new measurement comes in, update our estimate?
Yes… At time k Explains data at time k as function of x at time k At time k-1
And if these distributions are Gaussian turning the handle (see supporting material) leads to the Kalman filter……
Kalman Filtering • Ubiquitous estimation tool • Simple to implement • Closely related to Bayes estimation and MMSE • Immensely Popular in robotics • Real time • Recursive (can add data sequentially) • It maintains the sufficient statistics of a Multidimensional • Gaussian PDF It is not that complicated! (trust me)
Overall Goal To come up with a recursive algorithm that produces an estimate of state by processing data from a set of explainable measurements and incorporating some kind of plant model Measurement model Sensor H1 KF Estimate Sensor H2 Sensor Hn Plant Model Prediction/plant model True underlying state x
Covariance is….. Multi-dimensional analogy of variance mean P is a symmetric matrix that describes a 1-standard deviation contour ( ellipsoid in 3D+ ) of the pdf
The i|j notation true estimated Data up to t=j This is useful for derivations but we can never use it in a calc asx is unknown truth!
The Basics We’ll use these equations as a starting point – I have supplied a full derivation in the support presentation and notes – think of a KF as an off-the-shelf estimation tool
Crucial Characteristics • Asynchronisity • Prediction Covariance Inflation • Update Covariance Deflation • Observability • Correlations
Nonlinear Kalman Filtering • Same trick as in Non-linear Least Squares: • Linearise around a current estimate using jacobian • Problem becomes linear again Complete derivation is in the notes but…
Vehicle Models - Prediction control Truth model
Background T-Composition Compounding transformations
Deduce an Incremental Move These can be in massive error But the common error is subtracted out here
Use this “move” as a control Substitution into Prediction equation (using J1 and J2 as Jacobians): Diagonal covariance matrix (3x3) of error in uo
Feature Based Mapping and Navigation Look at the code!!
Landmarks / Features Things that standout to a sensor: Corners, windows, walls, bright patches, texture… Map Point Feature called “i”
Observations / Measurements • Relative • On Vehicle sensing environment: • Radar • Cameras • Odometry (really) • Sonar Laser • Absolute • Relies on infrastructure: • GPS • Compass How smart can we be with relative only measurements?
And once again… It is all about probability
From Bayes Rule….. Input is measurements conditioned on map and vehicle Data: We want to use Bayes rule to “invert this” and get maps and vehicles given measurements.
Problem 1 - Localisation Remove line p(.) = 1 from notes. Mistake
We can use a KF for this! Plant Model Remember: u is control, J’s are a fancy way of writing jacobians (composition operator). Q is strength of noise in plant model.
Implementation No features seen here
Problem II Mapping Map With known vehicle The state vector is the map
But how is map built? Key Point: State Vector GROWS! New, bigger map Obs of new feature Old map “State augmentation”
How is P augmented? Simple! Use the transformation of covariance rule.. G is the feature initialisation function
Leading to : Angle from Veh to feature Vehicle orientation
So what are models h and f? h is a function of the feature being observed: f is simply the identity transformation :
Turn the handle on the EKF: All hail the Oracle ! How do we know whatfeature we are observing?
Problem III SLAM “Build a map and use it at the same time” “This a cornerstone of autonomy”