1 / 56

Vegetation Science Lecture 4 Non-Linear Inversion

Vegetation Science Lecture 4 Non-Linear Inversion. Lewis, Disney & Saich UCL. Introduction. Previously considered forward models Model reflectance/backscatter as fn of biophysical parameters Now consider model inversion

Download Presentation

Vegetation Science Lecture 4 Non-Linear Inversion

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Vegetation ScienceLecture 4Non-Linear Inversion Lewis, Disney & Saich UCL

  2. Introduction • Previously considered forward models • Model reflectance/backscatter as fn of biophysical parameters • Now consider model inversion • Infer biophysical parameters from measurements of reflectance/backscatter

  3. Linear Model Inversion • Dealt with last lecture • Define RMSE • Minimise wrt model parameters • Solve for minimum • MSE is quadratic function • Single (unconstrained) minimum

  4. P1 RMSE P0

  5. P1 RMSE P0

  6. Issues • Parameter transformation and bounding • Weighting of the error function • Using additional information • Scaling

  7. Parameter transformation and bounding • Issue of variable sensitivity • E.g. saturation of LAI effects • Reduce by transformation • Approximately linearise parameters • Need to consider ‘average’ effects

  8. Weighting of the error function • Different wavelengths/angles have different sensitivity to parameters • Previously, weighted all equally • Equivalent to assuming ‘noise’ equal for all observations

  9. Weighting of the error function • Can ‘target’ sensitivity • E.g. to chlorophyll concentration • Use derivative weighting (Privette 1994)

  10. Using additional information • Typically, for Veg Sci, use canopy growth model • See Moulin et al. (1998) • Provides expectation of (e.g.) LAI • Need: • planting date • Daily mean temperature • Varietal information (?) • Use in various ways • Reduce parameter search space • Expectations of coupling between parameters

  11. Scaling • Many parameters scale approximately linearly • E.g. cover, albedo, fAPAR • Many do not • E.g. LAI • Need to (at least) understand impact of scaling

  12. LAI 1 LAI 4 LAI 0 Crop Mosaic

  13. LAI 1 LAI 4 LAI 0 Crop Mosaic • 20% of LAI 0, 40% LAI 4, 40% LAI 1. • ‘real’ total value of LAI: • 0.2x0+0.4x4+0.4x1=2.0. visible: NIR

  14. canopy reflectance over the pixel is 0.15 and 0.60 for the NIR. • If assume the model above, this equates to an LAI of 1.4. • ‘real’ answer LAI 2.0

  15. Options for Numerical Inversion • Iterative numerical techniques • Quasi-Newton • Powell • Knowledge-based systems (KBS) • Artificial Neural Networks (ANNs) • Genetic Algorithms (GAs) • Look-up Tables (LUTs)

  16. Local and Global Minima Need starting point

  17. How to go ‘downhill’? Bracketing of a minimum

  18. Slow, but sure How far to go ‘downhill’? w=(b-a)/(c-a) z=(x-b)/(c-a) Choose: z+w=1-w For symmetry Choose: w=z/(1-w) To keep proportions the same w 1-w z=w-w2=1-2w 0= w2 -3w+1 Golden Mean Fraction =w=0.38197

  19. Parabolic Interpolation Inverse parabolic interpolation More rapid

  20. Brent’s method • Require ‘fast’ but robust inversion • Golden mean search • Slow but sure • Use in unfavourable areas • Use Parabolic method • when get close to minimum

  21. Multi-dimensional minimisation • Use 1D methods multiple times • In which directions? • Some methods for N-D problems • Simplex (amoeba)

  22. Downhill Simplex • Simplex: • Simplest N-D • N+1 vertices • Simplex operations: • a reflection away from the high point • a reflection and expansion away from the high point • a contraction along one dimension from the high point • a contraction along all dimensions towards the low point. • Find way to minimum

  23. Simplex

  24. Direction Set (Powell's) Method • Multiple 1-D minimsations • Inefficient along axes

  25. Powell

  26. Direction Set (Powell's) Method • Use conjugate directions • Update primary & secondary directions • Issues • Axis covariance

  27. Powell

  28. Simulated Annealing • Previous methods: • Define start point • Minimise in some direction(s) • Test & proceed • Issue: • Can get trapped in local minima • Solution (?) • Need to restart from different point

  29. Simulated Annealing

  30. Simulated Annealing • Annealing • Thermodynamic phenomenon • ‘slow cooling’ of metals or crystalisation of liquids • Atoms ‘line up’ & form ‘pure cystal’ / Stronger (metals) • Slow cooling allows time for atoms to redistribute as they lose energy (cool) • Low energy state • Quenching • ‘fast cooling’ • Polycrystaline state

  31. Simulated Annealing • Simulate ‘slow cooling’ • Based on Boltzmann probability distribution: • k – constant relating energy to temperature • System in thermal equilibrium at temperature T has distribution of energy states E • All (E) states possible, but some more likely than others • Even at low T, small probability that system may be in higher energy state

  32. Simulated Annealing • Use analogy of energy to RMSE • As decrease ‘temperature’, move to generally lower energy state • Boltzmann gives distribution of E states • So some probability of higher energy state • i.e. ‘going uphill’ • Probability of ‘uphill’ decreases as T decreases

  33. Implementation • System changes from E1 to E2 with probability exp[-(E2-E1)/kT] • If(E2< E1), P>1 (threshold at 1) • System will take this option • If(E2> E1), P<1 • Generate random number • System may take this option • Probability of doing so decreases with T

  34. P= exp[-(E2-E1)/kT] – rand() - OK Simulated Annealing P= exp[-(E2-E1)/kT] – rand() - X T

  35. Simulated Annealing • Rate of cooling very important • Coupled with effects of k • exp[-(E2-E1)/kT] • So 2xk equivalent to state of T/2 • Used in a range of optimisation problems • Not much used in Remote Sensing

  36. (Artificial) Neural networks (ANN) • Another ‘Natural’ analogy • Biological NNs good at solving complex problems • Do so by ‘training’ system with ‘experience’

  37. (Artificial) Neural networks (ANN) • ANN architecture

  38. (Artificial) Neural networks (ANN) • ‘Neurons’ • have 1 output but many inputs • Output is weighted sum of inputs • Threshold can be set • Gives non-linear response

  39. (Artificial) Neural networks (ANN) • Training • Initialise weights for all neurons • Present input layer with e.g. spectral reflectance • Calculate outputs • Compare outputs with e.g. biophysical parameters • Update weights to attempt a match • Repeat until all examples presented

  40. (Artificial) Neural networks (ANN) • Use in this way for canopy model inversion • Train other way around for forward model • Also used for classification and spectral unmixing • Again – train with examples • ANN has ability to generalise from input examples • Definition of architecture and training phases critical • Can ‘over-train’ – too specific • Similar to fitting polynomial with too high an order • Many ‘types’ of ANN – feedback/forward

  41. (Artificial) Neural networks (ANN) • In essence, trained ANN is just a (essentially) (highly) non-linear response function • Training (definition of e.g. inverse model) is performed as separate stage to application of inversion • Can use complex models for training • Many examples in remote sensing • Issue: • How to train for arbitrary set of viewing/illumination angles? – not solved problem

  42. Genetic (or evolutionary) algorithms (GAs) • Another ‘Natural’ analogy • Phrase optimisation as ‘fitness for survival’ • Description of state encoded through ‘string’ (equivalent to genetic pattern) • Apply operations to ‘genes’ • Cross-over, mutation, inversion

  43. Genetic (or evolutionary) algorithms (GAs) • E.g. of BRDF model inversion: • Encode N-D vector representing current state of biophysical parameters as string • Apply operations: • E.g. mutation/mating with another string • See if mutant is ‘fitter to survive’ (lower RMSE) • If not, can discard (die)

  44. Genetic (or evolutionary) algorithms (GAs) • General operation: • Populate set of chromosomes (strings) • Repeat: • Determine fitness of each • Choose best set • Evolve chosen set • Using crossover, mutation or inversion • Until a chromosome found of suitable fitness

  45. Genetic (or evolutionary) algorithms (GAs) • Differ from other optimisation methods • Work on coding of parameters, not parameters themselves • Search from population set, not single members (points) • Use ‘payoff’ information (some objective function for selection) not derivatives or other auxilliary information • Use probabilistic transition rules (as with simulated annealing) not deterministic rules

  46. Genetic (or evolutionary) algorithms (GAs) • Example operation: • Define genetic representation of state • Create initial population, set t=0 • Compute average fitness of the set • Assign each individual normalised fitness value • Assign probability based on this • Using this distribution, select N parents • Pair parents at random • Apply genetic operations to parent sets • generate offspring • Becomes population at t+1 7. Repeat until termination criterion satisfied

  47. Genetic (or evolutionary) algorithms (GAs) • Flexible and powerful method • Can solve problems with many small, ill-defined minima • May take huge number of iterations to solve • Not applied to remote sensing model inversions

  48. Knowledge-based systems (KBS) • Seek to solve problem by incorporation of information external to the problem • Only RS inversion e.g. Kimes et al (1990;1991) • VEG model • Integrates spectral libraries, information from literature, information from human experts etc • Major problems: • Encoding and using the information • 1980s/90s ‘fad’?

  49. LUT Inversion • Sample parameter space • Calculate RMSE for each sample point • Define best fit as minimum RMSE parameters • Or function of set of points fitting to a certain tolerance • Essentially a sampled ‘exhaustive search’

More Related