340 likes | 369 Views
Sequential Bayesian Prediction in the Presence of Changepoints. Roman Garnett, Michael Osborne, Stephen Roberts Pattern Analysis Research Group Department of Engineering University of Oxford.
E N D
Sequential Bayesian Prediction in the Presence of Changepoints Roman Garnett, Michael Osborne, Stephen Roberts Pattern Analysis Research Group Department of Engineering University of Oxford
We employ Gaussian processes to perform Bayesian prediction when our data includes changepoints. missing correlated delayed
The Gaussian distribution allows us to produce distributions for variables conditioned on any other observed variables.
The Gaussian distribution allows us to produce distributions for variables conditioned on any other observed variables.
A Gaussian process is the generalisation of a multivariate Gaussian distribution to a potentially infinite number of variables.
A Gaussian process represents a powerful way to perform Bayesian inference about functions. • We want to consider functions of time x(t).
A Gaussian process produces a mean estimate along with an indication of the uncertainty in it.
Covariance functions inform the Gaussian process of how we expect the function to vary with input e.g. time. We commonly possess prior expectations that the function should be smooth.
Covariance functions are very flexible. They allow us to express knowledge of periodicity correlated sensors long-term drifts
Associated with our covariance function are a number of hyperparametersf, such as periods correlations amplitudes
Similarly, we require hyperparameters for changepoint covariance functions. input scale pre-changepoint changepoint location input scale post-changepoint
( j ) I p x x D ? ; R ( j ) ( j ) ( j ) Á Á Á d Á I I I p x x p x p D D ? ; ; ; d P i t t x : r e c a n s = ? R ( j ) ( j ) Á Á d Á I I p x p D d P i ; t x : r e c o r s D I C t t : o n e x We assign a prior distribution to each hyperparameter and then marginalise them. Unfortunately, as this is nonanalytic, we are forced to use numerical integration.
We evaluate our predictions for each of a sample set of hyperparameters fS,e.g. f = the period of x(t).
Bayesian Monte Carlo assigns another Gaussian process to the integrand as a function of the covariance hyperparameters f.
We propagate through time one Gaussian process for each of our sample set fS, adjusting the weights according to the data as we go.
We use similar techniques to determine posterior distributions for any hyperparameters of interest.
We use an iterative formulation of our Gaussian processes that allows us to efficiently update our predictions. Similarly, we discard old data when the algorithm judges it sufficiently uninformative.
A GP with a changepoint covariance is able to effectively track data with changepoints.
We can also determine the posterior distribution for the location of changepoints.
Bramblemet is a network of wireless weather sensors deployed in the Solent.
Bramblemet is used by port authorities and recreational sailors. We would like to use its local, recent observations to produce accurate predictions.
The posterior distribution for the changepoint location clearly identifies the fault.
We have introduced a sequential Gaussian process algorithm to perform prediction in the presence of changepoints.
2 3 , , µ µ µ T 1 ( [ ( ] [ 0 0 ) 0 ] ) ( ) ( 0 ) ( ) ( 0 ) l l l l d l l d d d K K K K i i t t t t c o s c o s c o s ¡ ¡ 1 2 4 s s a g g a g g l b l l b l l l 0 i t a e a e m e ; ; ; ; ; ; µ µ µ µ µ i i i 0 s n s n c o s s n c o s 1 2 3 4 5 6 7 s e g = 6 7 . . µ µ µ µ µ i i i i 0 0 s n s n s n s n c o s 2 3 4 5 6 4 5 µ µ µ i i i 0 0 0 s n s n s n 4 5 6 Our covariance is a product of a term over sensor label and a term over time. delays typically Matern class length scales spherical parameterisation