170 likes | 188 Views
Detailed comparison of error detection methods for enhancing elevation accuracy in DEMs, focusing on PCA, elongated DTM, and local best fit approaches. Evaluation based on Mt. Sainte Victorie DEM study. Insights on strengths, weaknesses, and application scenarios provided.
E N D
On the improving of elevation accuracy of Digital Elevation Models: a comparison of some error detection procedures Carlos López http://www.fing.edu.uy/~carlos Centro de Cálculo Facultad de Ingeniería Montevideo URUGUAY
Our goals: • Improve elevation accuracy of grid-based DEM • Use general methods as far as possible • Satisfy both DEM producer & end user needs
Some assumptions... • a better value for the elevation can be obtained, but it might be too expensive to acquire it directly (GPS?) • once detected, the errors are corrected (“perfect inspector” hypothesis) • the editing cost is proportional to the number of candidate points
Organization of the presentation • Description of the three procedures • Results for the Mt. Sainte Victorie DEM • Discussion • Conclusions
The method by López (1997) (IJGIS 1997, 11, 7, 677-689) • A brief presentation of PCA • The method for an elongated DTM • The generalization to any DTM
A brief presentation of PCA • attempts to explain the behavior of clouds of points in Rw reducing the dimensionality of the data • it is usually applied to tabular (not raster!) datasets • the starting point is the cloud; all ordering among points (profiles) is lost
The elongated DTM case • The process requires two phases: • Identify the ¨suspicious¨ profiles • Analyze each of those profiles trying to pick in each the best candidate(s) for being an error • Any other procedure for tabular dataset can be used instead
Some remarks... • Even though we use PCA, our approach is not the standard one used in image processing • We do not use nor assume at all any model of covariance in respect with distance for the elevation • We locate errors based only upon the elevation (we will not consider slope neither curvature)
The generalization to any DTM • Any DTM can be considered as build from elongated ones, without intersection • We might look within each of those to locate errors • The procedure can be applied row-wise as well as column-wise • The most likely errors are those which are candidates both for column and row-wise analysis
The method of Felicísimo (JP&RS 1994, 49, 4, 29-33) • Compares a local best fit with a low order polinomial; the gross errors are obtained after an analisys of the residual • It is based on very simple hypothesis • uncorrelated errors in space • gaussian distribution of the errors • It can be easily implemented
Some problems of both methods • The assumptions of errors weakly correlated in space do not hold at least in the considered example. • Moreover, the performance of both method decreases if the spatial correlation increases
The modified procedure (TOG 2000, 4, 1, 43-64) • Since adjacent profiles are too correlated, we formed the strip choosing every kth. row from the DEM. • The implementation and the rationale are almost exactly the same as before. • It considers as a particular case the previous code.
The experiment • We used a SPOT derived DEM as a test bed, and consider another DEM of higher accuracy as the ground truth. • Once a location is selected, we correct the noisy DEM using the values from the other. The same point cannot be corrected twice. • We corrected as much as 15 per cent of the DEM. Various measures of the accuracy were recorded.
Discussion(1) • The new method outperforms the previous in the low effort region • The Felicísimo’s method in the long run gets mostly systematic errors • The RMSE might drop from 12.7 m to 11.0 m by checking only 1 per cent of the DEM; the max error drops from 193 m to 100 m • For efforts over 2.5 per cent the Felicisísimo’s method becomes better
Discussion(2) • Given a DEM, there are some parameters to be defined for our method. The suggested rules gave reliable values. • Once a suspicious point is suggested, some action needs to be taken. Different users might have different goals. • Despite the complexity of the details, the procedure requires only modest computer resources.
Conclusion (1) • Some advantages of the procedure • It is valid for any raster dataset • It might be of use both for data producers as well as for end users • It has some free parameters which can be tailored for specific needs, but we provide rules suitable for a first guess • It does not require any “model” for the dataset, neither at local nor global scale
Conclusion(2) • Some drawbacks of the method presented • It has been tested only with one DEM • It left unexploited some (maybe) important information from the dataset like the spatial correlation • Future work • should compare the methods with other DEM’s representative of different terrain characteristics