160 likes | 175 Views
Explore why Central Baltic Lake GIG's results are acceptable, despite EC concerns. Learn about criteria, constraints, and recommendations for improving comparability and credibility in intercalibration results.
E N D
Comparability of Central Baltic Lake GIG results on macrophyte and phytoplankton composition Marcel van den Berg / Centre for Water Management The Netherlands Brussels, 5 March
Situation CB Lake GIG (ex DK) has agreed that their results are comparable (enough) but EC does not agree EC has proposed a comparability criterion for option 3 intercalibration that is not consistent with option 2 results
Why do we think that the results are acceptable? • CB GIG has clear criteria for comparability which are consistent with the ones as agreed in option 2 intercalibration (across GIGs rivers/lakes/co) • Assessment of reference sites results in c. 70% as ‘high’ for sites within MSs territory and c. 40% (20-90%) for across GIG (incomparability is thus not only inevitable but needed in order to comply with normative definitions) • Averaged opinion of all MS in the exercise shows acceptable relationship with pressure indicators given all limitations • Comparison is made not at site level but at lake level (=body level) • Remaining mismatches between MSs are not really due to size of the GIG but to the definitions of the types • The level of confidence of the weighed averaged comparability indicator is determined and is fairly acceptable
Why do we think that the results are acceptable? - Constrains In some cases data are incomplete to fully apply a method, and comparison was improved by testing full national method vs incomplete method (DE and BE) Small differences in EQR’s may result in different classes, specifically when the EQR-values are close to the class boundaries (by definition a misclassification of sites very close to the boundary is 50%). Typology is not best suited, and does especially not account for lake size. E.g. almost no overlap in data availability of largest lakes in Belgium vs smallest in other MSs.
CB Lake GIG has applied criteria consistent with other intercalibration results (e.g. option 2 rivers) Use of weighed averaged of misclassification of 0,25 quality class (=consistent with 0,05 EQR units) Use of frequency of misclassification of ca. 35% agreement (=60% with small stretch). This criterion is comparable with accepted statistical noise in option 2 intercalibration as expressed in R2 (Option 2 has agreed on R2=0,5, this criterion is not always achieved but results are in decisions for good reasons) Use of correlation coefficients of one MS against average (significance P<0.001)
CB Lake GIG has applied criteria consistent with other intercalibration results
CB Lake GIG has applied criteria consistent with other intercalibration results
Ultimate test: transpose results of option 3 to option 2; Example macrophytes LCB2 / G/M boundary
From option 3 to option 2; What about R2 for MSs methods vs Common metric? Results considering misclassifications are acceptable
Reference sites are predicted correctly as ‘high status’ for 70% of the sites within their own territory Reference sites But this percentage ‘high’ classifications is on average c.40% (15-90%) when a MS applies its method on outside its territory
Relation between averaged normalised EQRs of MSs vs pressure indicators and position of ref sites
Conclusion • CB Lake GIG work is comparable enough, and absolute disagreement is due to typological limitations and the fact that MSs methods are best suited for their own territory • Improvement for macrophytes can be achieved by agreement on European scale on common indicators e.g. maximum colonised depth • Improvement for phytoplankton can be achieved by agreeing on combination rules, and by using more sites in the comparison >>>BUT: To GIG will be proposed to exclude the results of phytoplankton composition, because of combination rules used by MSs are not clear.
Way forward / Recommendation • Comparability paper has to be rewritten slightly (but crucial) and has to be extended with more criteria for both credibility and acceptability of intercalibration results in the decision
Criteria for acceptability and credibility of intercalibration results **Example**