240 likes | 258 Views
2. Analytical Design. Method selection Validation. Background Reading: Quality Assurance in Analytical Chemistry, Chapter 4 VAM leaflet: Introduction to method validation (on Studentcentral). Define analytical requirements. Select/Develop candidate method. Plan validation experiments.
E N D
2. Analytical Design Method selection Validation Background Reading: Quality Assurance in Analytical Chemistry, Chapter 4 VAM leaflet: Introduction to method validation (on Studentcentral)
Define analytical requirements Select/Develop candidate method Plan validation experiments Conduct validation experiments NO Criteria met? YES Validation document Assess fitness for purpose Method selection vs Validation Based on: Quality Assurance in Analytical Chemistry, Chapter 4, Fig. 4.2
Principles of method selection (1) Components Practicalities Accuracy
Principles of Method Selection (2) • What are the analytes? • What properties do they have? • How much is present? • What else is present (matrix)? • Will it interfere – false positive or negative? • Over repeated runs? • How accurate do the results have to be? • How big/small an effect am I looking for? • What are the minimum amounts that must be detected? • Which must be eliminated – false positives or negatives? • What are the consequences of incorrect results? • How many samples must be analysed? • What resources are available? Components Accuracy Resources
Components: Analyte and Matrix The physico-chemical properties of components affect: • Separation methods: depends on all components (exploiting differences between analyte and matrix) • Detection method: must use analyte property • Level of selectivity needed: depends on scale and type of differences between analyte and matrix
Accuracy Successful analysis means setting, testing and meeting performance criteria • Qualititative or quantitative • Depend on the acceptable level of uncertainty (risk) • Criteria allow objective selection of sampling, separation and detection methods • Common criteria include measures of: • Precision • Selectivity/specificity • Bias • Ruggedness • Linearity (working range) • Limit of Detection / Quantitation
Resources Cost, availability of instruments, materials and staff will affect: • Number of samples • Available methods • Accuracy of measurements • Rate of turnaround Must be balanced against the consequences of an incorrect result
Finding a suitable method Sources of potential methods include: • Primary scientific literature • Patents • British & international standards (via UoB online library) • Manufacturer’s technical information You may find several potential methods, but: • It is rare to find one which is perfect • Objective criteria are needed to select the best method • Adapted methods must then be validated to make sure they are…
Thought Exercise –In groups • Write down five chemical properties • Name a separation technique that exploits differences in each of these • Name a detection method for each • Add an example of matrix and analyte for each of the five • Give an example of an analysis where each of the following might occur: • High risk (minimal consequence) false positive result • Low risk, false positive result • High risk, false negative result • Low risk, false negative result
Validation “Confirmation by the examination and the provision of objective evidence, that the particular requirements for a specific intended use are fulfilled.” Reference: ISO/IEC 17025:2005. General requirements for the competence of testing and calibration laboratories
How to validate? Purpose • What exactly are we analysing for? • Assay, iD, limit (impurities) • What are limits to the conditions the analysis covers? • What objective parameters will show whether the goals have been met? • Will these detect failures? • How best can the parameters be measured? • How should the data be compared to the specified parameters ?(statistics?, blind trials?) Performance Criteria Test Plan Interpretation
Precision Definition: “The closeness of agreement between independent test results obtained under specified conditions”. Includes reproducibility and repeatability. Affected by: Number of measurements, uncontrolled random errors. Measurement: Measures of spread (s, 95% CI etc.) Should include effect of factors that will not be consistent during normal use of method. Evaluation: Acceptable levels of precision depend on the levels of variability tolerated. Effect of concentration may be large. Bias also affects precision requirements. Student’s t and f tests. Are you hitting the bullseye?
Specificity Definition: “The extent to which a method can be used to determine particular analytes in mixtures or matrices without interference from the presence of components with similar behaviour.” Affected by: Types of components routinely present. Lack of specificity can give a false negative or positive. Measurement: Increasing concentrations of potential interferents added to samples. Can quantify at what concentrations interference becomes significant. Evaluation: Most important in trace analysis as contaminants can be significant. False positives can be neglected in screening assays, if followed by second confirmatory technique. Spot the Oak leaf?
Bias Definition: “The difference between the calculated value and the accepted reference standard”. A Measure of trueness Affected by: Systematic errors. Measurement: Spiking and recovery (how much of a known amount of analyte added at to the starting sample is measured by the analysis). Measurement with alternative validated method. Interlaboratory trials – to establish causes of bias. Evaluation: Simple t-test (compare result to known value). Bias and precision combine to give accuracy Altered gravity – or systematic building error?
Ruggedness Definition: The degree to which a method is affect by small changes in the operating conditions. Associated with both precision and bias. Affected by: type of technique, number of variable parameters Measurement: Deliberately vary conditions to quantify their effect on results, and identify critical parameters. Evaluation: Focused on identifying prime causes of variability, and setting controls for these. Don’t get stuck in the mud?
Linearity Definition: “The ability to produce test results that are proportional to the analyte concentration within a given range.” Affected by: Technique, interferents, recovery. Measurement: Calibration, ideally with CRMs or spiked samples. Concentrations must be evenly spaced. LOQ is often lower limit. Evaluation: May use visual inspection, r, runs test.
Limit of Detection (LOD) Definition: “The minimum concentration of analyte that can be detected with statistical confidence.” Affected by: method, uncertainty…the kitchen sink Measurement: Concentration (calculated from line of best fit) at which either (a) signal is equivalent to blank + 3 x sd of blank or (b) intercept (y0) + (3 x Sxy) Evaluation: No analysis should rely on a value below this. All but qualitative analyses should use the higher LOQ. Is this glass empty, or not?
Limit of Quantitation Definition: “The lowest concentration of analyte that can be determined with an acceptable level of uncertainty.” Affected by: Method performance…no really! Measurement: As for LOD but 10 x sd of blank Evaluation: This is the point at which quantitative analysis can be considered valid. May require multiple assays (alongside reproducibility studies to set limits appropriately). The world’s smallest ruler? 1 Division = 1.25µm
Thought exercise 2 You are to validate a new method for the analysis of calcium in infant formula: 1) What are the key features of this analysis? 2) What interferents might be important? 3) How would you decide what limits should be set for each of these parameters? • Precision • Selectivity/specificity • Bias • Ruggedness • Linearity (working range) • Limit of Detection / Quantitation 4) How would you determine if your analysis met the criteria specified?
Validation Documentation and QC Key Features include: • Description of method, including scope (what can it do, and what can it not do?) • All important technical details (how do I do it?) • Expected performance criteria (how well does it do it?) • Warning limits – normally 2-3 times within lab precision (how can you tell if it is not working?) • Responsible signatory, dated versions and revisions, document control to ensure currency See also end of Chapter 4: Quality Assurance in Analytical Chemistry
Typical Documentation • Analytical Procedure • Preparation of samples • Preparation of standards • Critical factors • Detailed description of all steps • Typical outputs; chromatograms, spectra, etc. • Recording and reporting of data • Method • Rounding and significant figures • Data treatments • Calculation of results • Calibration model • Calculation methods • Assumptions and limitations • Method performance • Statistical measures • Control charting • References & Bibliography • Scope and applicability • Samples • Analytes • Ranges • Description and principle of the method • Equipment • Specification • Calibration and qualification • Range of operability • Reference materials and reagents • Specification • Preparation • Storage • Health & Safety • Sampling • Methods • Storage • Limitations
Validation proposal activity 1 • Make a Word (or equivalent) template covering each of the items from the suggested table of contents for a validation document on the previous slide • Add your ideas from the thought exercise about calcium in infant formula • Decide which areas require further research and which may be answered during the practical sessions • Use the scientific literature to research possible answers to questions • Make notes on how you might decide which of two methods is more suitable – can you prioritise performance criteria? You will add to this document during the course of the module – it is designed to form the basis of your final assessment – the “validation proposal”.