430 likes | 621 Views
Maths and Statistics Seminar Series. 2: Experimental Design John Fenlon j.fenlon@warwick.ac.uk. A statistical career. RA @ Warwick Univ. (1969-71) SO, HSO, SSO @ GRI, nr Reading (1971-79) SSO, Water Quality, Severn-Trent (1979-83) PSO @ GCRI, L’hampton (1983-1995)
E N D
Maths and Statistics Seminar Series 2: Experimental Design John Fenlon j.fenlon@warwick.ac.uk
A statistical career • RA @ Warwick Univ. (1969-71) • SO, HSO, SSO @ GRI, nr Reading (1971-79) • SSO, Water Quality, Severn-Trent (1979-83) • PSO @ GCRI, L’hampton (1983-1995) • Head of Statistics @ HRI, Littlehampton, then Wellesbourne (1987-2004) • Director, RISCU & Reader in Statistics @ Univ. Warwick (2003-2011) • Consultant, RISCU & Teaching Fellow (2011-)
Two Farmers • A uses variety X and gets a yield of 24.1 • B uses variety Y and gets 21.4 • A grows X on a second field and gets 24.2
Two types of study • Observational studies • An observational study compares populations • Designed experiments • An experiment compares treatments • Example: randomised clinical trial vs. Similar study using hospital records
Statistical design allows us to • avoid poor experimentation • make efficient / ethical use of resources • distinguish between signal and noise • Choose the right size of experiment • work with general heterogeneity of experimental material, by grouping into homogeneous blocks • design a sequence of experiments within an overall research programme
Publications with ‘Experimental Design’ or ‘Design of Experiments’ in the title
Areas of application • Agricultural field experiments • Ecological / environmental • Food processing / baking, etc • Industrial (e.g. chemical, pharma., bio-genetics) • Engineering • Medical • Transport / environment • Educational • Computer experiments
Some simple examples • We wish to compare the potential of several different varieties of wheat. How might we test this, and what inferences can we draw from the results? • How might we determine what type of battery would give the longest life in a torch? • It is thought that a new method of road-management might work better at certain types of junction. How should we go about testing this?
Some more simple examples • How should we maximise the yield of a chemical product which we know is dependent on operating temperature and running time? • A pharmaceutical company has developed a new drug to treat a particular disease. How can they test its efficacy? • The Government need to test whether genetically-modified (GM) crops have an impact on the environment. How might they go about this?
HISTORICAL • Rothamsted • Agricultural field trials (Fisher & Yates) • World War II • Industrial statistics • Impact of Computers • Medicine + other disciplines • Engineering (Taguchi & robust design) • Optimal Design • Computer Experiments
Classes of Experimental Problems (Wu and Hamada, 2000) • Treatment comparisons • Variable screening • Response surface exploration • System optimisation • System robustness
Experimental Units • Experimental unit: smallest division of the experimental material such that any two units may receive different treatments in the actual experiment • Examples: • a plot of land • a patient in a hospital • a class of students • a lump of dough • a group of animals in a pen • a specific run on a machine with given conditions.
Comparative experiments • Treatments applied at different times / places will almost certainly produce different means • It is important to compare treatments on material / units that are as similar as possible • In designed experiments we can infer a causal effect of treatments
The Three R’s of Experimental Design • Replication • Randomisation • R...Blocking • Representativeness
Replication Replication is the process of running the same treatment on different (i.e. independ-ent) experimental units. Examples: • Different mice in an assay • Running a reaction again • More than one plot in a field trial It does NOT mean repeating the reading, or sampling within a unit!
Replication • Provides a true estimate of variance • Helps to avoid of outliers • How much replication depends on • resources available • variability of the experimental units • treatment structure • size of effect that we want to detect • relative importance of different comparisons
Randomisation In practice, randomisation means that, once the units for the trial have been selected, it is entirely a matter of chance which unit receives which treatment. Furthermore, the selection of one particular treatment-unit combination should have no influence on the treatment received by the unit that is adjacent in space or time.
RANDOMISATION • Provides a valid estimate of error • Guards against bias • Systematic bias • Selection bias • Accidental bias • Cheating • Use of ‘blind’, ‘double-blind’ principles in experimental trials
Mechanics of randomisation • ‘Bingo’ • Dice, coins and cards • Random number tables • Computer programs
Blocking – local control • Exploiting variation to advantage • Compare treatments on homogeneous material • Eliminate ‘known’ sources of variation in material • Examples • Plots in field trials • Animals in litters • Times of day • Patients of a given age / medical history • Positions in a glasshouse • Different labs
A food chemistry experiment We wish to compare the nutritional quality of 5 brands of pizza (A, B, C, D and E) using four different labs Here the labs are blocks
Representativeness How representative is the experiment for the material for which inference is to be made? Examples: • a nutrition experiment on specific breed of cow – is it representative of all cows? • we test various engine oils on a Ford engine – can we assume that the results will hold for a VW engine?
Design & Analysis • Analysis Of Variance (ANOVA) • A way of partitioning the variance in the experiment and attributing it to the design and treatment components • A Linear Model • The Design defines a specific model which determines the analysis. • The replication error gives us a measure against which we can compare other effects (e.g. Blocks & Treatments)
Treatments in Designed Experiments • Essentially factors can be classified in two ways: quantitative and qualitative. • Classical experimental design in agriculture was primarily qualitative (e.g. varieties of wheat, types of feed, etc.) although fertiliser experiments, sowing rates, etc are quantitative. • In contrast industrial experiments were often done with purely quantitative factors, which quite naturally led to consideration of the response surface for an experiment – in the same way that one might look at the response profile for a quantitative response in agriculture.
Factorial treatment sets Allow us to address more than one question in the same analysis • Not just “Are the treatments different?” • Are there differences between the levels of each factor? • Do the differences between the levels of one factor depend on the level of the other factor at which we observe the response? i.e. Is there an interaction? • Does the form of the interaction between two factors depend on the level of a third factor at which we observe the response?
Example (oil residues) Treatments: two reference oils RL206 and RL133 Design: one furnace running a.m. or p.m. two different tubes (3 & 4) two replicates
A Flour experiment Four flour formulations (four levels of factor A) X two yeast levels (factor N, low or high) X two proof times (factor S, short or long) X degree of mixing (factor Q, two levels) X dough time delay (factor T, short or long). This was a 4 x 24 experiment with 64combinations. Only 32 combinations could be used in two blocks of 16
Table 15.17 (DV1999, p.504): A blocked ½-fraction of a 4 x 24 experiment on bread Only one block (of two) shown
Antony & Antony (2001) • Objective: “to identify the optimal settings of control factors which would maximise the flight time of paper helicopters (with minimum variation)” • Control factors: those that can be easily controlled and varied by the designer or operator
Example of control factors in the helicopter expt (Table 2 from A²)
A Checklist for Designing Experiments • Define Objectives • Identify all sources of variation • Choose a rule for assigning units to treatments • Specify measurement & procedure • Run a pilot experiment • Specify the model • Outline the analysis • Determine the number of observations • Review the above and revise (if nec.)