480 likes | 699 Views
Design of Experiments – Methods and Case Studies. Dan Rand Winona State University ASQ Fellow 5-time chair of the La Crosse / Winona Section of ASQ (Long) Past member of the ASQ Hiawatha Section. Design of Experiments – Methods and Case Studies. Tonight’s agenda The basics of DoE
E N D
Design of Experiments – Methods and Case Studies • Dan Rand • Winona State University • ASQ Fellow • 5-time chair of the La Crosse / Winona Section of ASQ • (Long) Past member of the ASQ Hiawatha Section
Design of Experiments – Methods and Case Studies • Tonight’s agenda • The basics of DoE • Principles of really efficient experiments • Really important practices in effective experiments • Basic principles of analysis and execution in a catapult experiment • Case studies – in a wide variety of applications • Optimization with more than one response variable • If Baseball was invented using DoE
Design of Experiments - Definition • implementation of the scientific method. -design the collection of information about a phenomenon or process, analyze information, learn about relationships of important variables. - enables prediction of response variables. - economy and efficiency of data collection minimize usage of resources.
Advantages of DoE • Process Optimization and Problem Solving with Least Resources for Most Information. • Allows Decision Making with Defined Risks. • Customer Requirements --> Process Specifications by Characterizing Relationships • Determine effects of variables, interactions, and a math model • DOE Is a Prevention Tool for Huge Leverage Early in Design
Steps to a Good Experiment • 1. Define the objective of the experiment. • 2. Choose the right people for the team. • 3. Identify prior knowledge, then important factors and responses to be studied. • 4. Determine the measurement system
Steps to a Good Experiment • 5. Design the matrix and data collection responsibilities for the experiment. • 6. Conduct the experiment. • 7. Analyze experiment results and draw conclusions. • 8. Verify the findings. • 9. Report and implement the results
An experiment using a catapult • We wish to characterize the control factors for a catapult • We have determined three potential factors: • Ball type • Arm length • Release angle
One Factor-at-a-Time Method • Hypothesis test - T-test to determine the effect of each factor separately. • test each factor at 2 levels. Plan 4 trials each at high and low levels of 3 factors • 8 trials for 3 factors = 24 trials. • levels of other 2 factors? • Combine factor settings in only 8 total trials.
Randomization • The most important principle in designing experiments is to randomize selection of experimental units and order of trials. • This averages out the effect of unknown differences in the population, and the effect of environmental variables that change over time, outside of our control.
Detecting interactions between factors • Two factors show an interaction in their effect on a response variable when the effect of one factor on the response depends on the level of another factor.
Predicted distance based on calculated effects • Distance = 80.84 + 4.94 * X2_arm_length – 4.19* X3_Release_angle – 2.06* X1*X3 - 5.94*X2*X3 • X2 = -1 at arm length of 10, = 1 at arm length of 12 • X3 = -1 at release angle of 45, = 1 at release angle of 75
Poorly executed experiments • If we are sloppy with control of factor levels or lazy with randomization, special causes invade the experiment and the error term can get unacceptably large. As a result, significant effects of factors don’t appear to be so significant.
The Best and the Worst • Knot Random Team and the String Quartet Team. Each team designed a 16-trial, highly efficient experiment with two levels for each factor to characterize their catapult’s capability and control factors.
Knot Random team results Mean square error = 1268 Demonstration of capability for 6 shots with specifications 84 ± 4 inches , Cpk = .34
String quartet result Mean square error = 362Demonstration of capability for 6 shots with specifications 72 ± 4 inches , Cpk=2.02
String Quartet Best Practices • Randomized trials done in prescribed order • Factor settings checked on all trials • Agreed to a specific process for releasing the catapult arm • Landing point of the ball made a mark that could be measured to ¼ inch • Catapult controls that were not varied as factors were measured frequently
Knot Random – Knot best practices • Trials done in convenient order to hurry through apparatus changes • Factor settings left to wrong level from previous trial in at least one instance • Each operator did his/her best to release the catapult arm in a repeatable fashion • Inspector made a visual estimate of where ball had landed, measured to nearest ½ inch • Catapult controls that were not varied as factors were ignored after initial process set-up
Multivariable testing (MVT) as DoE • “Shelf Smarts,” Forbes, 5/12/03 • DoE didn’t quite save Circuit City • 15 factors culled from 3000 employee suggestions • Tested in 16 trials, 16 stores • Measured response = store revenue • Implemented changes led to 3% sales rise
Census Bureau Experiment • “Why do they send me a card telling me they’re going to send me a census form???” • Dillman, D.A., Clark, J.R., Sinclair, M.D. (1995) “How pre-notice letters, stamped return envelopes and reminder postcards affect mail-back response rates for census questionnaires,” Survey Methodology, 21, 159-165
1992 Census Implementation Test • Factors: • Pre-notice letter – yes/ no • SASE with census form – yes / no • Reminder postcard a few days after census form – yes / no • Response = completed, mailed survey response rate
Surface Mount Technology (SMT) experiment - problem solving in a manufacturing environment • 2 types of defects, probably related • Solder balls • Solder-on-gold • Statistician invited in for a “quick fix” experiment • High volume memory card product • Courtesy of LallyMarwah, Toronto, Canada
Problem in screening / reflow operations Solder paste screening Component placement Prep card Solder paste reflow Clean card Inspect (T2) insert Inspect (T1)
8 potential process factors • Clean stencil frequency: 1/1, 1/10 • Panel washed: no, yes • Misregistration: 0, 10 ml • Paste height: 9ml, 12 ml • Time between screen/ reflow: .5, 4 hr • Reflow card spacing: 18 in, 36 in • Reflow pre-heat: cold, hot • Oven: A, B
Experiment design conditions • Resources only permit 16 trials • Get efficiency from 2-level factors • Measure both types of defects • Introduce T1 inspection station for counting defects • Same inspectors • Same quantity of cards per trial
7 more columns contain all interactions • Each column contains confounded interactions
Normal plot for factor effects on solder ball defects C D CD
Which confounded interaction is significant? • AF, BE, CD, or GH ? • The main effects C and D are significant, so engineering judgement tells us CD is the true significant interaction. • C is misregistration • D is paste height
Conclusions from experiment • Increased paste height (D+) acts together with misregistration to increase the area of paste outside of the pad, leading to solder balls of dislodged extra paste. • Solder ball occurrence can be reduced by minimizing the surface area and mass of paste outside the pad.
Implemented solutions • Reduce variability and increase accuracy in registration. • Lowered solder ball rate by 77% • More complete solution: • Shrink paste stencil opening - pad accommodates variability in registration.
The Power of Efficient Experiments • More information from less resources • Thought process of experiment design brings out: • potential factors • relevant measurements • attention to variability • discipline to experiment trials
Optimization – Back to the Catapult • Optimize two responses for catapult • Hit a target distance • Minimize variability • Suppose the 8 trials in the catapult experiment were each run with 3 replicates, and we used means and standard deviations of the 3
Maximum Desirability • Modeled response equation allows hitting the target distance of 84, d=1 • Best possible standard deviation according to model is 1.475 • d (for stddev) = (3-1.475)/(3-1) = .7625 • D = SQRT(1*.7625) = .873
How about a little baseball? • Questions? • Thank you • E-mail me at drand@winona.edu • Find my slides at http://course1.winona.edu/drand/web/