430 likes | 545 Views
Summing Up: Research Process Revisited. Research Methods for Public Administrators Dr. Gail Johnson. What is Research?. A systematic search for answers to questions. Search: to uncover, examine, find by exploration, to investigate, to inquire.
E N D
Summing Up:Research Process Revisited Research Methods for Public Administrators Dr. Gail Johnson Dr. G. Johnson, www.ResearchDemystified.org
What is Research? • A systematic search for answers to questions. • Search: to uncover, examine, find by exploration, to investigate, to inquire. • Research: "the systematic inquiry into a subject in order to discover or revise facts, theories Dr. G. Johnson, www.ResearchDemystified.org
Deductive Vs. Inductive Logics • Deductive: Sherlock Holmes • Testing theory • Inductive: • Building theory from the ground up Dr. G. Johnson, www.ResearchDemystified.org
Planning Phase 1. Determining your questions 2. Identifying your measures and measurement strategy 3. Selecting a research design 4. Developing your data collection strategy A. Methods B. Sample 5. Identifying your analysis strategy 6. Reviewing and testing your plan Dr. G. Johnson, www.ResearchDemystified.org
Planning: An Iterative Process • Not as linear as its presented • The process is like going through a funnel • It feels like going around in circles but things get very narrow and focused-at the end • There are no shortcuts Dr. G. Johnson, www.ResearchDemystified.org
Step 1: What’s the Question? • Identify the issue or concern • What’s already known? • Literature review, talk with experts • Is there a theory? • Engage the stakeholders • Clarify the issues • What matters most? • Choosing the question is deceptively difficult and takes way longer than people expect Dr. G. Johnson, www.ResearchDemystified.org
Types of Questions • Descriptive: what is • Normative: comparison to a target • Impact/causal: cause-effect • Tough to answer and require: • Logical theory • Time-order • Co-variation • Eliminates all other rival explanations Dr. G. Johnson, www.ResearchDemystified.org
Step 2: Developing a Measurement Strategy • Identify and define all the key terms. • Do tax cuts cause economic stability? • What tax cuts? What is economic stability? • Develop operational definitions that translates a concept into something that can be concretely measured. • How much money? How is economic stability measured? Will need to think about natural variation too. Dr. G. Johnson, www.ResearchDemystified.org
Step 2: Developing a Measurement Strategy • Boundaries. • Scope of the study: who’s included and excluded, time frame, geographic location(s). • Tax cuts during what time frame? Are we talking U.S.? • Unit of analysis. • If talking about U.S., then we need to look at national data. • If looking at a state, then we need to look at state data Dr. G. Johnson, www.ResearchDemystified.org
Levels of Measurement • Nominal: names and categories • Ordinal: has an order to it but not real numbers • Interval and Ratio: real numbers • Remember: different levels of measurement use different analytical techniques Dr. G. Johnson, www.ResearchDemystified.org
Validity and Reliability • Valid measures: the exact thing you want to measure • The number of books in a library is not a good measure of the quality of the school • Infant mortality rate—often used as an indicator of the quality of the health care system • Reliable measures: the exact thing measured in the exact same way every time • A steel ruler rather than an elastic ruler Dr. G. Johnson, www.ResearchDemystified.org
Reliability and Validity • Remember: • The poverty measure is reliable because it is measured the same way every time • But many question its validity because it does not accurately measure actual living costs and various financial benefits that some people might receive Dr. G. Johnson, www.ResearchDemystified.org
Step 3: Research Design The Xs and Os framework • Experimental • Random assignment to treatment or control (comparison) groups • Quasi-experimental • Non-random assignment to groups • Non-experimental • The one-shot design: implement a program and then measure what happens Dr. G. Johnson, www.ResearchDemystified.org
Question-design Connection • One-shot designs make sense for descriptive questions and normative questions but are weakest for cause-effect questions Dr. G. Johnson, www.ResearchDemystified.org
Question-Design Connection • The best design for a cause-effect question is the classic experimental design • But: quasi-experimental designs, including using statistical controls, are more typically used in public administration research • But: sometimes a one-shot design is as good as it gets in public administration research • Sophisticated users exercise great caution in drawing cause-effect conclusions Dr. G. Johnson, www.ResearchDemystified.org
Research Design: Other Common Research Approaches • Secondary Data Analysis • Evaluation Synthesis • Content Analysis • Survey Research • Case Studies • Cost-Benefit Analysis Dr. G. Johnson, www.ResearchDemystified.org
Step 4: Data Collection Options The decision depends upon: • What you want to know • Numbers or stories • Where the data resides • Environment, files, people • Resources available; time, money, staff to conduct the research Dr. G. Johnson, www.ResearchDemystified.org
Data Collection Methods • Locate sources of information • Data collection methods: • Available data • Archives, documents • Data collection instruments (DCIs) • Observation • Interviews, focus groups • Surveys: mail, in-person, telephone, cyberspace Dr. G. Johnson, www.ResearchDemystified.org
Multiple Methods • Quantitative-Qualitative war is over • Neither is inherently better • Law of the situation rules • Each work well in some situations, less well in others Dr. G. Johnson, www.ResearchDemystified.org
Multiple Methods • Quantitative and qualitative data collection often used together • Available data with surveys • Surveys with observations • Observations with available data • Surveys with focus groups Dr. G. Johnson, www.ResearchDemystified.org
Nonrandom Sampling • Useful in qualitative research • Sometimes a nonrandom sample is the only choice that makes sense • Weakness: potential selection bias • Do they hold a particular point of view that suits the agenda of the researchers? • Limitations: reflects only those included • Results are never generalizable Dr. G. Johnson, www.ResearchDemystified.org
Nonrandom Sampling • Non-random sample options • Quota • Accidental • Snow-ball • Judgmental • Convenience • Face validity: does the choice make sense? • Size is not important Dr. G. Johnson, www.ResearchDemystified.org
Random Sample • Based on probability: ever item in the population (people, files, roads, whatever) has an equal chance of being selected • Size matters: must be large enough—sample size table is needed • Advantages: • Ability to make inferences or generalizations about the larger population based on what we learn from the sample • Eliminates selection bias Dr. G. Johnson, www.ResearchDemystified.org
Random Sample • Challenge: • To locate a complete listing of the entire population from which to select a sample • Analysis requires inferential statistics • There is a calculable amount of error in any random sample • Confidence level, confidence intervals and sampling error (also called margin of error) • Tests of statistical significance Dr. G. Johnson, www.ResearchDemystified.org
Step 5: Analysis Plan • Analysis techniques vary based on: • Level of data collected • Sampling choices • The analysis plan links the data collection instruments, the questions and the planned data analyses • Check to make sure all the needed data are collected to answer the questions • Check to make sure unneeded data are not collected Dr. G. Johnson, www.ResearchDemystified.org
Step 6: Test Your Plan • Test all data collection instruments and data collection plans to make sure they work the way expected • Pre-test in real settings • Expert review • Cold-reader review • Revise and re-test • Finalize the research proposal Dr. G. Johnson, www.ResearchDemystified.org
The Design Matrix • A tool that helps pull on the pieces of the research plan together • Helps focus on all the details to make sure everything connects • It is a visual • Focus is on content not writing style • It is a living document • Planning is an iterative process • This is generic format Dr. G. Johnson, www.ResearchDemystified.org
“Begin with the end in mind.” • It is worth the time planning and testing your plan. • “If you do not know where you are going, you can wind up anywhere.” • It is hard to correct mistakes after the data has been collected • Remember: no amount of statistical wizardry will correct planning mistakes Dr. G. Johnson, www.ResearchDemystified.org
Doing Phase • Gathering the data • Preparing data for analysis • Analyzing and interpreting the data Dr. G. Johnson, www.ResearchDemystified.org
Doing Phase • Collect the data • Accuracy is key • Prepare data collected for analysis • Data entry, test error rate • Analyze the data • Qualitative approaches • Control for bias • Quantitative approaches • Numeric analysis • Interpretation in the English language Dr. G. Johnson, www.ResearchDemystified.org
Analysis Techniques • Descriptive • Frequency, percents, means, medians, modes, rates, ratios, rates of change, range, standard deviation • Bi-Variate • Cross-tabs, comparison of means Dr. G. Johnson, www.ResearchDemystified.org
Analysis Techniques • Relationships • Correlation and measures of association • Association does not mean the variables are causally related • The closer 0, the weaker the relationship, the closer to 1, the stronger the relationship • No defined rules: .2 to .3, something to look at, . 4 to .5 moderately strong, and above .5 is strong • It is rare to get correlations above .9 • Signal to take a more careful look at the measures Dr. G. Johnson, www.ResearchDemystified.org
Analysis Techniques • Inference • To infer something to the larger population based on the results of a random sample. • Confidence intervals (standard is 95% precision) • Confidence levels (standard is 95% confidence) • I am 95% certain (confident) that the true average salary in the population is between $45,000 and $50,000. • Sampling error (standard is plus/minus 5%) • Think polling data Dr. G. Johnson, www.ResearchDemystified.org
Analysis Techniques • Inference • Statistical Significance: How likely are we to have gotten these results from chance alone? • Many different tests to meet specific situations but interpretation is always the same • Standard practice: If there is a 5% chance or less that the results are due to chance, researchers will conclude that the results are statistically significant. Dr. G. Johnson, www.ResearchDemystified.org
Reporting Phase • What’s Your Point? • Major message/story? • Who is your audience? • Reporting Options • Executive summary • Written Reports • Oral presentations • Use of charts and tables Dr. G. Johnson, www.ResearchDemystified.org
Communication Guidelines • Present what matters to your audience • The goal is to illuminate, not impress • What’s your hook? • Grab your audience’s attention • Use clear, accurate and simple language • Use graphics to highlight points • Avoid jargon Dr. G. Johnson, www.ResearchDemystified.org
Communication Guidelines • Organize around major themes or research questions • Decide on your message and stick to it • Leave time for reviews by experts and cold readers • Leave time for the necessary revisions Dr. G. Johnson, www.ResearchDemystified.org
Communication Guidelines • Provide information about your research methods so others can judge its credibility • Always provide the limitations of this study • For reports: • Place technical information in an appendix • Provide an executive summary for busy readers Dr. G. Johnson, www.ResearchDemystified.org
Oral Presentations • Consider the needs of the audience • Consider the requirements of the situation • High tech or low tech? • Formal or informal? • Powerpoints as appropriate: simple, clear, large font, no distracting bells and whistles • Not too many, not too few: just right • Handouts as needed Dr. G. Johnson, www.ResearchDemystified.org
Visual Display Of Data • Tables: better for presenting data • Graphs/charts: more effective in communicating the message. • Impact • Increases audience acceptance • Increases memory retention • Shows big picture and patterns • Visual relief from narrative. Dr. G. Johnson, www.ResearchDemystified.org
Ethics and Values • Tell the truth--always. • Interpretations can have spin: may not be agreed criteria for what is “good.” • Build in checks to assure accuracy. • Be honest about the limitations of your research. • Do no harm to subjects of your research. Dr. G. Johnson, www.ResearchDemystified.org
Ethics and Values • Don’t take cheap shots at other people’s research • Don’t accuse of wrong-doing without evidence • Be careful not to harm people who benefit from program by concluding a program does not work when all you know is that your study was not able to find an impact. Dr. G. Johnson, www.ResearchDemystified.org
Creative Commons • This powerpoint is meant to be used and shared with attribution • Please provide feedback • If you make changes, please share freely and send me a copy of changes: • Johnsong62@gmail.com • Visit www.creativecommons.org for more information Dr. G. Johnson, www.ResearchDemystified.org