280 likes | 442 Views
Clinical trials and pitfalls in planning a research project. Dr. D. W. Green Consultant Anaesthetist King's College Hospital Denmark Hill London SE5 9RS with grateful thanks to Professor Alan Aitkenhead. Seven deadly scientific sins. insufficient information poor research
E N D
Clinical trials and pitfalls in planning a research project Dr. D. W. Green Consultant Anaesthetist King's College Hospital Denmark Hill London SE5 9RS with grateful thanks to Professor Alan Aitkenhead
Seven deadly scientific sins • insufficient information • poor research • inadequate sample size • no power analysis • no confidence intervals • biased • confounding factors e.g. mixed sexes for PONV • vague end points • e.g. not clearly defined severity of pain • straying from hypothesis
New Drugs: Types of study • Laboratory structure/activity analysis • Animal does it work in animals ? is it toxic ? • Human volunteers • Phase 1 .... Is it toxic ? • Phase 2 .... Does it work ? • Phase 3 .... Does it work better than existing drugs ? • Phase 4 .... Post marketing surveillance • What's it like in the real world ?
Background • has it been done before? • is it worth doing? • clinical scientific essential step • has anything similar been done before? • methods used by others?
Protocol • Introduction • background information • justification • why, what gap will it fill, what benefits • succinct • don’t miss out relevant info
Methodology: Ethics and consent • Crucial • Declaration of Helsinki • benefit to patients • benefit to society • Information to patients • purpose, what it involves • potential benefits, ability to withdraw • risks and disadvantages without prejudice • children and incompetent adults
Selection of patients • Age • efficacy and current disease • ASA status • Sex • pharmacokinetics, dynamics • e.g. PONV • Type of surgery • applicability and availability • Ability to give consent e.g. ICU • Pregnancy
Designs • prospective vs retrospective • open vs blind (double or single) • randomisation • acceptable methods eg envelopes opened after entering the trial • use of placebo • ethics and other treatments • block design • blocks of patients: • analyse after each block to enable one to stop when results are available • stratification • sequential analysis
Pitfalls • Funding salaries drugs, equipment and investigations e.g. NHS costs • statistics and data collection design • time …. how long do we go on for? • negative result … do (should) we publish? • contradictory results vs other studies • statistical and clinical effects • rival investigators
Assessment and measurements • which techniques • validity, accuracy, objective, analysis • which observer • blinded, nurses, how many make measurement, are they trained • how often • science, statistics, practicality over long periods, placebo effect of frequent assessments • number of variables, fewer the better • availability of test e.g. troponin T
Documentation • Ethics committee approval • patient information • data collection forms • data type, storage, security, confidentiality,safety • consent forms
Disproving the null hypothesis • The ‘null’ hypothesis is that there is no difference between the treatments • a probability value ‘p’ tells you how often the difference between the treatments could have occurred by chance. • p < 0.05 is 1 in 20 or less (statistically significant) • p < 0.01 is 1 in 100 or less (highly statistically significant)
Disproving the null hypothesis • Type I error is where a difference is shown which could have occurred by chance • 1 in 20 trials will show a difference where none exists if ‘p’ is reported at the 0.05 level • multiple subgroup analysis in a trial may also give subgroup treatment differences • a statistically significant result is more likely to be reported!
Disproving the null hypothesis • Type II error is showing no difference where one actually exists • almost always due to insufficient numbers • can mask beneficial treatment effects • BUT! if trial is large enough it may produce a statistically significant effect where the clinical significance is marginal
Size of study • Power of study to show a difference in Rx • (e.g. 70% chance of demonstrating a 15% difference with a p < 0.05)) • able disprove the null hypotheses with minimal or no Type II error • may require pilot to determine treatment differences • requires large numbers if differences are small or if great variability in treatment outcomes • lower power (smaller numbers) may be acceptable if outcome is important (e.g. leukaemia)
Assessment of population size 15% of patients die within one year of admission to hospital for suspected myocardial infarction. Preventing 1/3rd of these deaths would be a major advance. Roughly, how many patients are needed for a clinical trial if doctors want to be 90% sure that a difference between treatments as large as the prevention of 1/3rd of deaths will not be missed at the p < 0.05 level?
Presentation of results • Significance: clinical versus statistical • p values • confidence intervals (95%) (+/- 2 SE) • risk reduction (relative and absolute) • numbers needed to treat • odds ratios
Measures of risk reduction • Relative risk reduction …. Is it meaningful? • Headline “50% reduction in mortality” • if normal mortality is 50/100 this is great (25) • if normal mortality is 1/100 … (1 in 200) • Number needed to treat is better measure • reciprocal of risk reduction e.g. 4 in first (25/100) • 200 in the second (0.5/100) • If cost of treatment is £10,000 ………. !!
Number needed to treat • Control event rate is 9 cases in 30 (0.3) • Experimental event rate is 1 case in 29 (0.033) Then, NNT = 1/(CER - EER) = 1/(0.3-0.033) = 4 • This method corrects for relative and absolute risk by relating to the control event rate
Number needed to treat • Diabetic neuropathy 6.5 year prospective trial • 9.6% developed DN (conventional) • 2.8% developed DN (intensive treatment) • Relative risk reduction = (9.6-2.8)/9.6 = 71% • Absolute risk reduction = 9.6-2.8 = 6.8% • Number needed to treat = 1/.068 = 15 people for 6.5 years to prevent one case of DN
Odds ratios • OR are used where it is difficult to calculate the relative risk e.g. case control studies • A value greater than 1 assumes increased risk • Confidence intervals (95%) will give the overall picture (e.g. if CI crosses 1 then the result may not be significant
Odds ratio calculation • Calculated as the ratio of the results of the control group divided by the experimental group • (9/21) divided by (1/29) = 0.08 • The relationship between OR and NNT is not linear and is very confusing … even to statisticians!
Evidence based medicine The process of systematically finding, appraising and using contemporaneous research findings as a basis for clinical decisions
Evidence based medicine • Accurate identification of the clinical question to be investigated • a search of the literature to select relevant articles • evaluation of the evidence • implementation of the findings into clinical practise