Abstract
Evidence-Based Medicine (EBM) developed from the work of clinical epidemiologists at McMaster University and Oxford University in the 1970s and 1980s and self-consciously presented itself as a "new paradigm" called "evidence-based medicine" in the early 1990s. The techniques of the randomized controlled trial, systematic review and meta-analysis have produced an extensive and powerful body of research. They have also generated a critical literature that raises general concerns about its methods. This paper is a systematic review of the critical literature. It finds the description of EBM as a Kuhnian paradigm helpful and worth taking further. Three kinds of criticism are evaluated in detail: criticisms of procedural aspects of EBM (especially from Cartwright, Worrall and Howick), data showing the greater than expected fallibility of EBM (Ioaanidis and others), and concerns that EBM is incomplete as a philosophy of science (Ashcroft and others). The paper recommends a more instrumental or pragmatic approach to EBM, in which any ranking of evidence is done by reference to the actual, rather than the theoretically expected, reliability of results. Emphasis on EBM has eclipsed other necessary research methods in medicine. With the recent emphasis on translational medicine, we are seeing a restoration of the recognition that clinical research requires an engagement with basic theory (e.g. physiological, genetic, biochemical) and a range of empirical techniques such as bedside observation, laboratory and animal studies. EBM works best when used in this context.
Similar content being viewed by others
Notes
The term “evidence-based practice” may be replacing EBM, acknowledging the fact that the practice of medicine requires not only physicians but other health care professionals.
Some canonical definitions are unhelpful for a general understanding of EBM, for example, that given in (Sackett et al. 1996): “Evidence-based medicine is the conscientious, explicit and judicious use of current best evidence in making decisions about the care of individual patients.” This particular definition is probably used widely because it is brief, and because it has a rhetorical purpose—to address the frequent criticism of EBM that it applies to populations, not individuals.
“Double-masked” has largely replaced “double-blinded,” in order to avoid inappropriate use of terms relating to disability.
A useful essay discussing the differences and interactions between EBM and MDM is (Elstein 2004)
They were not, however, claiming along with many Kuhnians that there is subjectivity or relativity involved in what gets to count as evidence.
There are many such hierarchies in use, but all put double-masked RCTs at the top, or right after meta-analyses of RCTs, and clinical experience, expert consensus and physiological rationale at the bottom.
Interestingly, Sehon and Stanley (2003) argue that EBM should not be thought of as a Kuhnian paradigm, but, instead, as part of a Quinean holistic network of beliefs. My focus here is on the historic, rather than the popular meaning of “paradigm” and on what EBM is, not on what it should be.
There are many other characteristics of Kuhnian paradigms, such as the narrative of paradigm change, the emphasis on high-level theory, and the idea that paradigms replace one another, that do not apply to this case. I am not trying to apply Kuhn exhaustively, just to use some of his concepts where they may be explanatorily useful.
Double-masking is the standard for high quality RCTs.
Critics of EBM have occasionally presented EBM as a political and rhetorical movement, e.g. (Charlton and Miles 1998), emphasizing the ways in which it appears to lack rationality.
Cartwright’s (2010) four conditions are knowledge of “Roman laws“(laws that are general enough), “the right support team” (all necessary conditions), “straight sturdy ladders” (for climbing up and down levels of abstraction) and “unbroken bridges” (no interfering conditions).
Worrall might respond that he is only criticizing those methodologists who make abstract and general claims about freedom from “all possible” biases. This is fine, so long as no conclusions are drawn for RCTs in practice.
Michael Rawlins is the head of NICE (National Institute of Health and Clinical Excellence) in the UK, which bases its policies and guidelines on the results of EBM.
In this context, a positive trial is one in which the experimental arm of the trial is more effective, a null result is one in which both arms are equally effective, and a negative trial is one in which the control arm is more effective.
I recommend that in the meantime we correct for funding bias by asking for a higher level of significance from the results of trials funded by pharmaceutical companies.
An editorial in the same issue of NEJM (Pocock and Elbourne 2000) strongly protests these conclusions, partly in the name of EBM orthodoxy, but partly also on the basis of some well known RCTs which contradicted the results of observational trials
In fact, EBM is the successor to ancient “empiric” approaches to medicine.
Cartwright is, of course, a realist about underlying causal processes.
Technically, “translational research” includes both the bench-to-bedside-and-back (T1) and the clinical research to everyday practice (T2) “translational blocks.” See (Woolf 2008). But most of the resources and rhetoric favor the former.
References
AHRQ. (2002). Systems to rate the strength of scientific evidence. Evidence Report/Technology Assessment 47,
Als-Nielsen, B., Chen, W., Gluud, C., & Kjaergard, L. L. (2003). Association of funding and conclusions in randomized drug trials: A reflection of treatment effect or adverse events? JAMA: The Journal of the American Medical Association, 290(7), 921–928.
Ashcroft, R. E. (2004). Current epistemological problems in evidence based medicine. Journal of Medical Ethics, 30(2), 131–135.
Baratz, S. R., Goodman, C., & Council on Health Care Technology. (1990). Improving consensus development for health technology assessment: An international perspective. Washington: National Academy Press.
Bekelman, J. E., Li, Y., & Gross, C. P. (2003). Scope and impact of financial conflicts of interest in biomedical research: A systematic review. JAMA: The Journal of the American Medical Association, 289(4), 454–465.
Benson, K., & Hartz, A. J. (2000). A comparison of observational studies and randomized, controlled trials. The New England Journal of Medicine, 342(25), 1878–1886.
Bero, L., Oostvogel, F., Bacchetti, P., & Lee, K. (2007). Factors associated with findings of published trials of drug-drug comparisons: Why some statins appear more efficacious than others. PLoS Medicine, 4(6), e184.
Berwick, D. M. (2005). Broadening the view of evidence-based medicine. Qual. Saf. Health Care, 14(5), 315–316.
Bluhm, R. (2005). From hierachy to network: a richer view of evidence for evidence-based medicine. Perspectives in Biology and Medicine, 48(4), 535–547.
Canadian Task Force on Preventive Health Care. CTFPHC History/Methodology. Retrieved 07/07, 2009, from http://www.ctfphc.org
Cartwright, N. (1989). Nature's capacities and their measurement. Oxford; New York: Clarendon Press; Oxford University Press.
Cartwright, N. (2007a). Are RCTs the gold standard? Biosocieties, 2(2), 11–20.
Cartwright, N. (2007b). Evidence-based policy: Where is our theory of evidence? Center for Philosophy of Natural and Social Science, London School of Economics, Technical Report 07/07,
Cartwright, N. (2009). Evidence-based policy: What's to be done about relevance. Philosophical Studies, 143(1), 127–136.
Cartwright, N. (2010). Will this policy work for you? predicting effectiveness better: How philosophy helps. Unpublished manuscript.
Charlton, B. G., & Miles, A. (1998). The rise and fall of EBM. Quarterly Journal of Medicine, 91(5), 371-371-374.
Cohen, A. M., Stavri, P. S., & Hersh, W. R. (2004). A categorization and analysis of the criticisms of evidence-based medicine. International Journal of Medical Informatics, 73(1), 35-35-43.
Concato, J., Shah, N., & Horwitz, R. I. (2000). Randomized, controlled trials, observational studies, and the hierarchy of research designs. The New England Journal of Medicine, 342(25), 1887–1892.
Daly, J. (2005). Evidence-based medicine and the search for a science of clinical care. Berkeley; New York: University of California Press; Milbank Memorial Fund.
Davidoff, F., Haynes, B., Sackett, D., & Smith, R. (1995). Evidence based medicine. BMJ (Clinical Research Ed.), 310(6987), 1085–1086.
Denny, K. (1999). Evidence-based medicine and medical authority. Journal of Medical Humanities, 20(4), 247–263.
Doll, R., Peto, R., & Clarke, M. (1999). First publication of an individually randomized trial. Controlled Clinical Trials, 20(4), 367–368.
Elstein, A. S. (2004). On the origins and development of evidence-based medicine and medical decision making. Inflammation Research, 53(Supplement 2), S184-S184-S189.
Epstein, S. (1996). Impure science: AIDS, activism, and the politics of knowledge. Berkeley: University of California Press.
Evidence-Based Medicine Working Group. (1992). Evidence-Based Medicine. A new approach to teaching the practice of medicine. JAMA: The Journal of the American Medical Association, 268(17), 2420–2425.
Feinstein, A. R., & Horwitz, R. I. (1997). Problems in the “evidence” of “evidence-based medicine”. The American Journal of Medicine, 103(6), 529–535.
Goodman, N. (1955). Fact, fiction and forecast. Cambridge: Harvard University Press.
Goodman, S. N. (2002). The mammography dilemma: A crisis for evidence-based medicine? Annals of Internal Medicine, 137(5 Part 1), 363–365.
Guyatt, G. H., Oxman, A. D., Vist, G. E., Kunz, R., Falck-Ytter, Y., Alonso-Coello, P., et al. (2008). GRADE: An emerging consensus on rating quality of evidence and strength of recommendations. BMJ (Clinical Research Ed.), 336(7650), 924–926.
Hampton, J. R. (2002). Evidence-based medicine, opinion-based medicine, and real-world medicine. Perspectives in Biology and Medicine, 45(4), 549–568.
Harari, E. (2001). Whose evidence? lessons from the philosophy of science and the epistemology of medicine. Australian and New Zealand Journal of Psychiatry, 35(6), 724-724-730.
Hopewell, S., Clarke, M. J., Stewart, L., & Tierney, J. (2007). Time to publication for results of clinical trials. Cochrane Database of Systematic Reviews, (1) doi:10.1002/14651858.MR000011.pub2
Howick, J. (2008). Against a priori judgments of bad methodology: Questioning double-blinding as a universal methodological virtue of clinical trials. PhilSci Archive
Howick, J. (2011). The philosophy of evidence-based medicine Wiley-Backwell.
Ioannidis, J. P. (2005). Contradicted and initially stronger effects in highly cited clinical research. JAMA: The Journal of the American Medical Association, 294(2), 218–228.
Juni, P., Witschi, A., Bloch, R., & Egger, M. (1999). The hazards of scoring the quality of clinical trials for meta-analysis. JAMA: The Journal of the American Medical Association, 282(11), 1054–1060.
Kuhn, T. S. (1962). The structure of scientific revolutions. Chicago: University of Chicago Press.
Kuhn, T. S. (1970). The structure of scientific revolutions (2nd ed.). Chicago: University of Chicago Press.
Kuhn, T. S. (1977). The essential tension: Selected studies in scientific tradition and change. Chicago: University of Chicago Press.
Kuhn, T. S., Conant, J., & Haugeland, J. (2000). The road since structure: Philosophical essays, 1970–1993, with an autobiographical interview. Chicago: University of Chicago Press.
Lakatos, I., & Musgrave, A. (1970). Criticism and the growth of knowledge. Cambridge: University Press.
LeLorier, J., Gregoire, G., Benhaddad, A., Lapierre, J., & Derderian, F. (1997). Discrepancies between meta-analyses and subsequent large randomized, controlled trials. The New England Journal of Medicine, 337(8), 536–542.
Lexchin, J., Bero, L. A., Djulbegovic, B., & Clark, O. (2003). Pharmaceutical industry sponsorship and research outcome and quality: Systematic review. BMJ (Clinical Research Ed.), 326(7400), 1167–1170.
Marks, H. M. (1997). The progress of experiment: Science and therapeutic reform in the united states, 1900–1990. Cambridge: Cambridge University Press.
Meldrum, M. L. (1998). A calculated risk: The salk polio vaccine field trials of 1954. British Medical Journal, 317, 1233-1233-6.
Meldrum, M. L. (2000). A brief history of the randomized controlled trial: From oranges and lemons to the gold standard. Hematology/Oncology Clinics of North America, 14(4), 745-745-760.
Montgomery, K. (2006). How doctors think: Clinical judgment and the practice of medicine. Oxford: Oxford University Press.
Pocock, S. J., & Elbourne, D. R. (2000). Randomized trials or observational tribulations? The New England Journal of Medicine, 342(25), 1907–1909.
Rawlins, M. (2008). De testimonio: On the evidence for decisions about the use of therapeutic interventions. Lancet, 372(9656), 2152–2161.
Sackett, D. L., Rosenberg, W. M., Gray, J. A., Haynes, R. B., & Richardson, W. S. (1996). Evidence based medicine: What it is and what it isn't. BMJ (Clinical Research Ed.), 312(7023), 71–72.
Sehon, S. R., & Stanley, D. E. (2003). A philosophical analysis of the evidence-based medicine debate. BMC Health Services Research, 3(14).
Shrier, I., Boivin, J. F., Steele, R. J., Platt, R. W., Furlan, A., Kakuma, R., et al. (2007). Should meta-analyses of interventions include observational studies in addition to randomized controlled trials? A critical examination of underlying principles. American Journal of Epidemiology, 166(10), 1203–1209.
Solomon, M. (2008). Epistemological reflections on the art of medicine and narrative medicine. Perspectives in Biology and Medicine, 51(3), 406–417.
Solomon, M. (in progress). Beyond the art and science of medicine. Unpublished manuscript.
Straus, S. E., & McAlister, F. A. (2000). Evidence-based medicine: A commentary on common criticisms. Canadian Medical Association Journal, 163(7), 837–841.
Tonelli, M. R. (1998). The philosophical limits of evidence-based medicine. Academic Medicine, 73(12), 1234–1240.
Tonelli, M. R. (2006). Integrating evidence into clinical practice: An alternative to evidence-based approaches. Journal of Evaluation in Clinical Practice, 12(3), 248–256.
Woolf, S. H. (2008). The meaning of translational research and why it matters. JAMA: The Journal of the American Medical Association, 299(2), 211–213.
Worrall, J. (2002). What evidence in evidence-based medicine? Philosophy of Science, 69(S3), 316–330.
Worrall, J. (2007a). Evidence in medicine and evidence-based medicine. Philosophy Compass, 2(6), 981–1022.
Worrall, J. (2007b). Why there's no cause to randomize. The British Journal for the Philosophy of Science, 58(3), 451–488.
Yank, V., Rennie, D., & Bero, L. A. (2007). Financial ties and concordance between results and conclusions in meta-analyses: Retrospective cohort study. BMJ (Clinical Research Ed.), 335(7631), 1202–1205.
Author information
Authors and Affiliations
Corresponding author
Additional information
I am grateful to two anonymous reviewers for the European Journal for Philosophy of Science for helpful corrections and comments on this paper.
Rights and permissions
About this article
Cite this article
Solomon, M. Just a paradigm: evidence-based medicine in epistemological context. Euro Jnl Phil Sci 1, 451 (2011). https://doi.org/10.1007/s13194-011-0034-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s13194-011-0034-6