Authors

  1. Ramis, Mary-Anne RN, BN, MPhil (Clinical Sciences)
  2. Chang, Anne RN, BN, PhD, FRCNA
  3. Nissen, Lisa PhD

Abstract

Review question/objective: The objective of this review is to find, critically appraise and synthesize the available quantitative evidence on the effectiveness of interventions that promote successful teaching of the evidence-based practice process in undergraduate health students, in preparation for them to become professional evidence-based practitioners.

 

More specifically, the question that this review seeks to answer is: What is the effectiveness of teaching strategies for evidence-based practice for undergraduate health students?

 

Background: Learning and teaching the evidence-based practice (EBP) process presents great challenges to undergraduate students and educators alike, as providing students with knowledge alone, may not automatically mean students will feel capable to practice EBP in their eventual respective clinical environments.1,2 However EBP is now part of professional practice registration standards for many health disciplines to which new graduates are expected to adhere.3-7 Consequently, there is a growing responsibility for educators to address the requirement to ensure students have the necessary required knowledge, skills, attitude, and confidence.

 

In 2005, the Sicily statement on evidence-based practice was published after a consensus process of 86 international delegates of the Evidence-Based Health Care Teachers and Developers conference.8 In an attempt to provide some consistency to the evidence-based practice movement, the statement addressed the demand for a clearer definition of what evidence-based practice really was, as well as provided recommendations for the future. The authors also supported the change in EBP language, suggesting that the term Evidence-based Medicine, be changed to and accepted as "evidence-based practice", to reflect a broader construct and recognize and encourage a shared approach to EBP across health care fields. In recognizing the growing evidence-to-practice gap, the statement acknowledged the importance of effective teaching of EBP to undergraduate students as a way of preparing practitioners for the future.8

 

An update to the Sicily Statement was published in 2011 to provide direction for educators when developing tools to measure EBP dimensions from a learner perspective.9 The paper also provided a platform for further discussion on dimensions of EBP learning and teaching and suggested that effective EPB learning and evaluation should be "matched to the needs and characteristics of the learner audience".9(pg.2) Prior to the update being published, a systematic review by Shaneyfelt et al10 identified over 104 tools used to evaluate EBP teaching strategies for the field of medicine alone, with varying levels of validity and reliability.

 

It is generally agreed that the five basic steps of the EBP process include: asking a structured clinical question; collecting the best evidence available; critically appraising the evidence to ensure validity, relevance and applicability; applying or integrating the results into practice and evaluating outcomes.3,8,11,12 The teaching of these steps to undergraduates from a generic foundation has potential applicability across health fields13-15; however, studies available to date reflect a primarily segregated approach to the EBP process across health disciplines.16 While the content retrieved and the context for implementing the evidence will differ according to a students' clinical field, the process for searching and appraising the evidence will be similar when based on the five steps outlined above. There is some literature emerging which supports EBP education for undergraduate students to embrace an interprofessional context.17

 

Despite growing support for EBP, implementing the best evidence in practice is still suboptimal, with reports that some patients are still subject to inadequate and potentially unsafe care.18-21 Concomitant to this problem is the suggestion that some treatments are being implemented too early, without rigorous analysis of the research being implemented.18 The problem of translating knowledge into practice has become a focal point for many clinicians and researchers, and frameworks to address barriers for best practice present a multifaceted and complex phenomenon.21 One of the main tenets within knowledge translation frameworks is that of continuing education, with acknowledgement that this requires modification according to the target audience, e.g. health professional, health care service provider, policy maker and/or the patient.21 Tertiary educators are in a unique position to create a positive attitude toward EBP among students in the early years through teaching fundamental principles of EBP and building on them appropriately14, so that students can continue to develop these skills after graduation; however evidence to support strategies specifically for undergraduates is required.

 

A systematic review specifically on the effectiveness of strategies for teaching EBP to undergraduate students across health disciplines has not been reported to date, however systematic reviews on teaching all or part of the EBP process to postgraduate medical students and combinations of student and clinician populations have been conducted.22-28 A systematic review conducted by Coomarasamy & Khan compared stand-alone teaching methods to integrated clinical problem solving strategies in postgraduate medical students.22,23 Results identified that clinically integrated courses provided improvements in knowledge, skills, attitudes, and behaviors; however the population for this review was postgraduate medical students. A very recently published review of systematic reviews by Young, Rohwer, Volmink, and Clarke attempted to synthesize the available information on EBP education and reviewed 15 earlier published systematic reviews and one unpublished one, from 1993 to 2013.28 Each of the included reviews evaluated single and/or multi-faceted educational interventions aimed at improving various EBP outcomes including, but not limited to, knowledge, critical appraisal skills, attitudes, and EBM behaviors among a mixture of populations including undergraduate and postgraduate students as well as health professionals from medicine, nursing and allied health fields. Recommendations from the findings suggested that teaching strategies should take into account individual student factors such as learning style and capability as well as external organisational factors such as the setting of the learning activity and delivery format. The review suggested that a variety of EBP methods (e.g. journal clubs, small group discussions, incorporating clinical scenarios, lectures) had the greatest effect on improving critical appraisal skills and EBP behaviour and knowledge. Small sample sizes and heterogeneity of study aims, outcomes and samples were methodological limitations of the included studies and were highlighted by the authors as preventing meta-analysis. Although this review provides some guidance for teaching strategies of EBP to some populations, it does not provide specific guidance for the differing needs of students as opposed to health professionals.28

 

The focus of research on promoting EBP has been on changing health professionals' behavior to ensure their practice is based on evidence. The eventual reduction of such a focus to one of determining effective teaching strategies on the fundamentals of EBP behaviors for undergraduate students would enable development of positive EBP behavior. Proponents of the EBP process suggest the process is a lifelong learning journey,14,29 therefore starting in the undergraduate years would provide a solid basis from which learning can continue to develop. Teaching EBP to undergraduates is complex and challenging and consequently a strong evidence base is required to assist educators to better prepare students for professional evidence-based practice.

 

Definitions: Evidence-based practice: a process by which decisions on patient care are made utilizing best available evidence for the context in which care is being delivered and incorporating patient considerations and clinical expertise.30,31

 

Evidence-based medicine: the practice of EBP as applied to the field of medicine. Based on the widely cited definition by Sackett et al29,31 whereby EBM was defined as:

 

"[horizontal ellipsis]the conscientious, explicit, and judicious use of current best evidence in making decisions about the care of individual patients. The practice of evidence based medicine means integrating individual clinical expertise with the best available external clinical evidence from systematic research. By individual clinical expertise we mean the proficiency and judgment that individual clinicians acquire through clinical experience and clinical practice."29(p.71)

 

Evidence-based healthcare: Incorporation of the process outlined above, throughout all health fields including nursing, medicine, allied health, etc.30

 

Evidence-Based Practice behavior - refers to what the student actually does in clinical practice, which may be in a simulated scenario or as part of a clinical unit. It incorporates all the processes within EBP including incorporating patient preferences and the student's own level of competence to answer a specific clinical question. Behavior demonstrates how knowledge and skills are applied.8 Assessment of behaviour helps to uncover incongruence between what the student thinks they should or can do and what they actually do.9

 

Evidence-Based Practice skills - refers to the tasks within the process of EBP, e.g. searching for evidence, appraising the evidence. Evidence-based practice skills can be measured using an appropriate tool.9

 

Article Content

Inclusion criteria

Types of participants

This review will consider studies that include undergraduate health students from any undergraduate health discipline, including but not limited to medicine, nursing and allied health. Post graduate and post-registration students will not be included.

 

Types of interventions

This review will consider studies that evaluate strategies or interventions aimed at teaching any or all of the five steps of evidence-based practice, namely asking a structured clinical question; collecting the best evidence available; critically appraising the evidence to ensure validity, relevance and applicability; applying or integrating the results into clinical practice, and evaluating outcomes. The strategy may take place solely within a tertiary education environment or may be combined with a clinical setting.

 

Types of outcomes

This review will consider studies that include the following outcome measures: EBP behavior, knowledge, skills, attitudes, self-efficacy (or self-confidence), beliefs, values, intention to use EBP (future use) and confidence levels. Tools used to measure these outcomes will be assessed for reported validity, reliability and generalizability. Outcomes will be measured during the student's education period up to graduation. If studies are conducted across different year levels this will be taken into account during analysis and reported accordingly.

 

Types of studies

This review will consider for inclusion the following study designs: randomized controlled trials, non-randomized controlled trials, quasi-experimental and before and after studies Prospective and retrospective cohort studies, case control studies and analytical cross sectional studies will also be considered for inclusion. This review will also consider descriptive epidemiological study designs including case series and descriptive cross sectional studies.

 

Search strategy

The search strategy aims to find both published and unpublished studies. A three-step search strategy will be utilized in this review. An initial limited search of PUBMED and CINAHL will be undertaken followed by analysis of the text words contained in the title and abstract, and of the index terms used to describe article. A second search using all identified keywords and index terms will then be undertaken across all included databases. Thirdly, the reference list of all identified reports and articles will be searched for additional studies. Due to unavailability for translation resources, studies published in English only will be considered for inclusion in this review. Taking into consideration the Sicily Statement recommendations were published in 2005, this review will consider studies from 2005 to 2014.

 

The databases to be searched include:

 

PUBMED

 

Cinahl

 

Scopus

 

ProQuest Health

 

ERIC

 

PsychInfo

 

Science Direct

 

Database of Reviews of Effectiveness.

 

The search for unpublished studies will include:

 

Trove

 

Google Scholar

 

Conference proceedings

 

The New York Academy of Medicine

 

Mednar.

 

Initial keywords to be used will be:

 

Evidence-based practice

 

Evidence-based education

 

Evidence-based medicine

 

Undergraduate

 

Baccalaureate

 

Critical appraisal

 

Knowledge

 

Skills

 

Attitudes

 

Self-efficacy

 

Teaching

 

Education, Nursing

 

Education, Medicine

 

Occupational therapy

 

Allied health

 

Speech Pathologist

 

Social Work

 

Physical therapist

 

Evidence-based dentistry

 

Psychology

 

Journal club.

 

A previous systematic review on health professional behaviour used a comprehensive list of terminology for health professionals and will be used as a guide for this aspect of the review.32 A specialist health librarian will also assist with search terms.

 

Assessment of methodological quality

Papers selected for retrieval will be assessed by two independent reviewers for methodological validity prior to inclusion in the review using standardized critical appraisal instruments from the Joanna Briggs Institute Meta-Analysis of Statistics Assessment and Review Instrument (JBI-MAStARI) (Appendix 1). Any disagreements that arise between the reviewers will be resolved through discussion, or with a third reviewer.

 

Data collection

Data will be extracted from papers included in the review using the standardized data extraction tool from JBI-MAStARI (Appendix II). The data extracted will include specific details about the interventions, populations, study methods and outcomes of significance to the review question and specific objectives. As this review forms part of a PhD study, the first author will perform the data extraction and will do double extraction to ensure accuracy. If required, primary authors will be contacted for further information.

 

Data synthesis

Quantitative data will, where possible be pooled in statistical meta-analysis using JBI-MAStARI. All results will be subject to double data entry. Effect sizes expressed as odds ratio (for categorical data) and weighted mean differences (for continuous data) and their 95% confidence intervals will be calculated for analysis. Heterogeneity will be assessed statistically using the standard Chi-square and also explored using subgroup analyses based on the different study designs included in this review. If Meta-analysis is possible, I2 calculations will be used to determine the extent of heterogeneity. Where statistical pooling is not possible the findings will be presented in narrative form including tables and figures to aid in data presentation where appropriate.

 

Conflicts of interest

The authors declare no known conflicts of interest.

 

References

 

1. Levin R, Feldman HR, Eds, Teaching Evidence-Based Practice in Nursing. New York: Springer Publishing Company; 2006. [Context Link]

 

2. Ciliska D. Educating for evidence-based practice. J Prof Nurs. 2005; 21(6):345-50. [Context Link]

 

3. Burns HK & Foley SM. Building a foundation for an evidence-based approach to practice: teaching basic concepts to undergraduate freshman students. J Prof Nurs. 2005; 21(6): 351-357 [Context Link]

 

4. Cordiner M & Davis LM. Changing academic practice to prepare nursing undergraduates for an evidence-based approach. Research and Development in Higher Education: Reshaping Higher Education. 2010; 33(1): 179-190 [Context Link]

 

5. Nursing and Midwifery Board of Australia. National Competency Standards for the Registered Nurse. Nursing and Midwifery board of Australia: Melbourne.2006 [Context Link]

 

6. Ryecroft-Malone J. The politics of the evidence based movements, legacies and current challenges. Journal of Research in Nursing.2006; 11(2): 95-108. doi: 10.1177/1744987106059793 [Context Link]

 

7. Ryecroft-Malone J, Seers, K, Tichen A, Harvey G, Kitson A, & McCormack B. What counts as evidence in evidence-based practice? J Adv Nurs. 2004;47(1): 81-90 [Context Link]

 

8. Dawes M, Summerskill W, Glasziou P, Cartabellotta A, Martin, J, Hopayian K, Porzsolt F, Burls A, Osborne J. Sicily statement on evidence-based practice. BMC Med Educ. 2005; 5(1):1 [Context Link]

 

9. Tilson, J, Kaplan SL, Harris JL, Hutchinson A, Ilic D, Niederman R, Potomkova J, Zwolsman SE Sicily statement on the classification and development of evidence based practice learning assessment tools. BMC Med Educ. 2011; 11(1): 78 [Context Link]

 

10. Shaneyfelt T, Baum K, Bell D, Feldstein D, Houston T, Kaatz S, Whelan C, Green M. Instruments for evaluating education in evidence-based practice: a systematic review. JAMA. 2006;296:1116-27. PubMed PMID: doi:10.1001/jama.296.9.1116 [Context Link]

 

11. Fineout-Overholt E. Synthesizing the evidence: How far can your confidence meter take you?. AACN Advanced Critical Care.2008; 19(3): 335-339 [Context Link]

 

12. Sackett DL & Rosenberg WM. The need for evidence-based medicine. J R Soc Med.1995; 88(11): 620-624 [Context Link]

 

13. Hoffmann T, Bennett S, Del Mar C. Evidence-based practice across the health professions. 2nd Ed. Sydney: Churchill Livingstone Elsevier.; 2013. [Context Link]

 

14. Ilic, D. Teaching evidence-based practice: perspectives from the undergraduate and post-graduate viewpoint. Ann Acad Med Singapore 2009;38(6): 559. [Context Link]

 

15. Chang AM, Crowe L. Validation of Scales Measuring Self-Efficacy and Outcome Expectancy in Evidence-Based Practice. Worldv Evid-Based Nu. 2011;8(2):106-15. [Context Link]

 

16. Satterfield JM, Spring B, Brownson RC, Mullen EJ, Newhouse RP, Walker BB, et al. Toward a Transdisciplinary Model of Evidence-Based Practice. Milbank Q. 2009; 87(2):368-90. [Context Link]

 

17. Melnyk BM. Educational programming in undergraduate and graduate curricula: Friend or foe to accelerating evidence-based practice? Worldv Evid-Based Nu. 2013;10(4):185-186 [Context Link]

 

18. Graham ID, Logan J, Harrison MB, Straus, SE, Tetroe J, Caswell W, Robinson N. Lost in knowledge translation: time for a map? J Contin Educ Health Prof. 2006;26(1):13-24 [Context Link]

 

19. Grimshaw JM, Eccles MP, Lavis JN, Hill SJ, Squires JE. Knowledge translation of research findings. Implement Sci.2012; 7(1):50 [Context Link]

 

20. Grol R & Wensing M. What drives change? Barriers to and incentives for achieving evidence-based practice. Med J Aust. 2004;180(6 Suppl):S57-60 [Context Link]

 

21. Straus S, Tetroe J, Graham D. Knowledge translation in health care: moving from evidence to practice. West Sussex. John Wiley & Sons. 2013. [Context Link]

 

22. Coomarasamy A. & Khan K. What is the evidence that postgraduate teaching in evidence based medicine changes anything? A systematic review. BMJ. 2004;329(7473):1017. [Context Link]

 

23. Coomarasamy A, Taylor R, Khan K. A systematic review of postgraduate teaching in evidence-based medicine and critical appraisal. Med Teach. 2003;25(1):77-81. [Context Link]

 

24. Flores-Mateo G. & Argimon J. Evidence based practice in postgraduate healthcare education: A systematic review. BMC Health Serv Res. 2007;7(1):119. [Context Link]

 

25. Norman G. & Shannon S. Effectiveness of instruction in critical appraisal (evidence-based medicine) skills: a critical appraisal. CMAJ.1998;158(2):177-181. [Context Link]

 

26. Horsley T, Hyde C, Santesso N, Parkes J, Milne R, Stewart, R. Teaching critical appraisal skills in healthcare settings (Review). Cochrane DB Syst Rev. 2011.Nov 9;(11):CD001270. doi: 10.1002/14651858.CD001270.pub2. [Context Link]

 

27. Taylor R, Reeves B, Ewings P, Binns S, Keast J, Mears R. A systematic review of the effectiveness of critical appraisal skills training for clinicians. Med Educ. 34; (2): 120-125. [Context Link]

 

28. Young T, Rohwer A. Volmink J. Clarke M. What are the effects of teaching evidence-based health care (EBHC)? Overview of systematic reviews. PLoS ONE.2014; 9(1): e86706. [Context Link]

 

29. Sackett, D. L. and W. M. Rosenberg The need for evidence-based medicine. J Roy Soc Med 1995; 88(11):620-624. [Context Link]

 

30. Pearson A, Wiechula, R, Lockwood, C. The JBI model of evidence-based healthcare. International Journal of Evidence-Based Healthcare.2005; 3(8): 207-215. [Context Link]

 

31. Sackett DL, Rosenberg WM, Gray, J, Haynes B, Richardson WS. Evidence based medicine: what it is and what it isn't. BMJ.1996; 312(7023):71. [Context Link]

 

32. Eccles MP, Hrisos S, Francis J, Kaner EF, Dickinson HO, Beyer F, et al. Do self-reported intentions predict clinicians' behaviour: a systematic review. Implement Sci 2006; 1(1):28.ea [Context Link]

Appendix I: Appraisal instruments

MAStARI appraisal instrument[Context Link]

Appendix II: Data extraction instruments

MAStARI data extraction instrument[Context Link]

 

Keywords: evidence-based practice; systematic review; undergraduate health education; evidence-based practice process; interdisciplinary