Back to Journals » Advances in Medical Education and Practice » Volume 14

Utilizing Evaluation and Development Frameworks to Engineer a College-Wide Evaluation and Reform of an Undergraduate Dental Curriculum

Authors Alwadei A 

Received 31 December 2022

Accepted for publication 17 February 2023

Published 28 February 2023 Volume 2023:14 Pages 145—156

DOI https://doi.org/10.2147/AMEP.S402059

Checked for plagiarism Yes

Review by Single anonymous peer review

Peer reviewer comments 3

Editor who approved publication: Prof. Dr. Balakrishnan Nair



Abdurahman Alwadei

Department of Pediatric Dentistry and Orthodontics, College of Dentistry, King Saud University, Riyadh, Saudi Arabia

Correspondence: Abdurahman Alwadei, King Saud University, College of Dentistry, Department of Pediatric Dentistry and Orthodontics, Building #23, 1st Floor, Room 17, P.O. Appendix 60169, Riyadh, 11545, Kingdom of Saudi Arabia, Email [email protected]

Purpose: To operationalize and analyze a college-wide evaluation of an undergraduate dental curriculum.
Materials and Methods: A descriptive case study design was used with extensive multiple data collection methods that included literature review, document review of existing data, survey questionnaires, focus group semi-structured interviews and observation of clinical and laboratory tasks. This approach was based on Kern’s curriculum development model and Fitzpatrick’s practical guidelines and evaluation standards.
Results: The evaluation outcomes indicated that a significant curricular change is needed. In hindsight, a thorough reflection on the evaluation strategy is provided highlighting several contextual factors. Actionable recommendations and comparisons are also drafted to shape a coherent curriculum reform implementation.
Conclusion: The process by which the evaluation was conducted, and the reform implementation is being instituted, while unique to this college, may offer insights for change at other dental colleges. In that, greater emphasis is placed on the general principles that remain applicable to other comparable contexts regardless of the distinctiveness in specificities.

Keywords: curriculum evaluation, curricular change, curriculum, dental curriculum, dental education

Introduction

Background

A contemporary dental curriculum should reflect, and adopt to, innovation and changes in dental education. And such perspective is realized by leaders from various dental schools who consider continuous reflection, improvement, and periodic review cycle of the curriculum to be an essential measure to equip students with the necessary tools to thrive in the 21st century and respond to a fast-changing world.1 Toward this end, others have published their experiences in conducting curriculum evaluation and implementing curriculum reform.1–7 Those shared experiences present a valuable resource for reflection on how to bring about a curriculum reform implementation. However, they highlight contextual challenges associated with initiating curriculum change in the most valid and effective approach.

Although each individual institution reported a different trajectory of reform, similar themes can be deduced under the umbrella of systemic changes in dental educational models. The similarities emphasize comprehensive and patient-centered care provided through a reorganized curriculum that integrates basic and clinical sciences and adopts a competency- and community-based education as well as an instructional approach that fosters deep learning of students via curiosity, autonomy, independence, and responsibility.1 On the other hand, curriculum evaluation with/without reform entails aspects of institutional uniqueness that reflects a broader complexity about its implementation. In fact, findings from education literature explicitly highlight the gap in curriculum reform implementation where there is a malalignment between intention and realization.8

The impact of published experiences for curriculum reform is essential to rationalize and realize. However, in the context of global dental education, there is scarcity of comprehensive analysis of curriculum reform implementation from a practical and universal perspective. Through the lens of multiple evaluation and implementation frameworks,9–12 this paper aims to add to health professions education (HPE) literature by documenting how to initiate institutional change in an operational way and provide insights into the “change readiness” of dental colleges to proceed with curricular developments and innovations. Specifically, it underlines the importance of several key dimensions associated with planning and implementing a large-scale curriculum evaluation and how to translate these dimensions into tangible actions.

Context

An overview of the existing/evaluated curriculum will provide a clearer context to better understand the evaluation process and outcome. The existing curriculum consists of three phases over a six-year period, leading to a Bachelor of Dental Science degree (BDS), at King Saud University (KSU), Saudi Arabia. The first phase for all students is the premedical program year and permits the completion of prerequisite biomedical sciences for the dental curriculum and initiation of the general education requirements of the university. The second phase consists of the preclinical portion of the dental curriculum, which lasts for two years and includes the remaining general education courses. The curriculum is based on a traditional lecture-based format and preclinical laboratory experiences where students are exposed to the techniques of dentistry in traditional laboratory courses. The third phase consists of the clinical portion of the dental curriculum and lasts three years. The primary method of learning is by providing a required number of procedures to patients overseen by specialty faculty. Course assessments employ a variety of traditional formats including multiple-choice questions, essays, oral exams, and practical exams.

Aim

The evaluation process implemented at King Saud University-College of Dentistry (KSU-COD) was aimed to review the written, taught, and assessed college-wide curriculum to determine:

  1. The effectiveness of each department in accomplishing its original goals and objectives.
  2. The alignment with global, national, local standards and requirements.
  3. Students’ performance in national and international licensing exams.
  4. That the curriculum promotes active student engagement and meets diverse learning needs.
  5. The incorporation of research-based best instructional practices and strategies.
  6. That instructional strategies and assessment methods are aligned with the curriculum and current trends in dental education.
  7. The satisfaction of stakeholders.

Materials and Methods

Before describing the methodology taken to evaluate the KSU-BDS program, it is essential to first chronicle the deliberate actions and processes taken by KSU-COD leaders which can be described via Kotter’s framework in transforming organizations.9 These actions are discussed more in-depth in the following text.

Pre-Evaluation Deliberative Processes

Problem Identification (Creating a Sense of Urgency)

The KSU-COD is well positioned as a leader in the region. The KSU-COD newly elected administration wanted to further enhance the institution’s position and reputation by attaining the Commission on Dental Accreditation (CODA). Therefore, the KSU-COD leaders decided to conduct a major curriculum evaluation including signing a formal contract with an external education consulting agency to guide and aid in the curriculum evaluation process. During their site visit, the external evaluators who have extensive knowledge and experience in program development and evaluation generated a detailed report. The gist of their report highlighted that the KSU-COD lacked an explicit plan in place for systematic ongoing review of the curriculum based on benchmarking CODA standards and outcomes assessment. The report also highlights the lack of basic teaching philosophy agreed upon collectively among the faculty and administration.

As a result, the administration, faculty, students, and alumni expressed concerns of such trajectory and stated desire to maintain the college as a preeminent dental program from an academic and scholarly perspective. The faculty and students have recognized the urgency to change, including the newly elected administration whose members were determined to deliver the COD updated strategic plan and vision (Box 1).

Box 1 Modified Strategic Plan, Vision, Mission, and Core Values for the College of Dentistry, King Saud University

Forming a Powerful Guiding Coalition and Creating a Vision with Strategies

To create buy-in, maintain institutional vitality and embrace diverse perspectives, leaders from the new administration created a Dental Education Division (DED) and hired five full-time junior faculty members to pursue postgraduate master’s and/or doctorate degrees in HPE. They also designated a Quality Control Committee (QCC) led by the Dean himself and included domestic external experts in HPE as well as diverse body of faculty, students and staff members representing each division. The committee has achieved the following:

  1. Revised and modified on the institution’s mission, vision, core values strategic priorities, and promotional policies (Box 1) that reward teaching excellence for the institution to be consistent with the apparent change in the institution culture that is influenced by predominantly junior faculty members, well-rounded and vocal students, and supportive leadership.
  2. Identified and created new set of competencies (Table 1) that are in parallel with the American Dental Education Association updated recommendations on best practice as well as the CODA requirements that have been considered, through consensus, to be the main goal for the new administration to meet the modified mission and vision of COD.

Table 1 Data Collection Methods (Environmental Scan)

Communicating the Vision and Empowering Others

To ensure alignment of support systems with the new vision, the new administration formed two committees to ensure effective communication, college-wide curriculum evaluation and attainment of change effort.

  1. The Curriculum Oversight Committee (COC) to establish ongoing program evaluation procedures, report on opportunities for improvement of curriculum, and suggest maintenance, revisions, deletions, and/or creation of new curricula as needed.
  2. The Faculty Development Programs Committee (FDPC) to develop necessary workshops for faculty, student, administrators, and staff members. The workshops were planned to be designed based on the identified needs and recommendations produced by the COC during and after completion of the curriculum evaluation process (ie, consolidating improvements and producing additional change).9

Evaluation of the Existing Curriculum

As mentioned earlier, the QCC signed a formal contract with a well-established consulting agency with experts in HPE from the United States, United Kingdom and Canada and commissioned the COC to conduct the evaluation. The COC was led by the director of DED and consisted of newly appointed faculty members at the DED, departments’ chairs, and administrator personnel, with periodic involvement of student representatives from each academic year. It also included two external evaluators from the consulting agency.

Procedure (Design and Data Collection)

A descriptive case study design was used with extensive multiple data collection methods that included literature review, document review of existing data, survey, focus group semi-structured interviews and observation of clinical and laboratory tasks (Table 1). Collection of qualitative data included 1) content analysis of several documents such as the evaluated curriculum, curricula of several international dental schools (Table 1), previous internal evaluation documents (open-ended responses within course and college-level satisfaction surveys etc.), and the outcomes of ad hoc reviews and previous national and international accreditations, 2) focus group interviews with multiple stakeholders (Appendix 1), and 3) field notes during real-time observation of clinical and laboratory work. Quantitative data included aggregate data of students and teachers’ feedback via Likert-scale questions of course evaluation surveys, American Dental Association’s (ADA) Survey of Dental Education and online survey questionnaires specific to KSU-BDS curriculum evaluation Appendices 25. The ADA survey, published in alternate years, reports the time devoted to the instruction, laboratory, and patient care activities from 59 United States dental schools in 3 major teaching areas: biomedical science; dental/clinical science; and behavioral, social, information, and research sciences. The overall data collection approach relied on multiple sources of evidence which provided a convergence of findings and allowed for a holistic view of the complexity of a college-wide curriculum evaluation.

While resource intensive, this descriptive in-depth design included different data collection methods, each with specific targeted participants and served specific purpose to detect the needs, interest, and responsibilities of stakeholders. Given the broad scope of evaluation purpose, it was challenging to concurrently address the various needs of stakeholders while maintaining a cost-effective evaluation design. Thus, it was important to implement a flexible, gradual, and needs-specific evaluation design. Such sequential implementation allowed the evaluators to address the purpose, questions and context while prioritizing a valid and practical evaluation approach. With a flexible evaluation design, it allowed for timely response to any emergent needs to be addressed as continuous refinement and reinforcement were expected and needed through the evaluation process.11

Evaluation Model

The model developed for the purpose of the evaluation of the BDS curriculum (Figure 1) allowed for the collection of formative (process and outcome) as well as summative (outcome) data.13 Formative data was used to monitor the process of curricular change, to suggest and support additional changes to the curriculum, and to help understand what was done to achieve program outcomes by identifying gaps between curriculum outcomes and implementation objectives.14–16 Furthermore, process evaluation data provided a context for interpreting the findings of the outcome and impact evaluation.17 On the other hand, formative outcome evaluation data primarily served to answer the question – To what extent were the outcomes objectives of the BDS curriculum were achieved?18,19 All formative data were timely, concrete, and useful. Findings were communicated to department chairs, course directors, faculty, and students on a regular basis.

Figure 1 Undergraduate dental program evaluation model at King Saud University.

Summative evaluation data guided course directors to make decision about the overall merit (or worth) of their courses and to assess the achievement of outcome objectives.11,15,20 When considering dissemination of the reporting, these data were used, for example, to decide upon the generalizability of curricular changes, the need for further restructuring of the curriculum, and/or the allocation of resources.16,20 Since the attainment of accreditation was the driving force for the change initiative, summative data is essential, as external evaluators will use it for accreditation purposes.

Evaluation Approach

Three approaches surfaced through the evaluation process in a complementary manner. During the initiation phase, the Management-Oriented Approach (MOA) served decision makers (QCC and COC), which was crucial since the focus of the evaluation was to reflect the vision, mission, and core values of the new administration and to adhere to the updated BDS program competencies and CODA standards. In reference to attainment of CODA standards as another focus for the evaluation, expertise-oriented approach was important to use during specific stages of the evaluation. Also, the external evaluators helped in avoiding a common pitfall associated with MOA for which preference is usually given to top management.11

Throughout the process, Participant-Oriented Approach (POA) was an overriding approach as COC members realized that utilization of evaluation results was significantly improved when there is greater collaboration with other stakeholders. During the evaluation, COC members consistently interacted with multiple stakeholders to obtain high-quality information, provide perspectives of important goals and objectives, and identify existing gaps in processes and procedures that were sometimes missed by evaluators.11 POA was critical in increasing stakeholders’ sense of ownership and augmenting their understanding of evaluation, laying the foundation for an environment of ongoing program reflection and improvement. Although, time-consuming and labor intensive, this approach had the advantage of gaining support of key stakeholders to respond to recommendations for change.

Results

Based on the environmental scan described in Table 1, general and targeted needs of relevant stakeholders were identified (Box 2). The COC recommended that significant curricular changes are needed. The nature of such recommendation was sensitive regarding the success or lack thereof of some courses, evaluations of specific faculty and course directors, and recommendations to modify the existing curriculum by potential elimination and integration of courses. From a political perspective, the external evaluators help alleviate conflict of interest, provide constructive feedback when needed, and increase the objectivity in the presentation of the findings from the evaluation given their qualifications and extensive experience with transformational curricular evaluation and reform projects.

Box 2 General and Targeted Needs Assessment Outcomes

Discussion

This paper describes the anatomy and functions of a viable curriculum evaluation and reform implementation strategy characterized by the synchronized promotion of institutional accountability and individual agency with insights related to the internal dynamics of the top-down–bottom-up approach, proactive strategies conducive to institutional development, and the interplay between various global, cultural, political, societal, and contextual factors within a college reform.21 Upon reflection on the development and implementation of the KSU-BDS program evaluation strategy, it can be synthesized according to several key principles including collaborative, regulated, reflective, timely and student-centered, reliable and valid, continuous, and based on professional standards. These principles are described next in greater details.

The evaluation strategy of the BDS curriculum, which was executed by new administration, QCC and COC, took a collaborative approach. The evaluation had been, and will continue to be, a negotiated process.15,22,23 It was characterized by a significant degree of collaboration among key stakeholders including administration, faculty, and students as well as consideration of patients and community, in both its development and implementation.24 Because key stakeholders shared responsibility and decision-making, the evaluation was responsive to the needs of the BDS curriculum as well as those of stakeholders.15 By following a collaborative approach, notable promising initial results were noted as reflected by the increased stakeholders’ cooperation and involvement in the evaluation and receptivity to the findings. Such results are considered short-term wins,9 and such supportive culture shall be fostered to serve in building ongoing evaluation capacity within the COD.

As emphasized earlier, stakeholders’ involvement and engagement are essential to the success of the implementation process, especially in a complex environment such as the one described in this paper. The early steps of the envisioned curriculum, starting with the updated KSU-COD vision, mission, and strategy plan, were co-constructed with stakeholders, which was essential in building public support, aligning interests, and supporting individual and collective sense-making.21 Early and continuous faculty development workshops were designed and conducted for faculty members to provide them with the competences needed to understand, adapt, and shape the curriculum change.25

The evaluation strategy involved a shift from a deregulated evaluation system previously administered by individual departments and courses to a more regulated system achieved through the creation of DED (ie, institutionalizing a new approach).9 It should be noted that the evaluation of the BDS program is the responsibility of the COC, which reports directly to the QCC. The regulation of the evaluation process facilitated the overall assessment of the undergraduate curriculum as well as planned curricular change in the future.14,26 This approach highlights greater awareness toward aligning accountability mechanisms to the curriculum reform which is conducive to various stakeholders by allowing them to adjust their practices and focus on the implementation of the new curriculum. Furthermore, the evaluation was designed to promote reflective practice. As part of the reflective process, a new policy was proposed which requires course directors to respond to student feedback. Hence, future evaluations will be central to curricular change and ongoing curriculum development.23,27

The importance of acknowledging and responding to feedback in a timely fashion was recognized by the evaluation approach which was predominantly POA. The evaluation strategy was characterized by considerable involvement from teachers and students. As such, it facilitated curricular improvement and student learning through the integration of the existing curriculum evaluation and future change processes.23 Through an autonomy-centered evaluation process, the fundamental role of faculty members in curriculum development and implementation was recognized.28 Specifically, faculty members’ feedback regarding curriculum and instruction is particularly relevant given their frequent interaction with students and understanding of students’ needs.25

Students and teachers were, and will continue to be, actively involved in the ongoing evaluation and monitoring of courses and clinical rotations. They were encouraged to express their opinions and to provide feedback on content and pedagogical strategies as well as to make suggestions for improving the exchange of information. Currently, course directors, with the support of DED, are pilot testing rubrics and work-based assessment tools to be used for direct observation formative and summative assessment when students perform psychometric clinical and laboratory tasks.29,30 It is also an example for short-term win of the assessment-based curricular changes.9 Such change was the result of an appropriate and prioritized response to students’ timely feedback. This will allow students to “witness changes to a course as they experience it, rather than moving on without ever knowing whether their recommendations had any affect”.27

To ensure the reliability and validity of the findings of the evaluation, data and methodological triangulation were used.31,32 Data were examined from different sources and over time and a combination of qualitative and quantitative methods were used. In addition, all evaluation instruments were designed in consultation with key stakeholders and external evaluators. Summary reports were presented to key stakeholders to validate the findings. The external evaluators were given high authority in the process, which ensured credibility by reducing biases as the QCC (led by the Dean) commissioned COC for review and reporting. According to Fitzpatrick et al, the greatest source of bias is when the evaluator is employed by the organization whose program is being evaluated.11 The external evaluators accompanied two members of COC during the direct observations and solely conducted the focus group interviews. This allowed stakeholders to express their concerns freely. In addition to improving the evaluation outcomes, adopting an expertise-oriented approach provided political legitimacy and effective implementation of timely (early and continuous) stakeholder engagement.33

Using meta-evaluation checklists recommended by the external evaluators, the evaluation was monitored on an ongoing basis by the COC and QCC to ensure that (1) the design was feasible; (2) activities were completed as planned and in a timely manner; and (3) instruments and products (data and reports) were of high quality.11,16 The checklist comprised 49 items arranged under six overarching themes including conceptualization of the evaluation, sociopolitical factors, contractual/legal arrangements, technical design, management plan, and moral/ethical/utility questions.34 The evaluation was modified as needed and as appropriate.

With regard to the revised curriculum, the following section provides an evidence-based and contextually appropriate perspective in attempt to guide future processes and decisions in implementing the revised curriculum. Guided by what have been described so far, Table 2 presents a comparison of the curriculum features between the existing/evaluated curriculum and proposed/revised curriculum. Through this comparison, and under the umbrella of guiding principles for the revised curriculum, an updated pedagogical paradigm is drafted that includes perspectives on content, instructional strategy, assessment methods as well as cultural and behavioral aspects. These foundational changes are important in moving the curriculum toward embracing the need for contemporary, empirically sound, and culturally sensitive dental care.35

Table 2 Comparison of the Curriculum Features Between the Proposed Curriculum and the Current Curriculum

The goal of the updated pedagogical paradigm is to address deficiencies in the existing curriculum and prepare students to integrate knowledge and skills, develop professional attitude, and seek excellence in providing oral health care for society members. This pedagogical perspective is rooted in reflective critique of the evaluation process and outcome and through the adoption of established concepts in the context of HPE literature with regard to competency-based education, interdisciplinary collaboration, promotion of student-centered learning and implementation of valid, reliable, and feasible assessment methods.36–39 To conclude, the development and implementation of the curriculum will be dependent on a whole-of-system approach with the provision of supportive leadership, adequate resources, effective communication, stakeholders’ consensus and commitment, and faculty development.

The DED will supervise and manage the evolution of the revised curriculum to warrant effective implementation as well as continuous improvement and alignment with the revised program objectives and updated competencies. With the help of the appropriate subcommittees (evaluation and assessment, IT/library resources, and faculty development subcommittees), the COC has created a centralized monitoring system evaluation to aid in monitoring the implementation of the revised curriculum and foster a robust transparent system for the curriculum reform. It depends on indicator frameworks (scores on internal examinations, results of national and international standardized examinations, performance-based assessments including an aggregate view on students’ e-portfolios, graduation questionnaire and exit interviews, course evaluations, and student advancement and graduation rates, etc.) tools (curriculum management system including curriculum mapping ie, the curriculum database) and processes (scholarly activities, ad hoc reviews, teacher training workshops, collaboration between dental and medical education departments at KSU, as well as with independent national and international agencies including the Saudi Commission for Health Specialties and ADA).

The intended KSU-BDS curriculum management approach is a methodical discipline of evaluating curriculum data (program and learning objectives, content, instructional approach including learning activities to session-, course-, and program-level objectives, and assessment) and making curriculum recommendations and/or changes. Utilizing export tools and gap analysis reports, the COC will continuously provide thorough yet targeted reports to the QCC and deanship office identifying any gaps or unforeseen redundancies as the curriculum delivery matures.

Implications and Recommendations

The deliberate actions and processes described in this paper highlight key implications in curriculum evaluation and reform implementation. The nature of this curriculum reform adopted a top-down-bottom-up approach as evident by the early involvement of relevant stakeholders.40 This highlights the importance of having supportive leadership that plays a crucial role in driving and communicating the reform initiative, and more importantly, creates a culture that cultivates changes, and promotes professional learning communities through faculty development programs.41 Indeed, stakeholders’ early involvement, beliefs, individual and collective sense-making, and consensus, are pivotal to ensure successful implementation.42,43

Although KSU-COD leaders engineered the evaluation process and subsequent reform recommendations, it was critical that relevant stakeholders co-constructed the process and perceived it to be autonomy-centered, advancing, and valuable, rather than prescribed, hindering or inconsequential.44 The quality of sense-making influences the stakeholders’ degree of agency toward any curriculum change (ie whether it is embraced, challenged, or rejected).45 However, galvanizing stakeholders’ support requires an effective policy design by developing a coherent implementation strategy that addresses the justification of the reform, clearness and consensus, and theory of change (ie the why, what, and how, respectively). As such, it implies developing a curriculum that intends to achieve clear objectives, according to a robust and attainable theory of change, and to realize a clear and valued educational vision. Furthermore, reform is costly and resource-intensive, and as highlighted in this paper, the implementation strategy should anticipate a high opportunity cost and needs to consider several influential factors such as available resources, existing capacity, and contextual factors.46 Just as important, it needs to develop means to adjust those factors in the short term and plan for the long term to ensure its continuity. It is essential that the COD leaders maintains this continuity since CODA requires participating programs to compile and submit semi-annual resident milestones data, along with annual programmatic information such as faculty surveys and patient-log data. This emphasis on ongoing data collection and trend analysis further highlights the need for building organizational capacity for evaluation into the BDS curriculum by creating the DED. It also ties back into the importance of using the different complementary approaches in the evaluation plan. Finally, educators are encouraged to turn their contextual evaluation experiences into scholarly activities mainly via publishing focused research on discrete aspects of programs’ evaluation such as learners/teachers’ experiences, development and assessment of evaluation instruments, evaluation process, data collection procedures, and critical analysis of evaluation outcomes.

Conclusion

This paper unfolded the implementation of a curriculum evaluation and reform from an operational perspective, translating the abstract into concrete dimensions by contextualizing the principal factors in curriculum renewal and by drawing on real-time practices and examples from the evaluation process to show the utility of this work. The paper also recognized and addressed the unique institutional history, value, culture, and structure that impact the evaluation process and curriculum reform. Nonetheless, greater emphasis was placed on the general principles that remain applicable to other comparable contexts regardless of the distinctiveness in specificities.

Abbreviations

BDS, Bachelor of Dental Science; KSU, King Saud University; COD, college of dentistry; CODA, commission on dental accreditation; DED, dental education division; HPE, health professions education; QCC, quality control committee; COC, curriculum oversight committee; FDPC, faculty development programs committee; MOA, management-oriented approach; POA, participant-oriented approach.

Acknowledgments

The author would like to thank the faculty of King Saud University (KSU), College of Dentistry, for their support and collaboration throughout the evaluation/implementation process. Without their dedication, this work would not have been possible. The author wishes to give a special mention to members of the curriculum oversight committee.

Disclosure

The author reports no conflicts of interest in this work.

References

1. Pyle MA. New models of dental education and curricular change: their potential impact on dental education. J Dent Educ. 2012;76(1):89–97. doi:10.1002/j.0022-0337.2012.76.1.tb05237.x

2. Association ADE. Beyond the Crossroads: Change and Innovation in Dental Education. Washington, DC: American Dental Education Association; 2009.

3. Iacopino AM. The influence of “new science” on dental education: current concepts, trends, and models for the future. J Dent Educ. 2007;71(4):450–462. doi:10.1002/j.0022-0337.2007.71.4.tb04296.x

4. Nadershahi NA, Bender DJ, Beck L, Alexander S. A case study on development of an integrated, multidisciplinary dental curriculum. J Dent Educ. 2013;77(6):679–687. doi:10.1002/j.0022-0337.2013.77.6.tb05519.x

5. Pyle MA, Goldberg JS. Engineering curriculum change at a private midwest school of dental medicine: a faculty innovation. J Dent Educ. 2008;72(3):288–298. doi:10.1002/j.0022-0337.2008.72.3.tb04495.x

6. Ryder MI, Sargent P, Perry D. Evolution and revolution: the curriculum reform process at UCSF. J Dent Educ. 2008;72(12):1516–1530. doi:10.1002/j.0022-0337.2008.72.12.tb04631.x

7. Spallek H, O’Donnell JA, Yoo YIJ. Preparing faculty members for significant curricular revisions in a school of dental medicine. J Dent Educ. 2010;74(3):275–288. doi:10.1002/j.0022-0337.2010.74.3.tb04873.x

8. Pak K, Polikoff MS, Desimone LM, Saldívar García E. The adaptive challenges of curriculum implementation: insights for educational leaders driving standards-based reform. AERA Open. 2020;6(2):2332858420932828. doi:10.1177/2332858420932828

9. Kotter JP. Leading change: why transformation efforts fail. In: Museum Management and Marketing. Routledge; 2007:20–29.

10. Thomas PA, Kern DE, Hughes MT, Tackett SA, Chen BY. Curriculum Development for Medical Education: A Six-Step Approach. JHU press; 2022.

11. Fitzpatrick JL. Program evaluation: alternative approaches and practical guidelines; 2011.

12. Fournier DM. The Program Evaluation Standards: How to Assess Evaluations of Educational Programs. JSTOR; 1994.

13. Tyler R, Gagne R, Scriven M. The methodology of evaluation. In: Perspectives of Curriculum Evaluation Chicago. Vols. 39. IL: Rand McNally; 1967.

14. Gerrity MS, Mahaffy J. Evaluating change in medical school curricula: how did we know where we were going? Acad Med. 1998;73(9 Suppl):S55–S59. doi:10.1097/00001888-199809001-00010

15. O’Sullivan RG. Practicing Evaluation: A Collaborative Approach. Sage; 2004.

16. Scriven M. Evaluation Thesaurus. 4th ed. ed, Sage Publications, Newbury Park, Calif., ©1991; 1991.

17. Issel LM, Wells R, Williams M. Health Program Planning and Evaluation: A Practical Systematic Approach to Community Health. Jones & Bartlett Learning; 2021.

18. Nestel D. Development of an evaluation model for an introductory module on social medicine. Assess Eval High Educ. 2002;27(4):301–308. doi:10.1080/0260293022000001328

19. Patton MQ. Utilization-Focused Evaluation. Sage publications; 2008.

20. Rossi PH, Freeman HE, Lipsey MW. Evaluation: A Systematic Approach. 6th ed. ed, Sage Publications, Thousand Oaks, Calif., ©1999; 1999.

21. Gouëdard P, Pont B, Hyttinen S, Huang P. Curriculum reform: a literature review to support effective implementation; 2020;.

22. Guba EG, Lincoln YS. Fourth Generation Evaluation. Sage; 1989.

23. Louie B, Byrne N, Wasylenki D. From feedback to reciprocity: developing a student-centered approach to course evaluation. Eval Health Prof. 1996;19(2):231–242. doi:10.1177/016327879601900206

24. Cousins JB, Donohue JJ, Bloom GA. Collaborative evaluation in North America: evaluators’ self-reported opinions, practices and consequences. Eval Pract. 1996;17(3):207–226. doi:10.1177/109821409601700302

25. McLean M, Cilliers F, Van Wyk JM. Faculty development: yesterday, today and tomorrow. Med Teach. 2008;30(6):555–584. doi:10.1080/01421590802109834

26. Nieveen N, Kuiper W. Balancing curriculum freedom and regulation in the Netherlands. Eur Educ Res J. 2012;11(3):357–368. doi:10.2304/eerj.2012.11.3.357

27. Hendry GD, Cumming RG, Lyon PM, Gordon J. Student-centred course evaluation in a four-year, problem based medical programme: issues in collection and management of feedback. Assess Eval High Educ. 2001;26(4):327–339. doi:10.1080/02602930120063484

28. Van den Akker J, Gravemeijer K, McKenney S, Nieveen N. Educational Design Research. Vol. 2. Routledge London; 2006.

29. Kirton J, Palmer N, Grieveson B, Balmer M. A national evaluation of workplace-based assessment tools (WPBAs) in foundation dental training: a UK study. Effective and useful but do they provide an equitable training experience? Br Dent J. 2013;214(6):305–309. doi:10.1038/sj.bdj.2013.302

30. Manekar VS, Radke SA. Workplace based assessment (WPBA) in dental education-A review. JETHS. 2018;5(2):80–85. doi:10.18231/2393-8005.2018.0016

31. Milburn K, Fraser E, Secker J, Pavis S. Combining methods in health promotion research: some considerations about appropriate use. Health Educ. 1995;54(3):347–356.

32. Whitman NA, Cockayne TW. Evaluating medical school courses: a user-centered handbook; 1984.

33. McGinn N, Welsh T. Decentralization of education: why, when, what and how?; 1999.

34. Stufflebeam DL. Meta-Evaluation. JMDE. 2010;7(15):99–158.

35. Donate‐Bartfield E, Lobb WK, Roucka TM. Teaching culturally sensitive care to dental students: a multidisciplinary approach. J Dent Educ. 2014;78(3):454–464. doi:10.1002/j.0022-0337.2014.78.3.tb05695.x

36. Chuenjitwongsa S, Oliver R, Bullock AD. Competence, competency‐based education, and undergraduate dental education: a discussion paper. Eur J Dent Educ. 2018;22(1):1–8. doi:10.1111/eje.12213

37. Harris N, Bacon CEW. Developing cognitive skills through active learning: a systematic review of health care professions. Athl Train Educ J. 2019;14(2):135–148. doi:10.4085/1402135

38. Williams JC, Baillie S, Rhind S, Warman S, Sandy J, Ireland A. A guide to assessment in dental education; 2015.

39. Yilmaz RBN, Nalbantgil D, Ozdemir F. Impact of discipline‐based vs. interdisciplinary orthodontic courses on dental students’ performance. J Dent Educ. 2019;83(1):64–71. doi:10.21815/JDE.019.008

40. Tyack DB, Cuban L. Tinkering Toward Utopia: A Century of Public School Reform. Harvard University Press; 1995.

41. Thompson SC, Gregg L, Niska JM. Professional learning communities, leadership, and student learning. sssRMLE Online. 2004;28(1):1–15. doi:10.1080/19404476.2004.11658173

42. Fullan M. The New Meaning of Educational Change. Routledge; 2001.

43. Spillane JP, Reiser BJ, Reimer T. Policy implementation and cognition: reframing and refocusing implementation research. Rev Educ Res. 2002;72(3):387–431. doi:10.3102/00346543072003387

44. Pierce JL, Kostova T, Dirks KT. The state of psychological ownership: integrating and extending a century of research. Rev Gen Psychol. 2003;7(1):84–107. doi:10.1037/1089-2680.7.1.84

45. Ketelaar E, Beijaard D, Boshuizen HP, Den Brok PJ. Teachers’ positioning towards an educational innovation in the light of ownership, sense-making and agency. Teach Teach Educ. 2012;28(2):273–282. doi:10.1016/j.tate.2011.10.004

46. Viennet R, Pont B. Education policy implementation: a literature review and proposed framework; 2017.

Creative Commons License © 2023 The Author(s). This work is published and licensed by Dove Medical Press Limited. The full terms of this license are available at https://www.dovepress.com/terms.php and incorporate the Creative Commons Attribution - Non Commercial (unported, v3.0) License. By accessing the work you hereby accept the Terms. Non-commercial uses of the work are permitted without any further permission from Dove Medical Press Limited, provided the work is properly attributed. For permission for commercial use of this work, please see paragraphs 4.2 and 5 of our Terms.