Skip to main content
Wiley - PMC COVID-19 Collection logoLink to Wiley - PMC COVID-19 Collection
. 2022 Nov 20:10.1111/eje.12869. Online ahead of print. doi: 10.1111/eje.12869

Structures, processes and outcomes of objective structured clinical examinations in dental education during the COVID‐19 pandemic: A scoping review

Yetunde Kemi Salawu 1, Derek Stewart 2, Alaa Daud 3,
PMCID: PMC9877700  PMID: 36337030

Abstract

Introduction

Objective structured clinical examinations (OSCEs) are an essential examination tool within undergraduate dental education. Fear of spread of the COVID‐19 virus led to dental institutions exploring alternative means of conducting OSCEs. The aim of this scoping review was to investigate what structures, processes and outcomes of dental OSCEs were reported during the COVID‐19 pandemic.

Materials and Methods

This scoping review was conducted and reported adhering to the Preferred Reporting Items for Systematic Reviews and Meta‐analyses extension for scoping review guidelines (PRISMA‐ScR). Published literature was identified through a systematic search of PubMed, Embase, Cumulative Index to Nursing and Allied Health Literature (CINAHL), Education Resources Information Center (Eric), ProQuest and Google Scholar. Identified articles were independently reviewed by two authors (KS, AD), followed by synthesis in terms of the reported structures, processes and outcomes. Articles reporting cancellation or rescheduling were also included, extracting data on reasons and any suggestions/recommendations.

Results

The search yielded a total of 290 studies of which 239 sources were excluded after removal of duplicates, leaving 51 studies for title and abstract evaluation. Thirty‐four articles were excluded as they did not report on the topic of interest, leaving 17 for full‐text evaluation, of which nine were analysed according to the pre‐set themes. All dental OSCEs taking place (n = 6) were conducted online whilst the remaining (n = 3) were either cancelled or rescheduled. Data on structures reported specific online videoconferencing software used and provision of staff and student training. Processes on the execution of online OSCEs varied significantly from one study to the other, providing rich data on how dental institutions may carry out such assessments tailored to their need. Information regarding outcomes was sparse, as little attention was paid to the results of the students compared to pre‐pandemic, lacking investigation into reliability and validity of online dental OSCEs.

Conclusion

Dental OSCEs could be conducted online implementing well‐planned structures and processes; however, further evidence is needed to prove its reliability and validity based on outcomes. Dental institutions may need to consider alternative methods to assess practical competencies if online OSCEs are to take place.

Keywords: COVID‐19, dental, education, objective structured clinical examination, scoping review

1. INTRODUCTION

Assessment is an exercise designed to measure student's fulfilment of learning outcomes deemed necessary for the completion of their programme. 1 In dental education, assessments evaluate abilities and competencies required for subsequent registration as a dental professional. 2 There are various forms of assessments undertaken throughout the undergraduate dental curriculum, including the objective structured clinical examination (OSCE). OSCEs are universally utilised as an assessment method in dentistry for various reasons. They are considered gold standard as they are able to demonstrate the “shows how” level of clinical competence as described by Miller's Pyramid 3 and assess areas most critical to performance of healthcare professionals such as communication skills, history‐taking and physical examination. 4 They are also a fair method of assessment, 5 removing any patient–examiner interaction that may be experienced in other forms of assessment. 6

Over the years, the scope of OSCE examination has been broadened to satisfactorily complement cognitive knowledge and clinical testing, in order to cover areas such as the ability to handle unpredictable patient behaviour, obtain/interpret data, problem‐solve and educate patients. 7 , 8 , 9 , 10 , 11 Most OSCEs use standardised patients (SP) for accomplishing the latter. SPs are individuals trained to act specific scenarios and display signs and symptoms of certain conditions within a fixed testing environment. 12 , 13 Traditional OSCEs usually take place face to face and comprise a circuit of assessment stations whereby marshals and timekeepers guide students to the next station ensuring time is adhered to accurately. 14

Prior to embarking on developing an OSCE, it is important to have a proper structure in place to ensure that the process, once established, is self‐sustaining. A structure is defined as “an arrangement and organization of interrelated elements in a material object or system”. 15 In assessments, effective structuring requires outlining the regimen clearly, creating a well‐trained assessment team, and defining the appropriate settings in terms of resources, manpower and administrative support. 16 , 17 A process is a “series or set of activities that interact to produce a result”. 18 Within a dental examination, a process entails the course of action executed to achieve the outcome. This includes but not limited to identifying and codifying examination content, assigning students to their stations and observing their performance, following the scoring rubric appropriately and finally, post exam analyses. 19 , 20

At the beginning of 2020, a novel coronavirus, COVID‐19 emerged. The pandemic took countless industries by surprise and resulted in unassailable changes to the way many things had previously taken place. In dental circles, there is a lot of discussion surrounding what may be the “new normal”. 21 It is generally accepted that things will be different for at least the near future, but perhaps even for the distant future. Dentistry is an industry that has had to undergo several changes in order to accommodate the new and advancing regulations owing to the COVID‐19 pandemic. The very nature of dentistry as an occupation is considered as high risk for the transmission of the coronavirus for obvious reasons. 22 Less documented, however, are the changes that undergraduate dentistry training programmes have had to endure.

Along with other university programmes, a great deal of dental educational activity has been moved online during the governmental lockdowns imposed. Moreover, even during periods outside of lockdowns, social distancing is still required to be adhered to. 23 Dentistry along with other university courses moved lectures, tutorials and even assessments to online portals wherever possible. An obvious aperture left consequently unfilled is that of OSCEs.

Due to the nature of OSCE examinations, in their current form they would not comply with COVID‐19 regulations and as such may need to be either removed, delayed or replaced. In order to explore this topic, a scoping review was conducted to ascertain the current status of the published literature. The aim of this scoping review was to investigate structures, processes and outcomes of OSCEs in dental education during the COVID‐19 pandemic.

2. METHOD

A scoping review was used to establish the extent of the existing literature relating to the chosen topic. 24 , 25 Despite having a similar purpose to a systematic review, a scoping review captures a broader range of studies with greater variation in study designs and is appropriate when the search results fall in a body of literature that is diverse in nature. 26 The strength of such method lies in its ability to summarise the existing evidence base and identify any gaps. However, the absence of assessing the quality of primary data presented is considered a limitation.

2.1. Design and search strategy

The review was performed systematically following the Preferred Reporting Items for Systematic Reviews and Meta‐Analyses extension for Scoping Reviews (PRISMA‐ScR) checklist 26 , 27 and Joanna Briggs Institute Reviewers' Manual. 28 A protocol was developed to pre‐define the objectives of the study and methods and allow for transparency of the process. The protocol detailed the criteria that the reviewers intend to use to include and exclude sources of evidence, identified what data were relevant and how the data will be extracted and presented. The search strategy was developed from the research question “what structures, processes and outcomes of OSCEs in dental education have been reported during the COVID‐19 pandemic?”. The following inclusion criteria were applied:

  1. Articles from all countries, published in English, during the COVID‐19 pandemic, related to OSCE assessments in dentistry.

  2. OSCEs performed during the undergraduate programme in dental institutions.

  3. OSCEs cancelled or rescheduled and rational.

  4. A method of assessment comparable to an OSCE in terms blueprinting, learning outcomes assessed and application method.

  5. Qualitative, quantitative and mixed‐methods studies were included.

Any studies reporting an OSCE experience in healthcare professions other than dentistry were excluded.

Key words were identified through titles, abstracts, full text and their references. Full‐text manuscripts were retrieved using institutional access, Google Scholar, Research Gate or by contacting the authors. Discussion papers and letters to the editor were excluded. Databases were searched using the Boolean operators AND and OR combined with truncation (Table 1) and phrase searches. The review was undertaken to identify the published literature and the grey literature in relation to the topic between January 2020 and the end of January 2021 to cover the peak COVID‐19 pandemic period. The following scientific databases were searched: PubMed, Embase, Cumulative Index to Nursing and Allied Health Literature (CINAHL) and educational databases: Education Resources Information Center (Eric) and ProQuest. Additionally, a Google Scholar search was performed. The last date of search was executed on the 31st of January 2021

TABLE 1.

Search words used and truncation

Search word used (Keyword) Truncation options
Dent* Dental, Dentistry, Dentist
Objective Structured Clinical Exam* OSCE, OSCEs
Coronavirus* COVID‐19, COVID 19, SARS2, SARS 2, SARS‐COV‐2
Education No truncation

For PubMed for example

the following strings were combined:

  • CONCEPT 1: OSCE

Key Words

osce*[tw] OR “objective structured clinical exam*” [tw] OR exam*[tw] OR assessment*[tw]

  • CONCEPT 2: Dental

Key Words

dental [tw] OR dentistry [tw] OR dentist*[tw]

  • CONCEPT 3: Education

Key Words

education [tw] OR “Education

Distance”[Mesh] OR “Education

Dental

Graduate”[Mesh] OR “Education

Dental”[Mesh]

  • CONCEPT 4: COVID‐19

Key Words

COVID‐19[tw] OR “COVID 19”[tw] OR SARS2[tw] OR “SARS 2”[tw] OR SARS‐COV‐2[tw] OR Coronavirus[tw] OR “severe acute respiratory syndrome coronavirus 2” [Supplementary Concept]

2.2. Screening process and inclusion criteria

All results obtained were exported to Endnote X9 and duplicates removed. Remaining articles were imported to Rayyan, a web application for screening scoping reviews. The titles and abstracts of 51 papers were scrutinised by AD and KS, who performed the screening independently. Titles were initially reviewed and filtered, excluding clearly irrelevant articles. Abstracts for the remaining articles were retrieved and inclusion/exclusion criteria were applied. Full text of the selected studies was considered for inclusion. The third author (DS) was involved when required to make a final decision. A reference check was conducted and included in the flowchart.

2.3. Data extraction and synthesis

Identified publications were independently scrutinised and summarised to present:

  • Author and year of study

  • Reported structures: What was the regime of the OSCE conducted (e.g. conventional or online), what settings took place (e.g. online systems, breakout rooms and IT support) and what training for staff and students was provided if any.

  • Reported processes: How the OSCEs were carried out, number of students and staff involved, logistics and any methods used to prevent cheating and disruption.

  • Reported outcomes: How it was determined if the OSCEs were successful, student and staff feedback and any suggestions for improvement.

  • Reported cancelled, rescheduled or modified dental OSCEs. Reasons for decisions and any suggestions made.

  • Reported methods of assessment equivalent/comparable to OSCEs in terms blueprinting, learning outcomes assessed and application method.

Analysis of the identified articles was performed based on the reported structures, processes and outcomes of OSCEs. Articles reporting cancellation or rescheduling of dental OSCEs were reviewed including the reason and any suggestions/recommendations made. This is to help explore why a dental institution was unable to run the OSCE; was it lockdown measures, unable to plan a safe, infection transmission‐free environment or even facility issues. In addition, if an institution reported rescheduling, was the reason explained, alternatives described, second attempt described?

3. RESULTS

3.1. Selection of sources of evidence

A search on OSCEs in dental education during the COVID‐19 pandemic yielded a total of 290 studies. As independent researchers searched various databases, a total of 239 sources were found to be duplicates, and hence removed, leaving 51 studies for title and abstract evaluation. A further 34 articles were excluded as they did not report on the topic of interest, leaving 17 for full‐text evaluation, of which nine were retained following application of the inclusion and exclusion criteria (Figure 1).

FIGURE 1.

FIGURE 1

Records selection in accordance with PRISMA for systematic scoping review

3.2. Characteristics of sources of evidence

A summary of the data extracted from each source is presented in Table 2. The table provides a deeper insight into the reasons behind decisions taken by institutions. Thematic analysis of the subject matter was conducted independently by AD and KS, to describe relevant data under each subheading, then a final agreement was achieved.

TABLE 2.

Table of results; evidence gathered from included articles

Reported OSCE Features during COVID‐19 pandemic
Study Structures Processes Outcomes
Kakadia et al. 2020 37

Regime: Online OSCE

Settings:

Qualtrics
  • For marking the exam
Zoom
  • Selected because of its ‘breakout’ room feature which allowed private sessions for each student.

Training: The purpose of training was to allow standardisation through examiner calibration. This was done on two occasions, 1 month and 1 week prior to the OSCE. Secondly, there was a trial OSCE held for both the students and the examiners.

Execution: OSCE ran with two parallel groups, Group Red and Green.

34 students examined, split into Group Red or Green. 20 examiners; 2 examiners for each of the 10 disciplines examined within either Group Red or Green.

One Zoom link with 21 pre‐assigned breakout rooms was created. Each breakout room had an examiner, 1 room used as a rest station.

A clinical case was disseminated to each student 1 hour before the start time. Zoom host placed examiners in breakout rooms before students. Students had to sign in with their full names. Once all students had joined, they were placed into their first breakout rooms by the Zoom host. Six minutes were allowed for each examination. Students were then moved to the next room of the appropriate colour. One minute warnings were provided. Examiners graded students on Qualtrics or on hard copy.

Prevention of cheating: Students were asked to keep cameras and microphones on at all times.

Prevention of disruption: This institution held a rehearsal of the OSCE beforehand

Results: Not reported

Feedback reported in article:

Positive:
  • Student: Time saving
  • Student: Less anxious
  • Student: More privacy
  • Staff: The rehearsal was appreciated
  • Staff: It went smoothly
Negative:
  • Student: Poor internet connectivity
  • Student: Poor image quality
  • Student: Poor sound quality
  • Staff & Students: Errors in entering/leaving breakout rooms and difficulty re‐entering
  • Staff: no hands‐on questions
  • Staff: difficult to grade between the students.
Suggestions made in article:
  • An established form of communication between examiners and students at all times e.g. a phone line
  • A supporting co‐host who is able to take over should the actual examiner become unavailable
  • Examination case/question screen shared for the entire duration to maximise the amount of time examiner has with the student.
Khalaf et al. 2020 42

Regime: Online OSCE

Settings:

Blackboard
  • Examination meeting
Phone lines
  • Communication if the Blackboard meeting suffered disturbance.
Teams
  • Meetings between faculty members in preparation for exam.
  • Ran simultaneously with Blackboard as an extra means of communication during the exam and as a backup in case the examination meeting using Blackboard suffered disturbance.
“Respondus”
  • lockdown browser – for prevention of any other programme running on the computer

Training: Online meetings were held on teams between staff. Training and mock exams were held for the students beforehand.

Execution: 71 students were examined in an online OSCE as part of an overall ‘exit’ examination for final year dental students in Sharjah, UAE. This examination consisted of MCQs, MEQs, oral examinations, and the OSCE. The OSCE formed 30% of the overall exit exam grade.

Prevention of cheating: The “Respondus” lockdown browser.

10 min checking: a video recording of the student ID and the room they were in. Students allowed a maximum of 10 min late arrival to prevent sharing of information.

Prevention of disruption: Teams meetings were held alongside Blackboard meetings in case the communication through blackboard was lost. They were held on mobile devices using mobile data, therefore if Wifi connectivity went down, the student could let their invigilator know this way. In the case of lost connectivity, the student was placed in a direct phone call with their invigilator to not disturb the other students, once connectivity was restored, the invigilator would arrange for the host to allow the student back into the examination.

Results: Outcome of the online Exit exam for the current cohort was similar to the conventional exam, however, specific OSCE results were not discussed.

Feedback reported in article:

Positive:
  • 69% of the 29 responding staff agreed they were satisfied with the online OSCE.
  • 60% of students felt the training and mock exams held beforehand were adequate to familiarise them with online format.
Negative:
  • Only 17% of staff agreed online OSCE assessed a different set of knowledge/skills than traditional OSCEs
  • 33% of students felt online exams were unreliable and 44% felt online exam favoured some students over others. 37% of the 65 students agreed they had adequate previous knowledge of online systems (Blackboard, teams, etc.) whilst 36% of students felt less anxious after completing the practice exams.
  • 37% of students were satisfied their final grade reflected their performance and only 32% of students were satisfied overall.
  • Teams meeting in the background created unnecessary noise.
  • Technical issues with internet connectivity.
Suggestions made in article:
  • More IT training needed

Donn et al. 2020

32

Regime: Online OSCE

Settings:

PowerPoint:
  • How examination information, questions and cases were presented.
Zoom
  • Pre examination meetings for examiners
  • For the exam meeting.

Training: Practice sessions carried out for both students and examiners.

Appropriate questions selected from the OSCE bank of questions that had previously been used. They were selected based on their appropriateness for this new style of examining.

Zoom meetings were held to go through each question and arrive at a passmark.

Execution: A total of 10 questions for the three resitting students sitting the Virtual OSCE (VOSCE). 12 min stations, each with cases presented via PowerPoint, showcasing radiographs, photographs and special investigation results. Each breakout room featured two questions, two examiners and one reserve examiner. Examiner took on the role of simulated patient (SP) if required. Zoom meeting set up with breakout rooms. As many breakout rooms as number of exam questions. An additional room for the external examiner, and a student meeting room. Waiting room feature enabled to allow students physically into the examination by examiners. Five co‐hosts created allowing examiners to navigate freely between rooms without needing host's permission to enter. Briefing held before examiners were placed into their individual breakout rooms. Students briefed on what to expect & how to customise screen displays. Exams marked independently at the end of the OSCE by two examiners then a final mark for each student agreed. Examiners received an electronic copy of the mark sheet beforehand or a printed copy sent by recorded delivery. Mark sheets were then sent back electronically.

Prevention of cheating: All students were asked to complete an academic integrity statement

Prevention of disruption: Host emailed an emergency contact number for any IT issues. Initial meetings with each student online allowed evaluation of Wi‐Fi connectivity and IT device requirements. Students consented to OSCE being recorded for Quality Assurance purposes. To compensate for new format of questions and any IT disruption, original time allotted was doubled. Practice sessions held allowing identification of logistical and timing challenges.

Results: No information was provided on the results obtained for the OSCE.

Feedback reported in article:

Positive:
  • For majority of staff and students, VOSCE progressed in the same manner as a conventional OSCE
  • All staff and students agreed the VOSCE was a good substitute for the OSCE given the circumstances.
  • Both groups felt it allowed students to give an accurate reflection of their abilities. They did not feel disadvantaged by the new format. Staff mentioned marking felt similar to marking the conventional OSCE.
  • Training sessions considered vital.
  • Staff happy to take on role of SP. Likewise, students felt at ease with it.
  • Recording considered helpful for reviewing and staff calibration
  • Cloud for storage useful
Negative:
  • Staff intensive.
  • Extended time per station unnecessary.
  • Inability to assess practical skills
Suggestions made in article:
  • Practical skills may be assessed through longitudinal evaluations and competency assessments.
  • The use of WhatsApp or FaceTime to continue the exam in the event of IT failures.
  • Increased time for moving students into correct rooms.
  • Modification of software, making it more user friendly; improving communication to provide two‐way contact between the virtual breakout rooms and the host, increasing the size of text whilst in ‘Broadcast’ function to aid readability.

Alon et al. 2020

31

Regime: Online OSCE

Settings:

Canvas
  • For information presenting prior to the exam.
  • For the exam meeting.

Training: A grading criteria table was formed as well as a content of the expected answers

Execution: A module was created on the university online portal publishing what's to be expected from the exam and studying materials. Exam integrated: open‐ended questions, multiple choice questions, and “fill in the blank”. Exam included questions about materials, diagnosis, treatment planning, clinical procedure, emergencies, complications, and communication. Pass grade decided after OSCEs were marked.

Prevention of cheating: The assignment was only available on the portal for a limited time.

Prevention of disruption: Not reported

Results: Majority of the students were deemed competent.

Feedback reported in article:

Positive:
  • The evaluation process was rapid and successful
Negative:
  • This virtual format did not allow practical assessments.

Suggestions made in article: Not reported

Hytonen et al 2020

30

Regime: Online OSCE

Settings:

Moodle
  • For the exam.

Training: Students had the choice to complete a practice test online with similar style questions to the new style OSCE

Execution: 179 dental students from all four dental schools in Finland examined simultaneously. Eight subject entities included up to 6 questions. The total exam was 90 min; 10 min per subject entity, and a 10‐min for changing between questions. An additional 10 min was provided for students with dyslexia. The questions had to be answered in the given order, students were unable to return to previous questions. Each subject entity had its own pass mark set between 50–70% and each passed individually. Students received model answers at the end. A re‐sit 2 weeks after the initial OSCE was allowed.

Prevention of cheating: Not reported

Prevention of disruption: Uniform setting for the online Moodle examination were created and edited nationally by the IT support personnel. A preferred browser was recommended. Use of a computer was advised instead of a phone or tablet. An emergency contact number was provided to the students

Results: The pass rate was lower than the traditional OSCE.

Feedback reported in article:

Positive:
  • Students: received adequate information prior to exam
  • Students: positive attitude to the exam
  • Students: practice test useful and technical implementation of the OSCE was good.
  • Students: sufficient teaching in relation to the OSCE
  • The time frame was appropriate and the difficulty of the questions was adequate
Negative:
  • Before the exam, students worried about the possibility of technical issues, inability of OSCE to asses practical skills, professional behaviour and communication skills.
Suggestions made in article:
  • Include more patient cases and clinical photos.
  • Split each subject entity into individual exams lasting 10 min each, instead of a single 80‐min exam.
  • Increasing the number of subject entities
  • Incorporating oral examinations as a new subject entity.

Wu et al. 2020

58

Decision: reschedule OSCE assessment.

Reason: Challenging to ensure patient and student safety if use of live patients is required.

Suggestions made in article: A new style of OSCE has been suggested; The Dental Licensure OSCE which is designed to examine clinical judgement and decision‐making. This style of OSCE does not require procedures to be performed on live patients

Iyer et al. 2020

23

Decision: Cancel or reschedule OSCE assessment.

Reason: To satisfy quarantine requirements

Suggestions made in article: A multidisciplinary OSCE combined with a written exam by sharing resources across different dental schools

Hattar et al. 2020

59

Decision: Cancel OSCE assessment.

Reason: To satisfy quarantine requirements.

Suggestions made in article: Online replacements for OSCE's were prepared. Students however, did not feel the online replacements were adequate

Assessment method comparable to OSCE
Stoopler et al. 2020 29

Regime: Online.

Settings: BlueJeans

For students and examiners meeting

Canvas

For reading the case presentation during exam.

Training: Not reported

Execution: Split into two components. Component A for students to review a case presentation using Canvas, 30 min allocated. In B, students met examiners using BlueJeans answering five questions in 15 min. Predeveloped answer sheet with correct answers, and critical failure answers used. Students told if they had passed or failed at the completion of the examination.

Prevention of cheating: Not reported.

Prevention of disruption: Not reported

Results: all students successfully completing the Oral Cancer CCE on their first attempt.

Feedback reported in article: Faculty members believed this method of examination was effective in assessing clinical competency related to oral cancer. Student and staff feedback was positive with particular focus on the efficiency of exam administration.

Suggestions made in article: Not reported

One study demonstrated a virtual oral cancer clinical competency examination (CCE). 29 The authors described the CCE as a “vehicle to ensure attainment of skill, knowledge, and value by dental students in a specific discipline”. The study was included in the current review as it was deemed similar to an OSCE in terms of level of competencies assessed, learning outcomes measured and methodology. Describing this assessment may inform the reader of how such assessment was approached in terms of structures, processes and outcomes.

3.3. Reported OSCE experience during the pandemic

Out of the nine included articles, six conducted the dental OSCE online, whilst the other three studies reported cancelling or rescheduling the OSCE, indicating that none of the studies ran their dental OSCEs face to face. All the studies demonstrating online OSCEs described the online setting utilised to run the OSCE; however, that varied from one institution to the other depending on many factors including familiarity with a particular portal, resources available and software specifications such as breakout rooms. The majority of articles described delivering online training for both staff and students to accustom with the new OSCE format, calibrate examiners, confirm appropriateness of questions and arrive at a pass mark. With regard to processes, conveying information on the execution of online OSCEs varied in depth amongst included articles, with some providing thorough informative information and others choosing to be brief. Most studies explored different ways of preventing cheating during the online OSCE, ranging from signing an academic integrity statement to using video recording and cameras open at all times.

Measuring the outcomes, all selected articles presented student and staff feedback to an extent, providing some idea regarding the perceptions and attitudes. No article clearly reported the results of the online OSCE compared to previously run face‐to‐face OSCEs. There was an apparent lack of consistency in approaches and data on effectiveness of methods reported.

4. DISCUSSION

4.1. Statement of key findings

This review was constrained to nine articles fulfilling the inclusion criteria. Notwithstanding, the reviewed articles were generally homogeneous in their themes, namely structures, processes and outcomes, which allowed for an in‐depth cross examination between them. Out of these, six dental OSCEs were performed online whilst the remaining were either cancelled or postponed. Although no reviewed article clearly mentioned why online methods were deemed preferable, intuitively they rigidly observed the quarantine protocols imposed during the pandemic. Moreover, this review observed that due to the longstanding opinion that OSCEs are an essential feature of the undergraduate dental assessment portfolio, most articles opted to run the OSCEs, clearly reinforcing the discernible efforts made in order to preserve them, even throughout the pandemic. Amongst the reviewed articles, the structures and processes of online OSCEs described diverse applications, however, not surprisingly, reported its inability to assess practical skills. 30 , 31 , 32 Outcomes generally presented staff and student feedback, however, were scarce in revealing reliability and validity when running the assessment online.

4.2. Interpretation

The scoping review has identified that different dental institutions have responded contrary to one another in their handling of this dilemma. Structures and processes attributed to dental OSCEs must ensure validity, reliability and reproducibility of the assessment style, in addition to its ability to examine practical skills; this makes it an unparalleled form of assessment in producing desired outcomes. 30 Exploration of OSCEs beyond the review revealed that in addition to its high fidelity, the strength of such an assessment tool lies in its ability to provide students with an opportunity to demonstrate essential competencies that are required of a graduating dentist to begin safe, independent, unsupervised dental practice. 33 On the contrary, a study suggested that other methods of assessments such as workplace‐based assessments (WPBAs) should now take centre stage in clinical assessments of undergraduate medical students and advocated a reduced role for OSCEs. 34 Another study exploring the perceptions of clinical assessment stakeholders (students, examiners, simulated patients and administrators) in relation to the future role of traditional OSCEs equally revealed WPBAs utilises a more holistic approach to ensuring undergraduates' work‐readiness, as it focuses on professional behaviours and identity in comparison to the OSCE. 35 Programmatic assessments have also been advocated, as it allows for comprehensive information gathering about the leaner's competence and the accumulation of information from a variety of assessment tools. 36

4.2.1. Structures

Based on the reviewed articles, the selection of a suitable online software is dependent on each institution with preference given to continuing with current online university portals. 29 Additionally, it may be considered more important to use a software with previous examiner and student familiarity; or use due to specific features within the software such as breakout rooms, which may aid in the architectural implementation of the OSCE. 37 Alternatively, Learning Management Systems such as “Moodle” and “Blackboard” have been suggested for a similar purpose. These are software applications allowing examiners to insert pictures and videos on which the student may then be examined. Further research within the medical field has disclosed their ability to use advance assessment settings, such as shuffling the order of question and answer options, which may be considered advantageous when designing an online OSCE. 38

To date, the fundamental feature underpinning all published studies holding a dental OSCE during the pandemic was the fact they took place online. Contrarily, a number of medical institutions did attempt to hold face‐to‐face OSCEs during the pandemic. 39 , 40 Auxiliary research within the medical field suggests online is less effective than the face‐to‐face OSCE in assessing practical skills. 40 Conversely, simulation‐based online OSCEs improve clinical outcomes, save time 41 and are preferred by the students. 40

Provision of training for staff and students must be a pivotal aspect within the structure of online OSCEs. Five out of the six studies highlighted the importance of running training sessions which may allow for standardisation, examiner calibration and evaluation of students' internet connectivity, 30 , 31 , 32 , 37 , 42 with one study emphasising on selecting OSCE questions based on their appropriateness for this new style of examining. 32 Furthermore, training must focus on improving familiarity and reducing anxiety of students facing this new examination format. 42 A similar study in the medical field recommended the need for a specifically trained team in the technology utilised. 43 In another medical study questioning alternative opportunities to enhance assessments during the pandemic, four structural ideas were suggested to accommodate the many changes which OSCEs may have to undergo: communication, flexibility, collegiality and compassion. 44

4.2.2. Processes

The execution of online OSCEs varied significantly from one study to the other, providing rich data on how dental institutions may carry out such assessments tailored to their need. Undoubtedly, this may be considered a weakness too, as lacking consensus and evidence‐based approaches may jeopardise the validity and reliability of an OSCE. It would have been valuable to the reader to present the rationale behind reported methods of implementing OSCEs, as the assurance of evidence to support the validity of OSCE scores is an essential condition to guarantee a high‐quality and effective assessment. 45 Each of the articles reviewed had varied in the depth of information provided relating to the execution of the OSCE. In terms of the timing per station, it ranged from 6 to 12 min each with no justification per se. Nevertheless, it was helpful to identify reported feedback on assigned timings for online OSCE stations. In one of the included studies, stations were allocated 6 min each, whereby students perceived the whole OSCE experience as “time saving” compared to their previous live OSCEs. 37 Another study assigned 12 min for each OSCE scenario; however, feedback from students and staff regarded the extended time per station as “unnecessary”. 32 Allocating appropriate and realistic time for tasks at individual stations improves the extent to which the test's content is representative of the actual skill; in other words, enhances validity. 17 For practical purposes, a broader search of literature has generally recommended that sufficient reliability could be achieved with 14–18 stations of 5–10 min duration each, 46 whilst the Cronbach's or Generalisability (G) value should fall between 0.7 and 0.8 for high‐stakes examinations. 47 Coincidently, the feedback of one of the current studies has therefore deemed 12 min per station unnecessary and could be reduced. 32 Administrative support has also been praised in a few of the included studies. Institutions planning to conduct OSCEs must guarantee adequate administrative support. 17 For online purposes, including tasks such as trialling online platforms suitable for purpose, allocating students to different breakrooms, logistics of moving students between stations and distribution of examiner mark sheets, having the right administrative support will enable the smooth running of the exam. 29

A repeatedly cited feature from our review was institutional efforts to ensure integrity of students' responses. Reflection on why cheating is wrong, its subsequent connotation of the student and their place within the profession is a topic of in‐depth ethical consideration. Outside of the review, the importance of this is emphasised by cases where cheating has led to a student's repetition of an academic year. The importance of academic integrity is ultimately due to its potential to echo into patient care. 48 The online OSCEs have recognised and actioned this in many forms including the completion of an academic integrity statement 32 ; a recorded “check‐in” process of the surrounding environment ensuring no elicit materials were accessible to the students; and use of software such as “Respondus” to prevent access of any other materials on the device being used for the examination. 42

Dependence on the unreliable nature of internet connectivity is an inherent constraint of online OSCEs. This review identified many attempts made to minimise the likelihood of such disruptions including simultaneous video calls alongside students' actual examination video calls in case it should fail 42 ; allowing more than usual time per OSCE station to countenance rectification of any connectivity matters, the latter considered unnecessary following later evaluation of the OSCE event as previously stated. 32 A discernible way of countering this would be the traditional face‐to‐face OSCE, which can be conducted in hospital wards, clinical skills laboratories or specially designed test canters. However, there is currently no evidence‐based method of conducting a COVID‐19 secure face‐to‐face dental undergraduate OSCE. A recent article concluded that both staff and students considered the virtual OSCE (VOSCE) a useful and fair assessment method, and a suitable alternative to the traditional OSCE, to overcome the difficulties inherent in traditional clinical examination, during pandemic restrictions, providing a number of improvements in the assessment process are undertaken. 49

4.2.3. Outcomes

The majority of dental OSCE experiences explored to date failed to confirm comparability of online OSCE results with previous live OSCEs, indicating a gap when relaying outcomes, leaving the reader short of establishing evidence‐based conclusions on the reliability and validity of online dental OSCEs in testing practical skills when needed. A fundamental principle of any assessment is making certain it is designed to assess stated learning outcomes. 50 In an article outside of our review outlining guidelines for online medical assessments in emergency remote teaching (ERT) during the COVID‐19 Pandemic, it was proposed that assessments may focus on the cognition involved in clinical skills instead of the practicality, such as history‐taking and patient counselling skills. 51 However, assessing the psychomotor domain has proven challenging in the scenario of university closures and social‐distancing orders. 52

Dentistry is a practical profession; therefore, assessing these skills is an essential feature of the undergraduate training programme. 53 Wider research suggests exploring methods through which this may be incorporated. These include development or modification of software devices or removal of the practical evaluation from the OSCE and into longitudinal observations throughout the year using clinical portfolios and practical competencies. 30 , 32 Interestingly, one of the included studies described the use of OSCE as a final‐year exit assessment alongside MCQs, MEQs and an oral exam, all administered online. However, the authors did not elucidate the reason for conducting an online OSCE at this stage; though, they stated that these exams are part of a graduation comprehensive assessment that “does not exist in other medical/dental colleges” and has been developed and implemented several years prior to COVID‐19 pandemic. 42 A recent study in a UK dental institution provided a proof of concept for the validity of virtual reality (VR) dental simulator assessments. It revealed the ability of VR to provide reliable and clinically relevant qualitative feedback on basic operative dental procedures, 33 which could in essence be explored as a validated method going forward to assess manual dexterity.

Additional study exposing elevated levels of anxiety and unpreparedness for the workforce of dental students as a result of the pandemic has been discussed in depth 54 ; it is therefore of vital importance to contemplate the impact of changes to the OSCE, on the already wavering confidence of the premature dental professionals. In most studies retrieved in this scoping review, both staff and students agreed that the online OSCE was a good substitute for the conventional OSCE given the circumstances, allowing students to give an accurate reflection of their abilities. 29 , 32 , 42 Students reported they felt less anxious, had more privacy and did not feel disadvantaged by the new online format. 37 Nevertheless, nearly half of the students in one study felt the online exam favoured some over others, 42 with others showing concerns regarding internet connectivity and the possibility of technical issues occurring during the exam. 30 , 37 , 42

Further review of literature outside of dentistry has revealed examples of live OSCEs held during the pandemic 39 ; Bouriscot et al have documented an example of how to execute an OSCE during the pandemic that is comparable to those before the occurrence of COVID‐19, but also defensible in its compliance to new regulations. Their take‐home message was to ensure that sufficient measures are needed to conduct a clinical examination safely including strict infection control practices, minimising the mixing of students, staff and patients, social distancing and videoconferencing‐facilitated briefings for training and calibration purposes. The study reported no COVID‐19 infections amongst participants afterwards, and the defensibility of the examination results was affirmed by the external examiners. 39

On the other hand, a paediatric medicine experience suggested that a “teleOSCE” is a comparable assessment to a live OSCE as no significant differences were found when compared to the previous 3 years' live OSCEs. 55 A similar study found that a virtual objective structured clinical examination (VOSCE) for medical undergraduate students was not only feasible to deliver, but also acceptable to both students and examiners, with no complaints or concerns displayed. Nonetheless, they plan to return to the traditional “gold standard” live OSCE format once the COVID‐19 situation allows, acknowledging the need to implement elements of programmatic assessments into their curriculum. 56

The expansive scope of the present review provides dental educators with a synopsis of key considerations to be aware of when designing a dental OSCE compliant within the COVID‐19 pandemic. Areas that lack depth of information modestly provide direction on precisely where further research in this area must be focussed.

4.3. Further research

Despite reporting that as many as 19% of dental institutions continued to hold face‐to‐face examinations, 57 the authors of the present review have yet to identify details of any traditional dental face to face OSCE held during the pandemic. Examples of how this may have been done are essential if there is to be evidence‐based analysis of the possible structures, processes and outcomes rendering this a potential option. Additionally, further research is needed to better inform pedagogic and assessment strategies to the possibility of utilising VR haptics for assessing operative and other dental practical skills.

4.4. Strengths and weaknesses

This scoping review is strengthened by its adherence to the robust PRISMA‐ScR guidelines. The generic nature of a scoping review deprives itself of a qualitative assessment of the research articles. Nevertheless, the authors of the present article considered that this review identified useful existing features, as well as highlighting areas requiring further exploration. The small number of available articles identified and the search period potentially limit the generalisability of the findings and their likely impact. This number was considered few in comparison to other healthcare professional fields; wider research within the medical profession revealed there were many more studies published relating to OSCEs during the COVID‐19 pandemic. Despite these limitations, the authors were reassured that the quality of the articles in the current review presents an in‐depth conversation of a similar nature of findings, allowing comparability and discussion of the key features for this scoping review.

Restricting the search to articles published in English only could potentially mean an OSCE experience during the pandemic might have been left out; yet, the authors opted to select a common scientific language and were in the opinion that including articles in languages other than English may not add to the findings of the scoping review. As with any search strategy employed, articles could be missed out; however, conducting a pilot search independently by the authors, searching a variety of relevant databases and running multiple regular searches reinforced the search strategy and consolidated the data set retrieved.

5. CONCLUSION

It is suggested that adopted structures and processes for online OSCEs may be utilised during the COVID‐19 pandemic to measure cognitive skills. Regarding reported outcomes, little information has been provided on the specifics relating to the results of online dental OSCEs; therefore, its genuine capability to replace the traditional face‐to‐face style remains undetermined. The COVID‐19 pandemic is likely to change dental training and assessment approaches; accordingly, dental institutions might need to consider alternative ways such as longitudinal and competency evaluations to assess manual dexterity. Further research into structures, processes and outcomes beyond this study period maybe required.

CONFLICT OF INTEREST

The authors have no conflicts of interest to disclose.

Salawu YK, Stewart D, Daud A. Structures, processes and outcomes of objective structured clinical examinations in dental education during the COVID‐19 pandemic: A scoping review. Eur J Dent Educ. 2022;00:1‐13. doi: 10.1111/eje.12869

DATA AVAILABILITY STATEMENT

The data that support the findings of this study are available from the corresponding author upon reasonable request.

REFERENCES

  • 1. Arnold RC, Walmsley AD. The use of the OSCE in postgraduate education. Eur J Dent Educ. 2008;12:126‐130. doi: 10.1111/j.1600-0579.2008.00469.x [DOI] [PubMed] [Google Scholar]
  • 2. Standards for Education . Standards and requirements for providers. Standards for Education; 2015. [Google Scholar]
  • 3. Davis MH. OSCE: the Dundee experience. Med Teach. 2003;25:255‐261. doi: 10.1080/0142159031000100292 [DOI] [PubMed] [Google Scholar]
  • 4. Harden RM, Gleeson FA. Assessment of clinical competence using an objective structured clinical examination (OSCE). Med Educ. 1979;13:39‐54. doi: 10.1111/j.1365-2923.1979.tb00918.x [DOI] [PubMed] [Google Scholar]
  • 5. Puryer J. Dental undergraduate views of objective structured clinical examinations (OSCEs): a literature review. Dent J. 2016;4:6. doi: 10.3390/dj4010006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Mossey D, Newton P, Stirrups J. Scope of the OSCE in the assessment of clinical skills in dentistry. Br Dent J. 2001;190:323‐326. [DOI] [PubMed] [Google Scholar]
  • 7. Jain SSM, DeLisa M, Joel AMD, et al. Further experience IN development of an objective structured clinical examination for physical medicine and rehabilitation residents. Am J Phys Med Rehabil. 1998;77:306‐310. [DOI] [PubMed] [Google Scholar]
  • 8. Hodges B. OSCE! Variations on a theme by Harden. Med Educ. 2003;37:1134‐1140. doi: 10.1111/j.1365-2923.2003.01717.x [DOI] [PubMed] [Google Scholar]
  • 9. Stillman PL, Wang Y, Ouyang Q, Zhang S, Yang Y, Sawyer WD. Teaching and assessing clinical skills: a competency‐based programme in China. Med Educ. 1997;31:33‐40. doi: 10.1111/j.1365-2923.1997.tb00040.x [DOI] [PubMed] [Google Scholar]
  • 10. Novack DH, Volk G, Drossman DA, Lipkin M. Medical interviewing and interpersonal skills teaching in US medical schools: Progress, problems, and promise. Jama J Am Med Assoc. 1993;269:2101‐2105. doi: 10.1001/jama.1993.03500160071034 [DOI] [PubMed] [Google Scholar]
  • 11. Leichner P, Sisler GC, Harper D. A study of the reliability of the clinical oral examination in psychiatry. Can J Psychiatry. 1984;29:394‐397. doi: 10.1177/070674378402900506 [DOI] [PubMed] [Google Scholar]
  • 12. Sanson‐Fisher RW, Poole AD. Simulated patients and the assessment of medical students' interpersonal skills. Med Educ. 1980;14:249‐253. doi: 10.1111/j.1365-2923.1980.tb02269.x [DOI] [PubMed] [Google Scholar]
  • 13. Swanson DB, van der Vleuten CPM. Assessment of clinical skills with standardized patients: state of the art revisited. Teach Learn Med. 2013;25:S17‐S25. doi: 10.1080/10401334.2013.842916 [DOI] [PubMed] [Google Scholar]
  • 14. Zayyan M. Objective structured clinical examination: the assessment of choice, Oman. Medizinhist J. 2011;26:219‐222. doi: 10.5001/omj.2011.55 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Structure ‐ Wikipedia ; n.d.
  • 16. Effective Structures for Assessment | University of Maryland ‐ Teaching and Learning Transformation Center ; n.d.
  • 17. Khan KZ, Gaunt K, Ramachandran S, Pushkar P. The objective structured clinical examination (OSCE): AMEE guide No. 81. Part II: organisation & administration. Med Teach. 2013;35(9):e1447‐e1463. doi: 10.3109/0142159X.2013.818635 [DOI] [PubMed] [Google Scholar]
  • 18. Process ‐ Wikipedia , (n.d.).
  • 19. Braun HI, Mislevy RJ. Intuitive Test Theory CSE Report 631; 2004.
  • 20. Neumann LM, MacNeil RL. Revisiting the national board dental examination. J Dent Educ. 2007;71:1281‐1292. doi: 10.1002/j.0022-0337.2007.71.10.tb04392.x [DOI] [PubMed] [Google Scholar]
  • 21. Proffitt E. What will be the new normal for the dental industry? Br Dent J. 2020;228:678‐680. doi: 10.1038/s41415-020-1583-x [DOI] [PubMed] [Google Scholar]
  • 22. Checchi V, Bellini P, Bencivenni D, Consolo U. COVID‐19 dentistry‐related aspects: a literature overview. Int Dent J. 2020;71:idj. 12601‐idj. 12626. doi: 10.1111/idj.12601 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Iyer P, Aziz K. Summary review/COVID‐19 and dental education abstract. J Dent Educ. 2020;21:46‐47. doi: 10.1002/jdd.12163 [DOI] [Google Scholar]
  • 24. Arksey H, O'Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol Theory Pract. 2005;8:19‐32. doi: 10.1080/1364557032000119616 [DOI] [Google Scholar]
  • 25. Williams B, Song JJY. Are simulated patients effective in facilitating development of clinical competence for healthcare students? A scoping review. Adv Simul. 2016;1:6. doi: 10.1186/s41077-016-0006-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Khalil H, Peters M, Godfrey CM, Mcinerney P, Soares CB, Parker D. An evidence‐based approach to scoping reviews. Worldviews Evidence‐Based Nurs. 2016;13:118‐123. doi: 10.1111/wvn.12144 [DOI] [PubMed] [Google Scholar]
  • 27. Moher D, Liberati A, Tetzlaff J, Altman DG. Preferred reporting items for systematic reviews and meta‐analyses: the PRISMA statement. BMJ. 2009;6:e1000097. doi: 10.1371/journal.pmed.1000097 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. (PDF) Methodology for JBI Scoping Reviews ; n.d.
  • 29. Stoopler ET, Kuperstein AS, Berardi TR, Sollecito TP. A virtual oral cancer clinical competency examination administered during the COVID‐19 pandemic. J Dent Educ. 2020;85(Suppl. 1):1004‐1005. doi: 10.1002/jdd.12291 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Hytönen H, Näpänkangas R, Karaharju‐Suvanto T, et al. Modification of national OSCE due to COVID‐19 – implementation and student feedback. Eur J Dent Educ. 2020;25:679‐688. doi: 10.1111/eje.12646 [DOI] [PubMed] [Google Scholar]
  • 31. Alon E, Amato R. Evaluation of endodontic competency in the COVID‐19 era: problem, solution and results. J Dent Educ. 2020;85(Suppl. 1):1000‐1003. doi: 10.1002/jdd.12237 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Donn J, Scott JA, Binnie V, Bell A. A pilot of a virtual objective structured clinical examination in dental education. A response to COVID‐19. Eur J Dent Educ. 2020;25:488‐494. doi: 10.1111/eje.12624 [DOI] [PubMed] [Google Scholar]
  • 33. Jonathan D, Ashley T, Nicolas M, James F. Re‐defining the virtual reality dental simulator: demonstrating concurrent validity of clinically relevant assessment and feedback. Eur J Dent Educ. 2020;25:108‐116. doi: 10.1111/eje.12581 [DOI] [PubMed] [Google Scholar]
  • 34. Khan H. OSCEs are outdated: clinical skills assessment should be centred around workplace‐based assessments (WPBAS) to put the ‘art’ back into medicine. Med EdPublish. 2017;6:189. doi: 10.15694/mep.2017.000189 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Malau‐Aduli BS, Jones K, Saad S, Richmond C. Has the OSCE met its final demise? Rebalancing clinical assessment approaches in the peri‐pandemic world. Front Med. 2022;9:1‐7. doi: 10.3389/fmed.2022.825502 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Schuwirth L, van der Vleuten C, Durning SJ. What programmatic assessment in medical education can learn from healthcare, Perspect. Med Educ. 2017;6:211‐215. doi: 10.1007/s40037-017-0345-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Kakadia R, Chen E, Ohyama H. Implementing an online OSCE during the COVID‐19 pandemic. J Dent Educ. 2021;85(Suppl. 1):1006‐1008. doi: 10.1002/jdd.12323 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Khan RA, Jawaid M. Technology enhanced assessment (TEA) in COVID 19 pandemic. Pakistan J Med Sci. 2020;36:S108‐S110. doi: 10.12669/pjms.36.COVID19-S4.2795 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Boursicot K, Kemp S, Ong TH, et al. Conducting a high‐stakes OSCE in a COVID‐19 environment. MedEdPublish. 2020;9:1‐8. doi: 10.15694/mep.2020.000054.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Gordon M, Patricio M, Horne L, et al. Developments in medical education in response to the COVID‐19 pandemic: a rapid BEME systematic review: BEME Guide NO. 63. Med Teach. 2020;42:1202‐1215. doi: 10.1080/0142159X.2020.1807484 [DOI] [PubMed] [Google Scholar]
  • 41. Tabatabai S. Simulations and virtual learning supporting clinical education during the covid 19 pandemic. Adv Med Educ Pract. 2020;11:513‐516. doi: 10.2147/AMEP.S257750 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Khalaf K, El‐Kishawi M, Moufti MA, Al Kawas S. Introducing a comprehensive high‐stake online exam to final‐year dental students during the COVID‐19 pandemic and evaluation of its effectiveness. Med Educ Online. 2020;25:1‐11. doi: 10.1080/10872981.2020.1826861 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Reid MD, Sam AH. Reflections on assessment in the wake of change from the COVID‐19 pandemic. Med Educ. 2021;55:128‐130. doi: 10.1111/medu.14368 [DOI] [PubMed] [Google Scholar]
  • 44. Fuller R, Joynes V, Cooper J, Boursicot K, Roberts T. Could COVID‐19 be our ‘there is no alternative’ (TINA) opportunity to enhance assessment? Med Teach. 2020;42:781‐786. doi: 10.1080/0142159X.2020.1779206 [DOI] [PubMed] [Google Scholar]
  • 45. Yazbeck Karam V, Park YS, Tekian A, Youssef N. Evaluating the validity evidence of an OSCE: results from a new medical school. BMC Med Educ. 2018;18:1‐8. doi: 10.1186/s12909-018-1421-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Epstein RM, Cox M, Irby DM. Assessment in medical education. New Engl J Med. 2007;(4):387‐396. [DOI] [PubMed] [Google Scholar]
  • 47. Pell G, Fuller R, Homer M, Roberts T. How to measure the quality of the OSCE: a review of metrics AMEE guide no. 49. Med Teach. 2010;32:802‐811. doi: 10.3109/0142159X.2010.507716 [DOI] [PubMed] [Google Scholar]
  • 48. Currie W, Dracopoulos S, Hendry G. Cheating in a dental practical exam. Int J Educ Integr. 2017;13:5. doi: 10.1007/s40979-017-0017-2 [DOI] [Google Scholar]
  • 49. Donn J, Scott JA, Binnie V, Mather C, Beacher N, Bell A. Virtual objective structured clinical examination during the COVID‐19 pandemic: an essential addition to dental assessment. Eur J Dent Educ. 2022:1‐10. doi: 10.1111/eje.12775 [DOI] [PubMed] [Google Scholar]
  • 50. Boitshwarelo B, Reedy AK, Billany T. Envisioning the use of online tests in assessing twenty‐first century learning: a literature review. Res Pract Technol Enhanc Learn. 2017;12:16. doi: 10.1186/s41039-017-0055-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Pharmacy | Qatar University ; n.d.
  • 52. Lewandowski R, Stratton A, Sen Gupta T, Cooper M. Twelve tips for OSCE‐style tele‐assessment. MedEdPublish. 2020;9:1‐16. doi: 10.15694/mep.2020.000168.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Scheutzel P, Gerhard‐Szép S. ‘Practical skills’ ‐ positioning of the GMA committee for dentistry. GMS Z Med Ausbild. 2016;33(4):1‐2. doi: 10.3205/zma001047 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Alraheam IA, Alashqar MA, Hattar S et al. Self‐perceived preparedness of dental school graduates and the impact of COVID‐19 pandemic on their condence ; 2020. doi: 10.21203/rs.3.rs-49483/v2. [DOI]
  • 55. Lara S, Foster CW, Hawks M, Montgomery M. Remote assessment of clinical skills during COVID‐19: a virtual, high‐stakes, summative pediatric objective structured clinical examination. Acad Pediatr. 2020;20:760‐761. doi: 10.1016/j.acap.2020.05.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Boyle JG, Colquhoun I, Noonan Z, McDowall S, Walters MR, Leach J. Viva la VOSCE? BMC Med Educ. 2020;20:514. doi: 10.1186/s12909-020-02444-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Quinn B, Field J, Gorter R, et al. COVID‐19: the immediate response of european academic dental institutions and future implications for dental education. Eur J Dent Educ. 2020;24:811‐814. doi: 10.1111/eje.12542 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Wu DT, Wu KY, Nguyen TT, Tran SD. The impact of COVID‐19 on dental education in North America—where do we go next? Eur J Dent Educ. 2020;24:825‐827. doi: 10.1111/eje.12561 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Hattar S, AlHadidi A, Sawair FA, Abd Alraheam I, El‐Ma'aita A, Wahab FK. Impact of COVID‐19 pandemic on dental academia. Students' experience in online education and expectations for a predictable practice. Res Sq. 2020;1:1‐16. doi: 10.21203/rs.3.rs-54480/v1 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.


Articles from European Journal of Dental Education are provided here courtesy of Wiley

RESOURCES