JCDR - Register at Journal of Clinical and Diagnostic Research
Journal of Clinical and Diagnostic Research, ISSN - 0973 - 709X
Dentistry Section DOI : 10.7860/JCDR/2019/37913.12594
Year : 2019 | Month : Feb | Volume : 13 | Issue : 02 Full Version Page : ZC13 - ZC17

Introduction and Comparison of Direct Observation of Procedural Skills (DOPS) with Conventional Method of Skill Assessment in Dental Students

Daisy Happy1, Amita Aditya2, Parag Hadge3, Nihal Devkar4, Akshay Vibhute5

1 Reader, Department of Periodontology, Sinhgad Dental College and Hospital, Pune, Maharashtra, India.
2 Reader, Department of Oral Medicine and Radiology, Sinhgad Dental College and Hospital, Pune, Maharashtra, India.
3 Reader, Department of Periodontology, DY Patil Dental College and Hospital, Lohegaon, Pune, Maharashtra, India.
4 Professor, Department of Periodontology, Sinhgad Dental College and Hospital, Pune, Maharashtra, India.
5 Reader, Department of Periodontology, Sinhgad Dental College and Hospital, Pune, Maharashtra, India.


NAME, ADDRESS, E-MAIL ID OF THE CORRESPONDING AUTHOR: Daisy Happy, Sinhgad Dental College and Hospital, Off Sinhgad Road Vadgaon (Bk), Pune, Maharashtra, India.
E-mail: drdaisyhappy@gmail.com
Abstract

Introduction

Procedural competence and professional behaviour are integral components of clinical skills in Dentistry. However till date, most of the clinical procedures performed by undergraduate dental students are only assessed at the end of the procedure. Because of this, there could be a lack of assessment of their actual procedural and patient handling skills. Also, because of lack of a standard format, there could be a lot of ambiguity in assessment scores given by different examiners for same performance.

Aim

The aim of this study was to introduce and compare Direct Observation of Procedural Skills (DOPS) with conventional method of case presentation for skill assessment of dental undergraduate students.

Materials and Methods

DOPS, as a skill assessment tool was analysed on the basis of its feasibility, acceptability, validity and reliability. Checklist for DOPS assessment of Scaling and Root Planing (SRP) procedure was generated and process of internal validation was completed. Forty final year Bachelor of Dental Surgery (BDS) students participated in this study during their clinical posting in the Department of Periodontology. One of the most routinely performed procedure by the students in the clinical posting; i.e., SRP was selected for skill assessment. Each student performed SRP on patients and was assessed by two assessors using conventional method (CA1 and CA2) and two assessors using DOPS method (DA1 and DA2). Three such assessments were carried out for each student i.e. on day 1, day 2 and after feedback i.e., at the end of their clinical posting of 14 days. Descriptive analysis, ANOVA, Tukey’s Post-Hoc and Spearman-Brown formula were used for statistical analysis concerned with reliability as compared to the conventional method. Feedback from students and assessors was used to analyse the feasibility and acceptability of DOPS.

Results

Correlation between DOPS examiners (74%, 71%) was significantly higher as compared to correlation between conventional examiners (23%, 55%). According to Spearman-Brown formula, reliability was found to be 80-90% in case of DOPS, and 50-70% in case of conventional method. Based on feedback collected from students and faculty regarding the feasibility and acceptability of DOPS, 97% students and all faculty members chose DOPS over conventional method.

Conclusion

According to this study, DOPS seems to be a reliable, valid, acceptable and feasible method of skill assessment in dental students.

Keywords

Introduction

Acquiring clinical competence and demonstrating professional behaviour on day to day basis has been emphasised upon lately in the field of health science education. Work Place Based Assessment (WPBA) is one of the modalities, which can be used to assess the trainee in authentic settings. The two cardinal elements of WPBA are ‘direct observation’ and ‘conducted in work place’ in addition to provision of ‘feedback’ to the trainee. The WPBA conforms to the highest (Level 4: ‘Does’) of Miller’s pyramid and also has the potential to assess at all four levels [1].

Direct Observation of Procedural Skills is mainly implied for assessment in medical settings; however few authors have reported it to be a good assessment tool in dentistry as well [2-6]. DOPS as a method of assessment can be applied in real time settings and can assess both performance skills and doctor patient interaction of a student. In fact, according to certain studies in the past, it has been reported to improve the skill learning of trainees significantly [7,8].

In dentistry, undergraduate training is said to be skill-based but most disciplines do not have any formative assessment program for the same in India. After going through the literature, need of incorporating a new assessment method for skill assessment was felt. One of the procedures performed by under graduate dental students on patients regularly is the SRP. However, during their clinical postings it is assessed only at the completion of procedure, because of which procedural skills and behaviour/interaction of student with patient is not assessed. In addition, due to lack of a standard and structured proforma of assessment in most of the departments, there seems to be a lot of ambiguity in assessment scores given by different examiners for same performance. This questions the reliability and validity of the existing assessment methods. So, in context of above mentioned problems, the present study was undertaken with the aim to assess reliability, validity, feasibility and acceptability of DOPS as an assessment method and compare it with the conventional post-procedural method of skill assessment in dentistry.

Materials and Methods

After taking approval from the Institutional Scientific Advisory Committee and Ethical Committee, a cross-sectional study was carried out involving final year BDS students during their clinical posting in the Department of Periodontology. As per the University norms, BDS students have three clinical postings in Department of Periodontology, two in their third year BDS and one in Final year. Final year students were considered for inclusion in this study as they are sufficiently trained for routine periodontal procedures during their third year and are expected to appear for the summative evaluation in fourth year BDS. Project was explained to students and informed written consent was obtained. Three batches i.e., 40 students participated in study. At the same time, sensitisation of faculty of the Department of Periodontology regarding the project was also done. Four faculty members volunteered to participate in the study amongst whom two were assigned as Conventional Assessors (CA1 and CA2) and two as DOPS Assessors (DA1 and DA2) by lottery method.

Checklist for DOPS assessment of SRP procedure was generated [Annexure 1] and process of internal validation was completed. To generate a valid checklist for DOPS, first of all the competencies required to perform SRP were identified and accordingly Specific Learning Objectives (SLOs) were prepared. Based on these SLOs, checklist for DOPS was generated. While doing SRP, along with dexterity to perform procedure on patient without injury, knowledge about instruments and principles of instrumentation is also needed. Above all, patient doctor interaction in terms of greeting patient, assurance and directions given to patient during procedure, answering patient’s doubts and post-procedural instructions are most important and an integral part of the procedure. So, checklist was designed and pre-validated by peers and senior faculty members to assess important aspects of all domains i.e., cognitive, psychomotor and affective. In addition, total score (65) used for assessment was same as score allotted by Maharashtra University of Health Sciences (MUHS) for university examination (summative assessment). A total of 13 sections were made in checklist, each section having weightage of score 5. Further, in each section, performance was rated on scores from 1 to 5 with 1 being the minimum and 5 being the maximum. In addition, feedback forms for students and faculty were prepared. The feedback forms for both the students and faculty had five questions each, pertaining to their perception regarding feasibility and acceptability of DOPS as assessment tool in comparison with conventional or the existing method [Annexure 2]. These feedback forms were personally handed over to the students and faculty and collected back.

The study was designed in such a way that reliability estimation was feasible. Each student was assessed using conventional method by two assessors (CA1 and CA2) and using DOPS by other two assessors (DA1 and DA2). Two such assessments were carried out before giving feedback to students, thus it was possible to calculate inter and intra examiner reliability and correlation for both the methods of assessment.

Students were subjected to both kinds of assessment i.e., conventional and DOPS by different examiners on three days of their clinical posting; Day 1, 2 and on the last two days (13th and 14th day) of their posting. Feedback was given to the students after day 2 that is the second assessment. Descriptive analysis and ANOVA was used to determine and compare the scores given by Conventional and DOPS assessors. To check pair-wise difference between both the groups, Tukey’s Post-hoc was applied.

Results

Forty (35 females and 5 males) final Year BDS students and four staff members of Department of Periodontology participated in the study. [Table/Fig-1] shows descriptive statistics and [Table/Fig-2] shows pair-wise comparison between scores assigned by conventional assessors and DOPS assessors on day 1, day 2 and after feedback (three assessments for each student). Comparison of the mean scores given by examiners (conventional and DOPS) showed significant difference in the scores given by conventional and DOPS assessors except for pair 5. [Table/Fig-3] represents inter-examiner correlation and reliability of Conventional and DOPS method. As evident, DOPS showed better correlation between different examiners and better reliability as compared to conventional method. As far as intra-examiner reliability was concerned, all the DOPS assessors and conventional assessor 2 had intra-examiner reliability above 0.7(70%) but conventional assessor 1 has achieved only 0.379 i.e., 37.9% reliability between his own two set of scores. ANOVA was used to compare the assessment scores on all days. ANOVA showed that groups (day1, day 2 and after feedback) were significantly different for both DOPS assessors but not for conventional assessors. Further to check pair-wise difference between groups, Tukey’s Post-Hoc was applied [Table/Fig-4]. Statistically significant improvement after feedback was seen as compared to Day 1 and Day 2 for both DOPS assessors (DA1, DA2). Although improvement was there after feedback as compared to Day 1 and Day 2 with conventional assessment also (CA1, CA2) [Table/Fig-3] but it failed to reach the level of significance.

Scores assigned by conventional and DOPS assessor.

nRangeMinimumMaximumMeanStd. Deviation
CA 1 Day 14014385243.084.122
CA 2 Day 14018304838.103.357
DA 1 Day 1407283530.882.377
DA 2 Day 14013253832.633.019
CA 1 Day 24021345544.054.437
CA 2 Day 24015294438.953.566
DA 1 Day 24014274132.103.193
DA 2 Day 24019254433.833.876
CA 1 after feedback4017355243.754.156
CA 2 after feedback4013334638.683.269
DA 1 after feedback4019345344.403.855
DA 2 after feedback4016355145.083.872

(CA=Conventional Assessor, Da=Dops Assessor).


Pair-wise comparison between scores assigned by conventional and DOPS assessors.

MeannStd. DeviationStd. Error MeantSig. (2-tailed)
Pair 1CA 1 Day 143.08404.1220.65215.7130.000
DA 1 Day 130.88402.3770.376
Pair 2CA 2 Day 138.10403.3570.5316.8110.000
DA 2 Day 132.63403.0190.477
Pair 3CA 1 Day 244.05404.4370.70217.8150.000
DA 1 Day 232.10403.1930.505
Pair 4CA 2 Day 238.95403.5660.5647.9010.000
DA 2 Day 233.83403.8760.613
Pair 5CA 1 after feedback43.75404.1560.657-0.8510.400
DA 1 after feedback44.40403.8550.610
Pair 6CA 2 after feedback38.68403.2690.517-9.0190.000
DA 2 after feedback45.08403.8720.612

Inter-Examiner Correlation and Reliability.

PairsDay 1 CA1, CA2Day 2 CA1, CA2After feedback CA1, CA2Day 1 DA1, DA2Day 2 DA1, DA2After feedback DA1, DA2
Pearson’s correlation (r-value)0.2380.558*0.4190.744*0.818*0.714*
Spearman brown coefficient0.3850.716*0.5900*0.853*0.900*0.833*

*Correlation is significant at the 0.01 level. DOPS showed better correlation between different examiners and better reliability (above 70%) as compared to conventional method.


Tukey’s Post-Hoc for multiple comparisons.

Dependent VariableGroupGroupMean DifferenceStd. ErrorSig.
CA 1Day 1After feedback-0.6750.9480.756
Day 2After feedback0.3000.9480.946
CA 2Day 1After feedback0.5750.7600.730
Day 2After feedback-0.2750.7600.930
DA 1Day 1After feedback13.50.7150.000
Day 2After feedback12.30.7150.000
DA 2Day 1After feedback12.40.8080.000
Day 2After feedback11.40.8080.000

Validity

Validity is a unitary concept requiring evidence from a variety of sources (content, response process, internal structure, relationship to other variables, and consequences) [9].

Valid assessment method means it is assessing what it meant to assess. Although improvement in scores were seen by both methods of assessment after giving feedback to students based on their two performances, this improvement was significantly more in DOPS score.

Reliability

According to Downing SM, reproducibility of assessment data or scores over time or occasions is a major source of validity evidence [10]. When inter-examiner statistics were applied, the Pearson’s correlation between two conventional examiners were found to be 0.238 and 0.55 on day 1 and day 2 of assessment, respectively. On the other hand, correlation was found to be 0.744 and 0.711 on day 1 and 2 respectively between two DOPS examiners. It concludes that correlation between DOPS examiners (74%, 71%) were significantly higher as compared to correlation between conventional examiners (23%, 55%). The reliability of the whole test can then be predicted using the Spearman-Brown formula. Downing expects reliability in the “range of 0.70-0.79 or so.” [10]. With DOPS it is found to be 80-90%, well within required range and with conventional method 50-70%, which is below the range suggested by downing as shown in [Table/Fig-3].

When intra-examiner reliability was analysed, it is a found to be above 70% for all examiners (CA2, DA1, DA2) except CA1 for which it was just 37.9%.

Feasibility

DOPS assessment was successfully implemented and was well appreciated by students and faculty. When students as well as faculty were asked about the feasibility of this method in feedback, all had given positive response.

Acceptability

Based on feedback, 97% students and all faculty members had chosen DOPS over conventional method, showing its acceptability. At last, all the students and faculty members agreed that experiencing DOPS during formative assessment will help students to better prepare for summative assessment i.e., their final university examination.

Educational Impact

Improvement in scores of all students after feedback was a clear indication of its educational impact. Performance of students significantly improved after feedback, when their DOPS scores were considered. Students were able to identify their weaker areas: 90% pre-procedural preparation, 50% patient position, 70% implementation of principles of instrumentation, 85% doctor patient interaction, 40% post procedural instructions. Assessment drives learning, thus 90% students had tried to improve.

Teachers were also able to identify problematic areas for students and were able to give timely specific feedback to students. In their feedback, teachers suggested few additions in teaching learning methods to improve problematic areas of students: 1) To improve doctor patient interaction, a live demonstration on patient will be helpful; 2) For post-procedural instructions (plaque control instructions) in addition to discussion and demonstration on jaw models, a simulated patient can also be used; 3) Demonstration on pre-procedural preparation should be included in teaching schedule; 4) Students should be encouraged to learn regional language to improve doctor patient interaction. Implementation of these proposals by teachers was suggested and further assessment of its impact on student’s performance is recommended.

Discussion

The assessment of actual performance that is what the doctor does in practice is the ultimate goal for a valid assessment of clinical competence. Unfortunately, traditional assessment method of post-procedural case presentation is limited in assessing this aspect of clinical skill [11]. DOPS is a well-established assessment tool in the context of work place, especially in Post Graduate (PG) setting in Medical field [12]. This study focuses on evaluating DOPS as an assessment tool for undergraduate students in dental setting in terms of the cardinal attributes of assessment methods that are: validity, reliability, feasibility, acceptability and educational impact.

Results showed statistically significant difference between scores given by conventional method of assessment and DOPS on all the days of assessment (except pair 5). It means same level of performance is being given different weightage by conventional and DOPS assessors. Overall, by conventional method score given on day 1 and 2 was more as compared to DOPS, reason being conventional method has not assessed the way procedure was performed and only the outcome of procedure was evaluated. Thus, it may be inferred by this study that in conventional method of assessment, score depended on trends of scoring in department and generosity of the examiner. On the other hand, DOPS assessed each and every step of procedure, thus score given is less, depending upon the performance of students in different areas and this score has increased only after feedback. It also resulted in improvement in performance of students and was actually reflected in their improved scores. On the other hand, this improvement was not significant when their conventional score were considered because scores given before and after feedback were not significantly different although performance of students has improved after feedback. So, it can be inferred that conventional method of assessment failed to appreciate the improvement in the performance of students.

Use of a checklist enabled the faculty members to provide student-specific and procedure-specific timely feedback to students. This enabled students to identify their weak areas and improve their performance as well as scores which is the main objective of any formative assessment, thus further confirming the validity of DOPS checklist.

In DOPS, use of checklist to score all important areas of procedure decreased the subjectivity; further categorising level of performance i.e., scoring 1-5 might have increased correlation between two examiners. It means that DOPS assessors were better in reproducing their assessment, reason being use of check list that has led to uniformity in assessment. However, one conventional examiner has also achieved reproducibility above 70% reliability which means that even conventional assessment can be made reliable if specific learning objective are clearly defined and task of assessment is performed with responsibility.

Despite its educational effectiveness and simplicity, the formal implementation of DOPS in dental education is still lacking and very few studies have been reported so far. Concerns may be raised regarding DOPS being time taking and less feasible, as mentioned by some previous authors [13]. However, it does not need any extra infrastructure but preparation and planning is required in terms of generation of valid checklist, sensitisation and involvement of faculty and supporting staff. Sensitising students and pre-hand sharing of checklist with them will further improve feasibility. Time management may prove to be a critical factor. According to our experience, DOPS requires approximately 30 minutes per student which is much higher as compared to conventional method which requires approximately 5 minutes per student. So, number of days required to assess whole batch and human resources (examiners and supporting staff) will be more. Still, it is recommended that DOPS should be used at least twice for each batch of students, first one to identify their weaker areas and to give feedback during clinical posting and second one to assess improvement at end-posting.

As far as feasibility and acceptability were concerned, almost all the participants found the DOPS to be a feasible and acceptable assessment tool. This has been earlier also concluded in a similar study done by Singh et al. that the faculty was comfortable using DOPS and found it to be feasible [13].

In fact, DOPS has been found to be useful as assessment learning tool in previous studies as compared to traditional methods and has been recommended to be used for formative assessment [14,15]. However, as with any other assessment tool, the quality of utility depends upon what is assessed rather than how it is assessed.

Limitation

This study had certain limitations, such as lack of systematic internal validation and small sample size. A further study with pre-validated tools on larger sample size of students is recommended.

Conclusion

Within the limitations of this study, DOPS was found to be valid and reliable method for skills assessment. A positive feedback from students and faculty makes it feasible and acceptable too. As it is time consuming, a great deal of support from faculty is needed for which sensitisation sessions for examiners and students are mandatory. Improvement in performance of students after feedback further reinforces the statement “assessment drives learning”.

(CA=Conventional Assessor, Da=Dops Assessor).*Correlation is significant at the 0.01 level. DOPS showed better correlation between different examiners and better reliability (above 70%) as compared to conventional method.

References

[1]Singh T, Modi JN, Workplace-based assessment: a step to promote competency based postgraduate training Indian Paediatrics 2013 50:554-59.10.1007/s13312-013-0164-3  [Google Scholar]  [CrossRef]

[2]Hengameh H, Afsaneh R, Morteza K, Hosein M, Marjan SM, Ebadi A, The effect of applying Direct Observation of Procedural Skills (DOPS) on nursing students’ clinical skills: a randomized clinical trial Global Journal of Health Science 2015 7(7):17-21.10.5539/gjhs.v7n7p1726153199  [Google Scholar]  [CrossRef]  [PubMed]

[3]Roghieh N, Fateme H, Hamid H, The effect of formative evaluation using “direct observation of procedural skills” (DOPS) method on the extent of learning practical skills among nursing students in the ICU Iran J Nurs Midwifery Res 2013 18(4):290-93.  [Google Scholar]

[4]Hoseini BL, Mazloum SR, Foroughipour M, Comparison of midwifery students’ satisfaction with direct observation of procedural skills and current methods in evaluation of procedural skills in Mashhad Nursing and Midwifery School Iran J Nurs Midwifery Res 2013 8(2):94-100.  [Google Scholar]

[5]Lele SM, A Mini-OSCE for Formative assessment of diagnostic and radiographic skills at a dental college in India Journal of Dental Education 2011 75(12):1583-89.  [Google Scholar]

[6]Kundra S, Singh T, Feasibility and acceptability of direct observation of procedural skills to improve procedural skills Indian Pediatrics 2014 51:59-60.10.1007/s13312-014-0327-x24561468  [Google Scholar]  [CrossRef]  [PubMed]

[7]Hattie JA, Identifying the salient facets of a model of student learning: A synthesis of meta-analyses Int J Educ Res 1987 11:187-212.  [Google Scholar]

[8]Veloski J, Boex JR, Grasberger MJ, Evans A, Wolfson DB, Systematic review of the literature on assessment, feedback and physicians’ clinical performance: BEME Guide No. 7 Med Teach 2006 28:117-28.10.1080/0142159060062266516707292  [Google Scholar]  [CrossRef]  [PubMed]

[9]Downing SM, Validity: on the meaningful interpretation of assessment data Med Educ 2003 37:830-37.10.1046/j.1365-2923.2003.01594.x14506816  [Google Scholar]  [CrossRef]  [PubMed]

[10]Downing SM, Reliability: on the reproducibility of assessment data Med Educ 2004 38:1006-12.10.1111/j.1365-2929.2004.01932.x15327684  [Google Scholar]  [CrossRef]  [PubMed]

[11]Al Wardy NM, Assessment methods in Under graduate Medical Education Sultan Qabas Univ Med J 2010 10(20):203-09.  [Google Scholar]

[12]Tabish SA, Assessment methods in Medical education Int J Health Sci 2008 2(2):3-7.  [Google Scholar]

[13]Singh G, Kauur R, Mahajan A, Thomas AB, Singh T, Piloting direct observation of procedural skills in dental education in India Int J Appl Basic Med Res 2017 7(4):239-42.10.4103/ijabmr.IJABMR_54_1729308361  [Google Scholar]  [CrossRef]  [PubMed]

[14]Profanter C, Perathoner A, DOPS (Direct observation od Procedural Skills) in Under graduate skills Lab: Does it work? Analysis of skills performance and curricular side effects GMS Z Med Ausbild 2015 32(4):Doc 45  [Google Scholar]

[15]Akbari M, Shamsabadi R, Direct Observations of Procedural Skills (DOPS) in restorative Dentistry: Advantages and Disadvantages in Student’s point of view Iranian Journal of Medical Education 2013 13(3):212-20.  [Google Scholar]