gms | German Medical Science

GMS Journal for Medical Education

Gesellschaft für Medizinische Ausbildung (GMA)

ISSN 2366-5017

DOPS (Direct Observation of Procedural Skills) in undergraduate skills-lab: Does it work? Analysis of skills-performance and curricular side effects

article Assessment

Search Medline for

  • corresponding author Christoph Profanter - Medical University Innsbruck, Department of Visceral, Transplant and Thoracic Surgery, Innsbruck, Austria
  • author Alexander Perathoner - Medical University Innsbruck, Department of Visceral, Transplant and Thoracic Surgery, Innsbruck, Austria

GMS Z Med Ausbild 2015;32(4):Doc45

doi: 10.3205/zma000987, urn:nbn:de:0183-zma0009879

This is the English version of the article.
The German version can be found at: http://www.egms.de/de/journals/zma/2015-32/zma000987.shtml

Received: January 14, 2014
Revised: December 19, 2014
Published: October 15, 2015

© 2015 Profanter et al.
This is an Open Access article distributed under the terms of the Creative Commons Attribution 4.0 License. See license information at http://creativecommons.org/licenses/by/4.0/.


Abstract

Objective: Sufficient teaching and assessing clinical skills in the undergraduate setting becomes more and more important. In a surgical skills-lab course at the Medical University of Innsbruck fourth year students were teached with DOPS (direct observation of procedural skills). We analyzed whether DOPS worked or not in this setting, which performance levels could be reached compared to tutor teaching (one tutor, 5 students) and which curricular side effects could be observed.

Methods: In a prospective randomized trial in summer 2013 (April – June) four competence-level-based skills were teached in small groups during one week: surgical abdominal examination, urethral catheterization (phantom), rectal-digital examination (phantom), handling of central venous catheters. Group A was teached with DOPS, group B with a classical tutor system. Both groups underwent an OSCE (objective structured clinical examination) for assessment.

193 students were included in the study. Altogether 756 OSCE´s were carried out, 209 (27,6%) in the DOPS- and 547 (72,3%) in the tutor-group.

Results: Both groups reached high performance levels. In the first month there was a statistically significant difference (p<0,05) in performance of 95% positive OSCE items in the DOPS-group versus 88% in the tutor group. In the following months the performance rates showed no difference anymore and came to 90% in both groups.

In practical skills the analysis revealed a high correspondence between positive DOPS (92,4%) and OSCE (90,8%) results.

Discussion: As shown by our data DOPS furnish high performance of clinical skills and work well in the undergraduate setting. Due to the high correspondence of DOPS and OSCE results DOPS should be considered as preferred assessment tool in a students skills-lab.

The approximation of performance-rates within the months after initial superiority of DOPS could be explained by an interaction between DOPS and tutor system: DOPS elements seem to have improved tutoring and performance rates as well.

DOPS in students ‘skills-lab afford structured feedback and assessment without increased personnel and financial resources compared to classic small group training.

Conclusion: In summary, this study shows that DOPS represent an efficient method in teaching clinical skills. Their effects on didactic culture reach beyond the positive influence of performance rates.

Keywords: DOPS, skills lab, WBA (workplace based assessment), curricular side effects


Introduction

Sufficient teaching and assessment of clinical skills already in the undergraduate setting becomes more and more important in view of changing requirements in health systems and planned reforms of medical curricula [1], [2], [3], [http://kpj.meduniwien.ac.at/fileadmin/kpj/oesterreichischer-kompetenzlevelkatalog-fuer-aerztliche-fertigkeiten.pdf]. For example in Austria a national competence level catalogue was recently implemented at all national medical universities [4]. Limitations of personnel and financial resources are challenging for academic teaching. Didactic methods will be evaluated in future not only in regard of performance but also of cost efficiency [5], [6], [7], [8].

Today there is much discussion in literature about efficiency of teaching and assessment methods [9], [10], [11], [12], [13]. MiniCEX (mini clinical evaluation exercise) and DOPS (direct observation of clinical skills) are used for some time as workplace based assessment (WBA) instruments in postgraduate settings [14], [15], [16], [17], [18], [19], [20], [21], [22], [23]. Recently, this didactic format is also used as WBA tool in final year undergraduate courses [24]. If DOPS is a reliable tool in teaching and assessment of clinical skills earlier in the curriculum, e.g. in the fourth year, is unknown [24]. Still open remains the question whether repeated assessment, e.g. with DOPS, is superior to conventional exercising in regard of performing clinical skills. Karpicke and Blunt stated in a prospective randomized study in the field of cognitive skills that retrieval practice produces more learning than elaborative studying [25]. Transferred to an undergraduate student´s skills lab DOPS would be the retrieval practice (repeated assessment) and the tutor teaching in small groups the elaborative learning. DOPS is well established as assessment tool in the context of workplace, especially in postgraduate settings [14], [15]. Simulation is a reliable instrument to create a workplace environment which is safe and provides a student centered setting [26].

In a prospective randomized study we analyzed the following questions:

  • Does DOPS work in an undergraduate skills-labs setting?
  • Does DOPS improve performance of clinical skills compared with tutor based teaching?
  • How good is assessment quality of DOPS in this setting?
  • Which curricular side effects result from the implementation of DOPS?

Methods

Setting/Study design

A prospective randomized trial at the Medical University of Innsbruck (MUI) from April to June 2013 was carried out in context of surgical practical studies in the eighth semester. The study was presented to the local ethics committee, which approved the study design and raised no objection to its realization.

The multidisciplinary surgical practical studies are conducted primarily by the department of Visceral, Transplant and Thoracic Surgery (VTT), the following departments contribute to the studies in accordance to the department´s size: Department of Anesthesia and Intensive Care Medicine, Department of Trauma Surgery and Sports medicine, Department of Orthopedics and Orthopedic Surgery, Department of Heart Surgery, Department of Vascular Surgery, Department of Plastic, Reconstructive and Esthetic Surgery, Department of Urology, Department of Neurosurgery.

193 students out of a year cohort (n=258) were included in the study. The students were randomized into two groups. In both groups the same four competence-level-based skills were teached: surgical abdominal examination, urethral catheterization (phantom), rectal-digital examination (phantom), handling of central venous catheters.

Group A was teached and assessed with DOPS only by lecturers of VTT. Group B was teached with a tutor system with lecturers of all other mentioned departments. Tutor teaching comprised small group lectures without a defined didactic method and one medical teacher present all the time.

Randomization/Design of practical studies

The students were randomized in group A “DOPS” and Group B “Tutor” at the registration by chance. According to the chronological order of the registration odd numbers were randomized in the DOPS group, even numbers into the tutor group. The initial size of each group for analysis after randomization was n=109. The practical studies took place at the same time but in different rooms for each study group, which consisted of maximum five students. Each unit of the practical studies took five days, 90 minutes a day. At the first day all students got a theoretical lecture in all four skills based on detailed handouts for the lecturers. In the following days two, three and four the students were teached with DOPS or tutor system according to randomization. The students in the DOPS group were repeatedly assessed with the goal (minimum) 6 DOPS per student. Day five was dedicated to the OSCE assessment for all students. The OSCE comprised four five-minute stations in accordance to the teached skills. The OSCE test reports corresponded exactly to the mentioned handouts for the teachers (see Figure 1 [Fig. 1]).

Inclusion/Exclusion criteria

The registration for the practical studies (“Chirurgisches Praktikum”) was the only inclusion criterion. To warrant validity of data attention was carefully paid to avoid group shifts during the lecture days. Therefore, the following criteria for exclusion were defined in the study design:

1.
Group shifting/students shifting form one group into the other (e.g. student of group A, four days in group A, one day in group B)
2.
Students of DOPS group who missed 1 or more days (goal of 6 DOPS minimum not within reach)
3.
Students in the Erasmus program, because they could not be obliged to undergo OSCE.
Skills/Materials/Assessment

The skills urethral catheterization and rectal-digital examination were teached and assessed on phantoms produced by Things and Limbs (Catheterization-Simulator, Rectal Examination Trainer MK2). The skill surgical abdominal examination was teached by the students as models themselves. For the corresponding OSCE station special trained student-actors were recruited. The skill handling of central venous catheters was teached and assessed on real devices that are used daily in the university hospital of Innsbruck. All OSCE examiners and actors were blinded regarding the randomization of the students.

The OSCE report forms were generated in the Innsbruck design (see Figure 1 [Fig. 1]). Content and layout of DOPS forms are in accordance with the original publications of Norcini and Darzi [14], [15].

Analysis/Statistics

Calculating OSCE results, the cut off for positive/negative rating was set between the category “completely accomplished” and “partially accomplished”. Thus, only item results from the category “completely accomplished” were rated as positive. The OSCE items were divided into two fractions according to the report form: skills and communication. The OSCE results were analyzed separately for each of the four skills and every month. The results were expressed in absolute numbers (points) and percent values.

Statistical analysis of OSCE results was performed by t-test. A p value <0,05 indicated a statistically significant difference.

DOPS items were also divided in two fractions – skills and communication. The skills fraction consisted of preparation/aftercare/certainty, technical performance, clinical judgment, organization and efficiency. The other items as professionalism were related to the communication fraction.

DOPS results in the categories satisfactory and superior were rated as positive and were summarized. The calculated percent values were compared with the OSCE results.

Instruction/Resources

All tutors and examiners were experienced residents or consultants/assistant professors and underwent specific OSCE teaching. Lecturers (and examiners) of VTT had an additional DOPS training. Special attention was directed to achieve a sufficient selectivity in assessment with OSCE and DOPS. The study design defined personnel and time related resources identically in both groups. Thus, the goal of 6 DOPS in summary or 2 DOPS each day was a result of this definition. Five students and 90 minutes per day amount to nine minutes for each DOPS as a total of observation and feedback time.


Results

From the total number of students (n=193), n=52 (26,9%) were in the DOPS and 141 (73,1%) in the tutor group. 756 singular OSCE´s were performed, n=209 (27,6%) in the DOPS, n=547 (72,3%) in the tutor group. 180 (93,3%) students underwent all OSCE stations, 13 (6,7%) less than four. As these students could unmistakably be related to one of the groups (DOPS or tutor) in accordance to the criteria mentioned above, these OSCE results could be rated as well. 90,8% of the students in the DOPS group and 89,8% in the tutor group underwent all OSCE stations.

The size difference between the two groups (DOPS n=52, tutor n=141) can be explained by data adjustment in consequence of exclusion criteria: The most frequent reasons were groups shifting from DOPS to tutor and not suitable DOPS reports. Furthermore, some student groups randomized to DOPS did not participate in the practical studies at all and had to be cancelled.

OSCE Skills

Overall, there cannot be found any difference in positive OSCE results between the study groups: 90,8% DOPS versus 89% tutor. Analyzing the monthly OSCE results leads to a different finding: in April the DOPS group was statistically significantly superior to tutor group with 94,1% versus 87,5% (p< 0,05). During the following months the performance rates approximated to each other. The rate of completely accomplished items was in May 91,2% (DOPS) and 91,1% (tutor), in June 89,3% versus 90,8% respectively (see Figure 2 [Fig. 2])

Also the analysis of the singular skills in the months May and June furnishes results close to each other. The percent mean values for DOPS and tutor were 91% versus 90,8% in May and 90% versus 90,5% in June.

OSCE Communication

Generally, the performance in communication items was worse with 76,6% DOPS and 78,4% tutor (percent mean values over all months).

DOPS/Correspondence DOPS-OSCE

320 DOPS were performed at all. Analysis of skills related DOPS items amounted to 92,4% positive results. This value corresponds well with 90,8% positive OSCE results in the DOPS group.

We found a rate of 91,3% positive DOPS communication items. The correspondence with 76,6% positive OSCE results is substantial lower (see Table 1 [Tab. 1])

Resources

Due to the study design there was no difference regarding time and personnel resources. In both groups 5 students were assessed (DOPS) or supervised (tutor) by one teacher 90 minutes per day. The expenditure of time in personnel teaching was 6 hours. An instructor trained lecturers and examiners in groups with a maximum size of 15 people. Staff members of VTT and staff members of the other departments were always instructed separately.


Discussion

DOPS as a teaching tool in an undergraduate setting achieve high performance levels of clinical skills. In our study, the performance rates in the DOPS group were found to be significantly (p<0,05) superior to the tutor system only in the first month of the semester (April). In the following months, May and June, and over the whole semester results showed no difference anyway (see Figure 2 [Fig. 2] and Table 1 [Tab. 1]) Thus, we conclude that DOPS works well in a student´s skills lab setting and scores at least equally with a tutoring in small groups and academic teachers.

Our data show a high correspondence between DOPS and OSCE results regarding skills. We venture to conclude that DOPS as assessment tool in students skills lab may be sufficient at least as pass and fail rating. Therefore, the OSCE could be dropped without diminishing the assessment quality. This would allow saving staff and time resources. According to the slogan “assessment time also is teaching time” either some more skills could be teached during the same time or the practical studies itself could be shortened e.g. four instead of five days. In our opinion this is an interesting curricular side effect that comes along with DOPS implementation in teaching undergraduates. Compared with a tutor system DOPS provide not only assessment but also structured feedback without generating additional costs.

The approximation of skills performance rates after initial superiority of the DOPS group may be explained by interaction between DOPS and tutor system. Essential DOPS elements as structured feedback and repeated assessment were taken over from the teachers in the tutor system in the course of increasing mental embodiment. The study design provided, that only staff members of VTT supervise the DOPS groups and staff members of the associated departments the tutor groups. Only during the first weeks of these practical studies this mode could be realized. Later on, frequently staff members of VTT had to supervise also tutor groups out of organizational reasons. Nevertheless, both groups were always teached in different rooms.

In contrast to the positive results regarding skills as psychomotor competences, OSCE outcome concerning communicative items was substantial poorer. In both groups we found only about 77% positive results. This can be interpreted on the one hand as an achievement of adequate selectivity in assessment, which was an important topic in teachers’ instructions. On the other hand these data show weaknesses of students in this regard. Thus, this means a commission to focus more on these communicative aspects in the future within the skills lab setting. However, it remains open, which method is the best to achieve this improvement. Strictly speaking, we would have expected better results in the DOPS group. Probably, the intended focus on the psychomotor aspects in our setting and course design may be an explanation for the poorer OSCE results, but not for the also poor correspondence between DOPS and OSCE regarding these communicative aspects.


Conclusion

In summary, our study shows that DOPS represents an efficient method in teaching clinical skills. Due to the high correspondence between DOPS and OSCE results regarding clinical skills it could be considered to carry out assessment in a student´s skills lab only with DOPS. Furthermore, DOPS implementation seems to have positive influence on the didactic culture of academic institutions.


Competing interests

The authors declare that they have no competing interests.


References

1.
Frenk J, Chen L, Bhutta ZA, Cohen J, Crisp N, Evans T, Fineberg H, Garcia P, Ke Y, Kelley P, Istnasamy B, Meleis A, Naylor D, Pablos-Mendez A, Reddy S, Scrimshaw S, Sepulvenda J, Serwadda D, Zurayk H. Health professionals for a new century: transforming education to strenghten health systems in an interdependent world. Lancet. 2010;376(9756):1923-1958. DOI: 10.1016/S0140-6736(10)61854-5 External link
2.
Mann KV. Theoretical perspectives in medical education: past experience and future possibilities. Med Educ. 2011;45(1):60-68. DOI: 10.1111/j.1365-2923.2010.03757.x External link
3.
Epstein RM. Assessment in medical education. N Engl J Med. 2007;356(4):387-396. DOI: 10.1056/NEJMra054784 External link
4.
Frank JR. The CanMEDS 2005 physician competency framework. Better standards. Better physicians. Better care. Ottawa: The Royal College of Physicians and Surgeons of Canada; 2005.
5.
Finucane P, Shannon W, McGrath D. The financial costs of delivering problem-based learning in a new, graduate-entry medical programme. Med Educ. 2009;43(6):594-598. DOI: 10.1111/j.1365-2923.2009.03373.x External link
6.
Neville AJ. Problem-based learning and medical education forty years on. A review of its effects on knowledge and clinical performance. Med Princ Pract. 2009;18(1):1-9. DOI: 10.1159/000163038 External link
7.
Norman SR, Schmidt HG. Effectiveness of problem-based learning curricula: theory, practice and paper darts. Med Educ. 2000;34(9):721–728. DOI: 10.1046/j.1365-2923.2000.00749.x External link
8.
Yaqinuddin A. Problem-based learning as an instructional method. J Coll Physicians Surg Pak. 2013;23(1):83-85. DOI: 01.2013/JCPSP.8385 External link
9.
Colliver JA. Effectiveness of problem-based learning curricula: research and theory. Acad Med. 2000;75(3):259-66. DOI: 10.1097/00001888-200003000-00017 External link
10.
Bannister SL, Hilliard RI, Regehr G, Lingard L. Technical skills in pediatrics: a qualitative study of acquisition, attitudes and assumptions in the neonatal intensive care unit. Med Educ. 2003;37(12):1082-1090. DOI: 10.1111/j.1365-2923.2003.01711.x External link
11.
Remmen R, Derese A, Scherpbier A, Denekens J, Hermann I, van der Vleuten C, Van Royen P, Bossaert L. Can medical schools rely on clerkships to train students in basic clinical skills? Med Educ. 1999;33(8):600-605. DOI: 10.1046/j.1365-2923.1999.00467.x External link
12.
McGaghie WC, Issenberg SB, Cohen ER, Barsuk JH, Wayne DB. Does simulation-based medical education with deliberate practice yield better results than traditional clinical education? A meta-analytic comparative review of the evidence. Acad Med. 2011;86(6):706-711. DOI: 10.1097/ACM.0b013e318217e119 External link
13.
Buss B, Krautter M, Möltner A, Weyrich P, Werner A, Jünger J, Nikkendei C. Can the Assessment Drives Learning" effect be detected in clinical skills training? – Implications for curriculum design and reource planning. GMS Z Med Ausbild. 2012;29(5):Doc70. DOI: 10.3205/zma000840 External link
14.
Norcini JJ, Blank LL, Duffy FD, Fortna GS. The MiniCEX: A method for assessing clinical skills. Ann Intern Med. 2003;138(6):476-481. DOI: 10.7326/0003-4819-138-6-200303180-00012 External link
15.
Darzi A, Mackay S. Assessment of surgical competence. Qual Health Care. 2003;10:64-69.
16.
Norcini J, Burch V. Workplace-based assessment as an educational tool: AMEE Guide Nr. 31. Med Teach. 2007;29(9):855-871. DOI: 10.1080/01421590701775453 External link
17.
Wragg A, Wade W, Fuller G, Cowan G, Mills P. Assessing the performance of specialist registrars. Clin Med. 2003;3(2):131-134. DOI: 10.7861/clinmedicine.3-2-131 External link
18.
Miller A, Archer J. Impact of workplace based assessment on doctor´s education and performance: a systematic review. BMJ. 2010;341:c5064. DOI: 10.1136/bmj.c5064 External link
19.
Govaerts MJ, Schuwirth LW, Van der Vleuten CP, Muijtjens AM. Workplace-based assessment: effects of rather expertise. Adv in Health Sci Educ. 2011;16(2):151-165. DOI: 10.1007/s10459-010-9250-7 External link
20.
Pelgrim EA, Kramer AW, Mokkink HG, van der Vleuten CP. The process of feedback in workplace-based learning: organistation, delivery, continuity. Med Educ. 2012;46(6):604-612.
21.
Norcini J. The power of feedback. Med Educ. 2010;44(1):16-17. DOI: 10.1111/j.1365-2923.2009.03542.x External link
22.
Mitchell C, Bhat S, Herbert A, Baker P. Workplace-based assessments of junior doctors: do scores predict training difficulties? Med Educ. 2011;45(12):1190-1198. DOI: 10.1111/j.1365-2923.2011.04056.x External link
23.
Bindal N, Goodyear H, Bindal T, Wall D. DOPS assessment: A study to evaluate the experience and opinions of trainees and assessors. Med Teach. 2013;35(6):e1230-e1234. DOI: 10.3109/0142159X.2012.746447 External link
24.
McLeod R, Mires G, Ker J. Direct observed procedural skills assessment in the undergraduate setting. Clin Teach. 2012;9(4):228-232. DOI: 10.1111/j.1743-498X.2012.00582.x. External link
25.
Karpicke JD, Blunt JR. Retrieval practice produces more learning elaborative studying with concept mapping. Science. 2011;331(6018):772-775. DOI: 10.1126/science.1199327 External link
26.
Ker J, Bradley P. Simulation in Medicine Education Booklet. Understanding Medical Education Series. Ediburgh: ASME; 2007.