ABSTRACT
At the University of Minnesota, fourth-year veterinary students assessed their clinical competence after completion of a small–animal, internal-medicine clinical rotation using the same rotation assessment form used by supervising faculty. Grades were compared between the two groups. Students identified by faculty as low-performing were more likely to overestimate their competence in the areas of knowledge, clinical skill, and professionalism than were students identified by faculty as higher performing. This finding mirrors research results in human health professional training. Self-assessment should not be used as the primary or sole measure of clinical competence in veterinary medical training without the introduction of measures to ensure the accuracy of student self-assessment, measures that include active faculty mentoring of student self-assessment, student goal-setting and reflection, and availability of subsequent opportunities to practice additional self-assessment.
The American Veterinary Medical Association Council on Education (COE), the accrediting body for all North American and some foreign veterinary schools, mandates the demonstration of new graduates' clinical competence as follows:
The specific clinical competencies expected of entry-level practitioners are comprehensive patient diagnosis (problem solving), appropriate use of clinical laboratory testing, and record management; comprehensive treatment planning including patient referral when indicated; anesthesia and pain management, patient welfare; basic surgery skills, experience, and case management; basic medicine skills, experience, and case management; emergency and intensive care management; health promotion, disease prevention/biosecurity, zoonosis, and food safety; client communications and ethical conduct; and strong appreciation for the role of research in furthering the practice of veterinary medicine.1
Competency-based education is a growing trend in human and veterinary health professional training. The general goals of competency assessment are that the student first demonstrates knowledge of clinical medicine and also demonstrates the abilities to integrate knowledge from several sources (history, physical examination, records review, laboratory testing), to make sound decisions, and to relay those decisions to clients. The COE does not dictate how schools demonstrate their students' mastery of competencies. The creation of appropriate assessment tools must take into account the ability of those tools to meet these specific learning objectives and must expose health students to a broad sampling of cases to ensure that clinical competence assessment is reliable across the breadth of the discipline.2 Many modalities have been created to try to assess competency in health students, including objective structured clinical examinations (OSCEs) in which the student interacts with a standardized patient or client; case logs or portfolios that may or may not contain descriptive and reflective elements; clinical competency examinations that can be completed either on paper, online, or as practical examinations; assessments by faculty and practitioners as students move through a variety of clinical rotations and externships; skills lists; and students' self-assessments.3–18
Self-assessment can be used for formative or summative evaluation. Formative evaluation is constructive review for the sole purpose of improving subsequent performance. Summative evaluation also includes some aspect of grading. Self-assessment for educational purposes is a natural offshoot of human behavior in other contexts. As social life has evolved, the need for social comparison—that is, the need to compare one's own attributes to those of others—has done so as well.19 High-performing students, including veterinary students, have been proven to spend a considerable amount of mental energy evaluating and comparing themselves to others.20 This self-evaluation can be channeled to good use for learning and, in an educational setting, it also has non-cognitive benefits including increased student motivation and increased communication between students and teachers as students strive to understand how they can best achieve mastery.21
The primary concern with using self-assessment for learning is that human beings are not naturally good at self-evaluation. The ability to assess one's own competence requires many of the same skills that are required to perform other tasks well. As a result, it is difficult for low-performing individuals to judge their own performance accurately. Low-performing individuals might at times overestimate their abilities because of the ambiguity involved in the process.19 Almost any individual, for example, if asked to rate him or herself as a companion, can find some aspect of the large skill set that makes up companionship at which he or she excels. Accuracy of self-assessment also varies with knowledge of external factors, the type of ability being rated, and the context in which the self-assessment occurs.5,22,23 High-performing individuals tend to underestimate their skills, perhaps because of social norms against bragging, unrealistically high personal goals, or unrealistic beliefs about others' comparative skills.9,24 Finally, one might underestimate or overestimate one's skills based on confidence, which has been defined as the willingness to undertake an activity.23 Many facets of human health professional training have demonstrated that low-performing students overestimate their abilities when they self-assess.5–9,11–13,25–30 Veterinary students, like other elite performers, are generally quite self-confident,20 and this confidence can be reflected in an increased self-evaluated level of competence.23
Our hypothesis is that veterinary students' self-assessment of their performance on a small–animal, internal-medicine clinical rotation will not be representative of faculty assessment, especially in the cases of the highest and lowest performing students, as has been shown in other health professions.
The study protocol was approved by the Institutional Review Board (IRB) of the authors' home institution. Fourth-year veterinary students were solicited to participate in the research study and assured that their participation was voluntary, anonymous, and confidential. Students were assured that the collected data would not be accessible to faculty and could in no way affect their grades. The only qualifying criterion was the completion of Small Animal Medicine-B (SAM-B), the students' second two-week block of small–animal, internal-medicine clinical rotation. At the conclusion of the rotation, students completed the same online rotation assessment form used by faculty. This assessment form is a rubric that includes anchors denoting specific standards for A–F grades (see Figure 1). Student self-assessments were coded by staff so that the investigators would not be aware of which students participated. Student self-assessed grades were their own; faculty grades for each student were determined using the combined input of all attending faculty and house officers on that rotation. Grades were reported on a standard A to F scale for knowledge, clinical skills, and professionalism, and an overall grade was assigned. Sensitivity and specificity of the student grade for the determining faculty grade were calculated for each competency separately, as were mean and standard deviation. For the purposes of this analysis, A and B grades denote high competence whereas C and D grades denote low competence.
One hundred students, including fourth-year students from the graduating classes of 2009 and 2010, completed the self-assessment. Of the 100, 88 were from the home institution and 12 were from other schools. Class rank was available for 88 students. Twenty-one students were in the uppermost quartile of their class, 28 in the second quartile, 21 in the third quartile, and 18 in the lowest quartile. Of the 100 students, 45 received an overall grade of A, 42 an overall grade of B, 12 an overall grade of C, and 1 an overall grade of D. The mean grade was 3.3 on a 4-point scale with A equaling 4 points. By way of comparison, the students of the entire class of 2009 (N=84) earned an average grade of 3.1 on their first completion of SAM-B.
The sensitivity and specificity of using student self-assessment to predict the faculty evaluation of C or D (low competency) are shown in Table 1. The sensitivity is the proportion of students assessed by faculty as low-competency students whose self-assessment grade was also C or D. The specificity is the proportion of students assessed by faculty as high–competency students whose self-assessment grade was also A or B. As an example, the sensitivity for knowledge is 11%. This means that of the 19 students who received a C or D grade from faculty for this competency, only 2 (11%) of them would have been identified by self-assessment. Conversely, of the 81 students who received an A or B grade from faculty for knowledge, 74 (91%) would have been identified by self-assessment.
|
| Correct assessmentof C or D | Sensitivity | Correct assessmentof A or B | Specificity | |
|---|---|---|---|---|
| Knowledge | 2/19 | 11% | 74/81 | 91% |
| Clinical skills | 1/14 | 7% | 84/86 | 98% |
| Professionalism | 0/3 | 0% | 97/97 | 100% |
| Overall | 1/13 | 8% | 87/87 | 100% |
To explore the relationship between student self-assessment and faculty assessment further, the mean and standard deviation were calculated for both types of evaluation in all four areas. As Table 2 shows, the average faculty score was lower than the average student self-assessment score and the standard deviation was higher for faculty scores. For all four categories, these differences were statistically significant at the 0.05 level. Means were compared with paired t-tests and variances were compared with F-tests. This indicates that students were more likely to give themselves a grade toward the middle of the range than the faculty, and suggests that students who perform higher are more likely to give themselves lower grades and students who perform lower are more likely to give themselves higher grades.
|
| Score | Faculty | Student |
|---|---|---|
| Knowledge | 3.03 (0.70) | 3.08 (0.51) |
| Clinical skills | 3.31 (0.71) | 3.50 (0.56) |
| Professionalism | 3.86 (0.43) | 3.97 (0.17) |
| Overall | 3.31 (0.72) | 3.42 (0.52) |
This tendency is demonstrated by the difference in overall grades. Of the 30 students with one A grade and one B grade, 18 students gave the lower grade than the faculty (60%). However, of the 13 students with one C or D grade and one higher grade, none gave the lower grade than the faculty grade (0%). Using Fisher's exact test, the difference between these proportions is statistically significant (p=.0005). The differences in the proportions for the knowledge and clinical skills scores are also statistically significant (p=.0075 and p=.019); for professionalism, too few low grades were reported to detect any differences. Figure 2 displays these differences graphically.
These data support formerly published studies that document low-performing students' overestimation of skills by self-assessment when compared to expert assessment. This was a voluntary sampling of students and it is possible that this was not a representative population. Objective data, including grade distribution and the number of participating students from each quartile, suggest that this was indeed a representative population. It is nevertheless possible that because participation in this study was voluntary only students who thought they were doing well volunteered, which would explain the low sensitivity. Conversely, it may be that the sample population had greater self-awareness than the larger population, which would have led to even more significant variation between self-assessed and instructor-assessed scores if the entire population had participated.
This study did not evaluate the reported effect of gender. A study evaluating human medical students in a surgery rotation documented the consistent overestimation of abilities by male students and consistent underestimation of abilities by female students.25 This might be the result of social norms suggesting that women are taught not be assertive or boastful.31 This effect was not evaluated in this study because students were identified only by number and because odds are very high that most of the participants were female considering the current gender inequity in veterinary medical training.
Expert evaluation has been proven to be most valuable when the student is assessed by a variety of instructors, all of whom are looking for different aspects of the skills described.7,17 In this study, while one clinician was the instructor of record for each rotation, the students' grades were compiled from individual evaluations of all the faculty and house officers with whom they had worked, a process that increased the accuracy of instructor assessment.
A related concern is the “halo effect,” defined as the tendency to rate students higher based on single attributes that receive undue weight. Because all of these students are chosen for admission from a much larger pool, there might be a tendency to rate them highly due to the instructors' innate belief in the high performance of all students in their program.4 This effect is especially significant to the general public, who expect experts to accurately identify individuals in health professional training who are not yet ready to practice. As a result of these same concerns, it is evident that students' self-assessment of competence is unsuitable as the primary or sole evidence of their skills. Again, public safety requires competence to be accurately determined before individuals are licensed to practice veterinary medicine.
Veterinary medical students need to develop the ability to recognize strengths and weaknesses in clinical practice in order to guide their clinical decision-making and career development, the latter of which includes continuing education chosen to enhance students' skills over the course of their entire career. As demonstrated here and elsewhere, self-assessment is unlikely to be an accurate indicator of true ability.
Valid self-assessment requires knowledge of the standards of expected performance and the ability to make informed judgments about how closely one's ability meets those standards.32 Timely instructor feedback coupled with questioning, reflection, and subsequent practice enhances self-assessment skills.6,13,14 Students should set specific goals against which they can judge their own progress.6,29 Opportunities for practice must be made readily available as frequency of performance is well correlated with accuracy of self-assessment.11
This study is the first report in veterinary medicine to demonstrate the same inaccuracy of student self-assessment as a measure of ability as has been demonstrated in other health professions. If student self-assessment is to be used as a measure of the levels of competence that are required to meet accreditation standards, certain specific measures to ensure the accuracy of student self-assessment must be employed, including active faculty mentoring of student self-assessment, student goal-setting and reflection, and availability of subsequent opportunities for practice.
| 1. | (2010). Accreditation policies and procedures of the AVMA Council on Education (COE) [Internet].cited 2010 Aug 6Schaumburg (IL):American Veterinary Medical AssociationAvailable from: http://www.avma.org/education/cvea/coe_self_study.asp Google Scholar |
| 2. | Wass, V, van der Vleuten, C, Shatzer, J, Jones, R (2001).Assessment of clinical competence.Lancet.35,945-, 9doi:https://doi.org/10.1016/S0140-6736(00)04221-5 Google Scholar |
| 3. | Miller, GE (1990).The assessment of clinical skills / competence / performance.Acad Med.65,563-, 7doi:https://doi.org/10.1097/00001888-199009000-00045 Google Scholar |
| 4. | Thompson, WG, Lipkin, M, Gilbert, DA, Guzzo, RA, Roberson, L (1990).Evaluating evaluation: assessment of the American Board of Internal Medicine resident evaluation form.J Gen Intern Med.5,214-, 7doi:https://doi.org/10.1007/BF02600537 Medline, Google Scholar |
| 5. | Weiss, PM, Koller, CA, Hess, W, Wasser, T (2005).How do medical student self-assessments compare with their final clerkship grades?.Med Teach.2,445-, 9doi:https://doi.org/10.1080/01421590500046999 Google Scholar |
| 6. | Gordon, MJ (1991).A review of the validity and accuracy of self-assessments in health professions training.Acad Med.66,762-, 9doi:https://doi.org/10.1097/00001888-199112000-00012 Medline, Google Scholar |
| 7. | Johnson, D, Cujec, B (1998).Comparison of self, nurse, and physician assessment of residents rotating through an intensive care unit.Crit Care Med.26,1811-6 Medline, Google Scholar |
| 8. | Arnold, L, Willoughby, TL, Calkins, EV (1985).Self-evaluation in undergraduate medical education: a longitudinal perspective.J Med Educ.60,21-, 8doi:https://doi.org/10.1097/00001888-198501000-00004 Medline, Google Scholar |
| 9. | Rudy, DW, Fejfar, MC, Griffith, CH, Wilson, JP (2001).Self- and peer-assessment in a first year communication and interviewing course.Eval Health Prof.24,436-, 45doi:https://doi.org/10.1177/01632780122035000 Medline, Google Scholar |
| 10. | Mossop, LH, Senior, A (2008).I'll show you mine if you show me yours! Portfolio design in two UK veterinary schools.J Vet Med Educ.35,599-, 606doi:https://doi.org/10.3138/jvme.35.4.599 Link, Google Scholar |
| 11. | Fincher, RME, Lewis, LA (1994).Learning, experience, and self-assessment of competence of third-year medical students in performing bedside procedures.Acad Med.69,291-, 5doi:https://doi.org/10.1097/00001888-199404000-00012 Medline, Google Scholar |
| 12. | Schneider, JR, Verta, MJ, Ryan, ER, Corcoran, JF, DaRosa, DA (2008).Patient assessment and management examination: lack of correlation between faculty assessment and resident self-assessment.Am J Surg.95,16-, 9doi:https://doi.org/10.1016/j.amjsurg.2007.08.050 Google Scholar |
| 13. | Rees, C, Shepherd, M (2005).Students' and assessors' attitudes toward students' self-assessment of their personal and professional behaviours.Med Educ.39,30-, 9doi:https://doi.org/10.1111/j.1365-2929.2004.02030.x Medline, Google Scholar |
| 14. | Musolino, GM (2006).Fostering reflective practice: self-assessment abilities of physical therapy students and entry-level graduates.J Allied Health.35,30-42 Medline, Google Scholar |
| 15. | Sidhu, RS, Vikis, E, Cheifetz, R, Phang, T (2006).Self-assessment during a 2-day laparoscopic colectomy course: can surgeons judge how well they are learning new skills?.Am J Surg.191,677-, 81doi:https://doi.org/10.1016/j.amjsurg.2006.01.041 Medline, Google Scholar |
| 16. | Davis, JD (2002).Comparison of faculty, peer, self, and nurse assessment of obstetrics and gynecology residents.Obstet Gynecol.99,647-, 51doi:https://doi.org/10.1016/S0029-7844(02)01658-7 Medline, Google Scholar |
| 17. | Bryan, RE, Krych, AJ, Carmichael, SW, Viggiano, TR, Pawlina, W (2005).Assessing professionalism in early medical education: experience with peer evaluation and self-evaluation in the gross anatomy course.Ann Acad Med Singap.34,486-91 Medline, Google Scholar |
| 18. | Edwards, RK, Kellner, KR, Sistrom, CL, Magyari, EJ (2003).Medical student self-assessment of performance on an obstetrics and gynecology clerkship.Am J Obstet Gynecol.188,1078-, 82doi:https://doi.org/10.1067/mob.2003.249 Medline, Google Scholar |
| 19. | Haidt, J (2006).The rose-colored mirror.The happiness hypothesis: finding modern truth in ancient wisdom.New York:, Basic Books66-69 Google Scholar |
| 20. | Zenner, D, Burns, GA, Ruby, KL, DeBowes, RM, Stoll, SK (2005).Veterinary students as elite performers: preliminary insights.J Vet Med Educ.32,242-, 8doi:https://doi.org/10.3138/jvme.32.2.242 Medline, Google Scholar |
| 21. | Gordon, MJ (1992).Self-assessment programs and their implications for health professions training.Acad Med.67,672-, 9doi:https://doi.org/10.1097/00001888-199210000-00012 Medline, Google Scholar |
| 22. | Mabe, PA, West, SG (1982).Validity of self-evaluation of ability: a review and meta-analysis.J Appl Psychol.67,280-, 96doi:https://doi.org/10.1037//0021-9010.67.3.280 Google Scholar |
| 23. | Stewart, J, O'Halloran, C, Barton, JR, Singleton, SJ, Harrigan, P, Spencer, J (2000).Clarifying the concepts of confidence and competence to produce appropriate self-evaluation measurement scales.Med Educ.34,903-, 9doi:https://doi.org/10.1046/j.1365-2923.2000.00728.x Medline, Google Scholar |
| 24. | Dunning, D, Johnson, K, Ehrlinger, J, Kruger, J (2003).Why people fail to recognize their own incompetence.Curr Dir Psychol Sci.12,83-, 7doi:https://doi.org/10.1111/1467-8721.01235 Google Scholar |
| 25. | Lind, DS, Rekkas, S, Bui, V, Lam, T, Beierle, E, Copeland, EM (2002).Competency-based student self-assessment on a surgery rotation.J Surg Res.105,31-, 34doi:https://doi.org/10.1006/jsre.2002.6442 Medline, Google Scholar |
| 26. | Donald, JB, Thomson, D (1993).Evaluation of computerized assessment package for general practitioner trainees.Brit J Gen Pract.43,115-8 Medline, Google Scholar |
| 27. | Wooliscroft, JO, TenHaken, J, Smith, J, Calhoun, JG (1993).Medical students' clinical self-assessments: comparisons with external measures of performance and the students' self-assessments of overall performance and effort.Acad Med.68,285-94 Medline, Google Scholar |
| 28. | Panek, PE, Sterns, HL (1985).Self-evaluation, actual performance and preference across the life-span.Exp Aging Res.11,221-3 Medline, Google Scholar |
| 29. | Brydges, R, Carnahan, H, Safir, O, Dubrowski, A (2009).How effective is self-guided learning of clinical technical skills? It's all about process.Med Educ.43,507-, 15doi:https://doi.org/10.1111/j.1365-2923.2009.03329.x Medline, Google Scholar |
| 30. | Violato, C, Lockyer, J (2006).Self and peer assessment of pediatricians, psychiatrists and medicine specialists: implications for self-directed learning.Adv Health Sci Educ.11,235-, 44doi:https://doi.org/10.1007/s10459-005-5639-0 Medline, Google Scholar |
| 31. | Tannen, D (1994).Talking from 9 to 5: women and men at work.New York:, Quill35-39 Google Scholar |
| 32. | Sargeant, J, Mann, K, van der Vleuten, C, Metsemakers, J (2008).“Directed” self-assessment: practice and feedback within a social context.J Contin Educ Health.28,47-, 54doi:https://doi.org/10.1002/chp.155 Medline, Google Scholar |
