1
|
Daghrery A, Alwadai GS, Alamoudi NA, Alqahtani SA, Alshehri FH, Al Wadei MH, Abogazalah NN, Pereira GKR, Al Moaleem MM. Students' performance in clinical class II composite restorations: a case study using analytic rubrics. BMC MEDICAL EDUCATION 2024; 24:1252. [PMID: 39497127 PMCID: PMC11536952 DOI: 10.1186/s12909-024-06261-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Accepted: 10/28/2024] [Indexed: 11/06/2024]
Abstract
BACKGROUND The analytical rubric serves as a permanent reference for guidelines on clinical performance for undergraduate dental students. This study aims to assess the rubric system used to evaluate clinical class II composite restorations performed by undergraduate dental students and to explore the impact of gender on overall student performance across two academic years. Additionally, we investigated the relationship between cumulative grade point averages (CGPAs) and students' clinical performance. METHODS An analytical rubric for the assessment of clinical class II composite restoration in the academic years of 2022/2023 and 2023/2024 was used by two evaluators. These two evaluators were trained to use the rubric before doing the evaluations. The scores were based on a 4-point scale for the evaluation of five major parameters for pre-operative procedures (10 points), cavity preparation (20 points), restoration procedures (20 points), and time management (4 points). At the same time, chairside oral exam parameter was 15 points based on a 5-point scale. Descriptive statistics were calculated for the different analytical rubric parameters, and the independent t-test was used to compare the scores between the student groups and the evaluators. Other tests, such as the Kappa test and Pearson's correlation coefficient, were used to measure the association among CGPA, evaluators, and gender participants. RESULTS The overall score out of 69 slightly increased for females/males (61.28/59.42) and (61.18/59.49) in the 2022/2023 and 2023/2024 academic years, respectively, but the differences were not statistically significant. In the 2022/2023 academic year, female students scored significantly higher than male students in pre-operative procedures, as evaluated by both evaluators (p = 0.001), and in time management, as assessed by both evaluators (p = 0.031). The Kappa test demonstrated a moderate to substantial level of agreement between the two evaluators in both academic years. Strong and significant correlations were noted between students' CGPA and some tested parameters (p = 0.000). CONCLUSION The overall performance was very good and high among both genders, but it was marginally higher among females than among males. This study found some differences in performance between male and female students and variability in the evaluations by the two raters ranging from moderate to substantial agreement and similar performances for students with different CGPA.
Collapse
Affiliation(s)
- Arwa Daghrery
- Department of Restorative Dental Sciences, College of Dentistry, Jazan University, Jazan, 45142, Saudi Arabia.
| | - Ghadeer Saleh Alwadai
- Restorative Dental Science, College of Dentistry, King Khalid University, Abha, Saudi Arabia
| | - Nada Ahmad Alamoudi
- Restorative Dental Science, College of Dentistry, King Khalid University, Abha, Saudi Arabia
| | - Saleh Ali Alqahtani
- Restorative Dental Science, College of Dentistry, King Khalid University, Abha, Saudi Arabia
| | - Faisal Hasan Alshehri
- Restorative Dental Science, College of Dentistry, King Khalid University, Abha, Saudi Arabia
| | | | - Naif Nabel Abogazalah
- Restorative Dental Science, College of Dentistry, King Khalid University, Abha, Saudi Arabia
| | - Gabriel Kalil Rocha Pereira
- Post-Graduate Program in Oral Sciences (Prosthodontics Unit), Faculty of Dentistry, Universidade Federal de Santa Maria (UFSM), Santa Maria, Rio Grande do Sul, Brazil
| | - Mohammed M Al Moaleem
- Department of Prosthetic Dental Science, College of Dentistry, Jazan University, Jazan, 45142, Saudi Arabia.
| |
Collapse
|
2
|
Yeo SW, Signorelli C, Vo K, Smith G. A Retrospective Cohort Analysis Comparing Analytic and Holistic Marking Rubrics in Medical Research Education. JOURNAL OF MEDICAL EDUCATION AND CURRICULAR DEVELOPMENT 2024; 11:23821205241277337. [PMID: 39211294 PMCID: PMC11359436 DOI: 10.1177/23821205241277337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Accepted: 08/07/2024] [Indexed: 09/04/2024]
Abstract
OBJECTIVES The use of analytic rubrics in assessing and grading students' performance has become more prominent among instructors due to its reliability and validity in ensuring consistency in student evaluation. However, there is limited evidence demonstrating the consistency of examiner judgments between analytic marking rubrics and holistic marking rubrics. METHODS Therefore, we aimed to compare the consistency of marks given using holistic marking methods and analytic rubrics at an Australian university by analyzing the mean mark differences and number of adjudications between two rubric types as well as the inter-rater reliability between two assessors. We analyzed all scores for project manuscripts between 2016 and 2021 for Honours medical students. We compared the mean mark differences graded using the Kruskal-Wallis test and Welch t-test. We used chi-squared tests to compare the frequency of adjudications for each rubric type. We assessed interrater reliability by comparing the marks between the two examiners utilizing Pearson correlation. RESULTS We found that analytic rubrics have lower mean mark differences and fewer adjudicators are required. We showed a strong positive association between the consistency of marks given and the use of analytic rubrics when compared to holistic marking. Pearson correlation showed a low but stronger correlation between marks awarded by the two assessors when analytic rubrics were used (r = 0.36), compared to holistic marking rubrics (r = 0.24). CONCLUSIONS Our findings suggest that the use of analytic rubrics may increase the consistency and reliability between two independent examiners in marking medical students' work.
Collapse
Affiliation(s)
- Siew Wan Yeo
- UNSW Medicine and Health, UNSW Sydney, Kensington, Australia
| | - Christina Signorelli
- UNSW Medicine and Health, UNSW Sydney, Kensington, Australia
- Kids Cancer Centre, Sydney Children's Hospital, Randwick, Australia
| | - Khanh Vo
- UNSW Medicine and Health, UNSW Sydney, Kensington, Australia
| | - Greg Smith
- UNSW Medicine and Health, UNSW Sydney, Kensington, Australia
| |
Collapse
|
3
|
Olvet DM, Bird JB, Fulton TB, Kruidering M, Papp KK, Qua K, Willey JM, Brenner JM. A Multi-institutional Study of the Feasibility and Reliability of the Implementation of Constructed Response Exam Questions. TEACHING AND LEARNING IN MEDICINE 2023; 35:609-622. [PMID: 35989668 DOI: 10.1080/10401334.2022.2111571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Accepted: 07/27/2022] [Indexed: 06/15/2023]
Abstract
PROBLEM Some medical schools have incorporated constructed response short answer questions (CR-SAQs) into their assessment toolkits. Although CR-SAQs carry benefits for medical students and educators, the faculty perception that the amount of time required to create and score CR-SAQs is not feasible and concerns about reliable scoring may impede the use of this assessment type in medical education. INTERVENTION Three US medical schools collaborated to write and score CR-SAQs based on a single vignette. Study participants included faculty question writers (N = 5) and three groups of scorers: faculty content experts (N = 7), faculty non-content experts (N = 6), and fourth-year medical students (N = 7). Structured interviews were performed with question writers and an online survey was administered to scorers to gather information about their process for creating and scoring CR-SAQs. A content analysis was performed on the qualitative data using Bowen's model of feasibility as a framework. To examine inter-rater reliability between the content expert and other scorers, a random selection of fifty student responses from each site were scored by each site's faculty content experts, faculty non-content experts, and student scorers. A holistic rubric (6-point Likert scale) was used by two schools and an analytic rubric (3-4 point checklist) was used by one school. Cohen's weighted kappa (κw) was used to evaluate inter-rater reliability. CONTEXT This research study was implemented at three US medical schools that are nationally dispersed and have been administering CR-SAQ summative exams as part of their programs of assessment for at least five years. The study exam question was included in an end-of-course summative exam during the first year of medical school. IMPACT Five question writers (100%) participated in the interviews and twelve scorers (60% response rate) completed the survey. Qualitative comments revealed three aspects of feasibility: practicality (time, institutional culture, teamwork), implementation (steps in the question writing and scoring process), and adaptation (feedback, rubric adjustment, continuous quality improvement). The scorers' described their experience in terms of the need for outside resources, concern about lack of expertise, and value gained through scoring. Inter-rater reliability between the faculty content expert and student scorers was fair/moderate (κw=.34-.53, holistic rubrics) or substantial (κw=.67-.76, analytic rubric), but much lower between faculty content and non-content experts (κw=.18-.29, holistic rubrics; κw=.59-.66, analytic rubric). LESSONS LEARNED Our findings show that from the faculty perspective it is feasible to include CR-SAQs in summative exams and we provide practical information for medical educators creating and scoring CR-SAQs. We also learned that CR-SAQs can be reliably scored by faculty without content expertise or senior medical students using an analytic rubric, or by senior medical students using a holistic rubric, which provides options to alleviate the faculty burden associated with grading CR-SAQs.
Collapse
Affiliation(s)
- Doreen M Olvet
- Department of Science Education, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, New York, USA
| | - Jeffrey B Bird
- Department of Science Education, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, New York, USA
| | - Tracy B Fulton
- Department of Biochemistry and Biophysics, University of California San Francisco School of Medicine, San Francisco, California, USA
| | - Marieke Kruidering
- Department of Cellular & Molecular Pharmacology, University of California at San Francisco School of Medicine, San Francisco, California, USA
| | - Klara K Papp
- Center for Medical Education, Case Western Reserve University School of Medicine, Cleveland, Ohio, USA
| | - Kelli Qua
- Research and Evaluation, Case Western Reserve University School of Medicine, Cleveland, Ohio, USA
| | - Joanne M Willey
- Department of Science Education, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, New York, USA
| | - Judith M Brenner
- Department of Science Education, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, New York, USA
| |
Collapse
|
4
|
Kotlyar I, Sharifi T, Fiksenbaum L. Assessing Teamwork Skills: Can a Computer Algorithm Match Human Experts? INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION 2022. [DOI: 10.1007/s40593-022-00318-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
5
|
Thomas L, Rosenberg I, R Weinstein A, Maffeo C, Feinn R. Implementation and Inter-Rater Reliability of a Comprehensive Scoring Rubric for Preclerkship Medical Documentation. ACADEMIC MEDICINE : JOURNAL OF THE ASSOCIATION OF AMERICAN MEDICAL COLLEGES 2022; 97:S145-S146. [PMID: 37838878 DOI: 10.1097/acm.0000000000004961] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/20/2023]
Affiliation(s)
- Listy Thomas
- Author affiliations: L. Thomas, I. Rosenberg, A.R. Weinstein, C. Maffeo, R. Feinn, Frank H. Netter MD School of Medicine at Quinnipiac University
| | | | | | | | | |
Collapse
|
6
|
Gilani S, Pankhania K, Aruketty M, Naeem F, Alkhayyat A, Akhtar U, Chaudhary M, Sinha S. Twelve tips to organise a mock OSCE. MEDICAL TEACHER 2022; 44:26-31. [PMID: 33656952 DOI: 10.1080/0142159x.2021.1887465] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The objective structured clinical examination (OSCE) is central to assessing clinical competence in undergraduate and postgraduate exams for medical and allied healthcare professions. A mock OSCE on the other hand is a simulation of the OSCE and a unique learning experience for the examinee. They benefit in a variety of ways; from enhancing their time management skills to receiving feedback that can improve their clinical skills. Unfortunately, opportunities to participate in simulated OSCEs remain limited. Reasons include difficulty in fulfilling organisational requirements and equipment-related costs. Mock OSCEs can be set up by undergraduate students or junior trainees for peers, without senior supervision or direct guidance. This article will discuss 12 tips regarding the arrangement of mock OSCEs to guide organisers, including accessing resources and establishing the content of the exam.
Collapse
Affiliation(s)
- Syed Gilani
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
- Lancashire Teaching Hospitals NHS Trust, Lancashire, UK
| | - Kishan Pankhania
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
| | - Maneesha Aruketty
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
| | - Farah Naeem
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
| | - Abdulaziz Alkhayyat
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
| | - Usmaan Akhtar
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
- Royal Bolton Hospital, Bolton, UK
| | - Muhammad Chaudhary
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
- Macclesfield District General Hospital, Manchester, UK
| | - Siddharth Sinha
- Faculty of Biology, Medicine and Health, University of Manchester, Manchester, UK
- King's College Hospital, London, UK
| |
Collapse
|
7
|
Virk A, Joshi A, Mahajan R, Singh T. The power of subjectivity in competency-based assessment. J Postgrad Med 2020; 66:200-205. [PMID: 33037168 PMCID: PMC7819378 DOI: 10.4103/jpgm.jpgm_591_20] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
With the introduction of competency-based undergraduate curriculum in India, a paradigm shift in the assessment methods and tools will be the need of the hour. Competencies are complex combinations of various attributes, many of which being not assessable by objective methods. Assessment of affective and communication domains has always been neglected for want of objective methods. Areas like professionalism, ethics, altruism, and communication—so vital for being an Indian Medical Graduate, can be assessed longitudinally applying subjective means only. Though subjectivity has often been questioned as being biased, it has been proven time and again that a subjective assessment in expert hands gives comparable results as that of any objective assessment. By insisting on objectivity, we may compromise the validity of the assessment and deprive the students of enriched subjective feedback and judgement also. This review highlights the importance of subjective assessment in competency-based assessment and ways and means of improving the rigor of subjective assessment, with particular emphasis on the development and use of rubrics.
Collapse
Affiliation(s)
- A Virk
- Adesh Medical College & Hospital, Shahabad (M), Haryana, India
| | - A Joshi
- Pramukhswami Medical College, Karamsad, Gujarat, India
| | - R Mahajan
- Adesh Institute of Medical Sciences & Research, Bathinda, Punjab, India
| | - T Singh
- SGRD Institute of Medical Sciences and Research, Amritsar, Punjab, India
| |
Collapse
|
8
|
Sripathi KN, Moscarella RA, Yoho R, You HS, Urban-Lurain M, Merrill J, Haudek K. Mixed Student Ideas about Mechanisms of Human Weight Loss. CBE LIFE SCIENCES EDUCATION 2019; 18:ar37. [PMID: 31418653 PMCID: PMC6755313 DOI: 10.1187/cbe.18-11-0227] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/21/2018] [Revised: 04/08/2019] [Accepted: 04/30/2019] [Indexed: 06/10/2023]
Abstract
Recent calls for college biology education reform have identified "pathways and transformations of matter and energy" as a big idea in biology crucial for students to learn. Previous work has been conducted on how college students think about such matter-transforming processes; however, little research has investigated how students connect these ideas. Here, we probe student thinking about matter transformations in the familiar context of human weight loss. Our analysis of 1192 student constructed responses revealed three scientific (which we label "Normative") and five less scientific (which we label "Developing") ideas that students use to explain weight loss. Additionally, students combine these ideas in their responses, with an average number of 2.19 ± 1.07 ideas per response, and 74.4% of responses containing two or more ideas. These results highlight the extent to which students hold multiple (both correct and incorrect) ideas about complex biological processes. We described student responses as conforming to either Scientific, Mixed, or Developing descriptive models, which had an average of 1.9 ± 0.6, 3.1 ± 0.9, and 1.7 ± 0.8 ideas per response, respectively. Such heterogeneous student thinking is characteristic of difficulties in both conceptual change and early expertise development and will require careful instructional intervention for lasting learning gains.
Collapse
Affiliation(s)
- Kamali N. Sripathi
- CREATE for STEM Institute, Michigan State University, East Lansing, MI 48824
| | - Rosa A. Moscarella
- Biology Department, University of Massachusetts Amherst, Amherst, MA 01003
| | - Rachel Yoho
- Project Dragonfly and Department of Biology, Miami University, Oxford, OH 45056
| | - Hye Sun You
- Department of Physical Sciences, Arkansas Tech University, Russellville, AR 72801
| | - Mark Urban-Lurain
- CREATE for STEM Institute, Michigan State University, East Lansing, MI 48824
| | - John Merrill
- Department of Microbiology and Molecular Genetics, Michigan State University, East Lansing, MI 48824
| | - Kevin Haudek
- Department of Biochemistry and Molecular Biology, Michigan State University, East Lansing, MI 48824
| |
Collapse
|
9
|
Systems Thinking for Transitions of Care: Reliability Testing for a Standardized Rubric. CLIN NURSE SPEC 2019; 33:128-135. [PMID: 30946110 DOI: 10.1097/nur.0000000000000443] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
PURPOSE The purpose of this study was to develop a standardized rubric for systems thinking across transitions of care for clinical nurse specialists. DESIGN The design was a mixed-methods study using the Systems Awareness Model as a framework for bridging theory to practice. METHODS Content validity was determined using a content validity index. Reliability was established using statistical analysis with Cronbach's α and intraclass correlation coefficient. Usability of the rubric was established using content analysis from focus group discussions about their experiences in using the rubric. RESULTS Content validity was established with a content validity ratio of 1.0. Statistical analysis showed a high interrater reliability (α = 0.99), and sections of the rubric showed a strong degree of reliability with α's ranging from 0.88 to 1.00. Content analysis revealed several overall themes for usability of the rubric: clarity, objectivity, and detail. The area for improvement included adding more detail in the scholarly writing section. CONCLUSION The research team recommends using the rubric to reflect application of systems thinking across transitions of care.
Collapse
|