1
|
Desy J, Harvey A, Martin K, Naugler C, McLaughlin K. Giving partial credit during a multiple-choice question assessment reappraisal does not make the assessment process fairer. CANADIAN MEDICAL EDUCATION JOURNAL 2024; 15:95-96. [PMID: 38827902 PMCID: PMC11139799 DOI: 10.36834/cmej.77957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 06/05/2024]
Affiliation(s)
- Janeve Desy
- Office of Undergraduate Medical Education, Cumming School of Medicine, University of Calgary, Alberta, Canada
| | - Adrian Harvey
- Office of Undergraduate Medical Education, Cumming School of Medicine, University of Calgary, Alberta, Canada
| | - Kerri Martin
- Office of Undergraduate Medical Education, Cumming School of Medicine, University of Calgary, Alberta, Canada
| | - Christopher Naugler
- Office of Undergraduate Medical Education, Cumming School of Medicine, University of Calgary, Alberta, Canada
| | - Kevin McLaughlin
- Office of Undergraduate Medical Education, Cumming School of Medicine, University of Calgary, Alberta, Canada
| |
Collapse
|
2
|
Progress testing: An educational perspective exploring the rationale for progress testing and its introduction into a Diagnostic Radiography curriculum. J Med Imaging Radiat Sci 2023; 54:35-42. [PMID: 36681618 DOI: 10.1016/j.jmir.2022.12.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2022] [Revised: 12/06/2022] [Accepted: 12/13/2022] [Indexed: 01/22/2023]
Abstract
INTRODUCTION In March 2020, the first diagnostic radiography degree apprenticeship programme in England was launched at the authors' institution. As part of the programme development and design, the programme development team explored and then implemented progress testing into a strand of the programme. The objective of this educational perspective is to scrutinise the literature around the use of progress testing in higher education programmes, namely medicine, to explain how and why this decision was reached. METHODS The initial search strategy was developed using the electronic databases CINHAL Complete and SCOPUS. Key words included 'progress test' and 'medicine' or 'health' or 'education' or 'higher education'. Eliminating articles that were not relevant, and also identifying and adding additional articles by key authors and experts resulted in thirty-three key articles being considered for review. RESULTS The thirty-three articles were a mixture of review articles, empirical research, case studies and conference presentations. Five key themes were identified which are discussed in this article; the evolution of progress testing; advantages of progress testing, disadvantages of progress testing, developing a test framework and academic progression and student feedback. DISCUSSION Progress testing is now well established in pre-registration medical programmes globally. The advantages of progress testing and the use of frequent look rapid remediation appear to be undisputed. Key disadvantages with progress testing were identified as it being an administrative heavy assessment process as well as a perceived bias towards male students undertaking this type of assessment. CONCLUSION Now this assessment practice is established within medicine, it seems reasonable to explore its use in other areas of healthcare, such as radiography.
Collapse
|
3
|
Goggs R, Kerl M, Jandrey KE, Guillaumin J. Prospective investigation of factors associated with success on the American College of Veterinary Emergency and Critical Care certification examination (2016-2018). J Vet Emerg Crit Care (San Antonio) 2021; 32:196-206. [PMID: 34714977 DOI: 10.1111/vec.13153] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2020] [Revised: 08/30/2020] [Accepted: 08/31/2020] [Indexed: 11/27/2022]
Abstract
OBJECTIVE To assess the association of candidate attributes and residency training factors with success on the American College of Veterinary Emergency and Critical Care (ACVECC) board certification examination and to develop multivariable models of first-attempt success. DESIGN Prospective survey-based study. SETTING Post-assessment ACVECC examination candidates. ANIMALS None. INTERVENTIONS None. MEASUREMENTS AND MAIN RESULTS Comprehensive surveys were distributed to ACVECC examination candidates in 2016 to 2018 after completion of their assessments, but prior to publication of examination results. Unique anonymous candidate identification numbers were used to match survey responses to outcome data from the office of the ACVECC Executive Secretary. After curation to retain only the first response from each candidate, there were 97 unique candidate responses available for analysis. Univariate analyses identified multiple factors associated with first-attempt success and multiple differences between academic and private practice residency programs. Multivariable logistic regression modeling suggested that 5 factors were independently associated with first-attempt success on the ACVECC examination, specifically younger age, more weeks of study prior to the examination, training at a facility with more ACVECC Diplomates, training at a facility with more ACVECC residents, and having no requirement to manage both Emergency Room (ER) and Critical Care (CC) cases simultaneously. CONCLUSIONS Numerous resident and training center factors are associated with success in the ACVECC board certification examination. Residents and training centers might be able to use these data to enhance training, but caution must be exercised because these data are associative only.
Collapse
Affiliation(s)
- Robert Goggs
- Department of Clinical Sciences, Cornell University College of Veterinary Medicine, Ithaca, New York, USA
| | - Marie Kerl
- Regional Operations, Heartland Group, VCA Inc., Los Angeles, California, USA.,Department of Veterinary Medicine and Surgery, College of Veterinary Medicine, University of Missouri, Columbia, Missouri, USA
| | - Karl E Jandrey
- Department of Surgical and Radiological Sciences, University of California, Davis, California, USA
| | - Julien Guillaumin
- Department of Clinical Science, College of Veterinary Medicine and Biomedical Sciences, Colorado State University, Fort Collins, Colorado, USA
| |
Collapse
|
4
|
Interdimensional Travel: Visualisation of 3D-2D Transitions in Anatomy Learning. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2020; 1235:103-116. [PMID: 32488638 DOI: 10.1007/978-3-030-37639-0_6] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
Clinical image interpretation is one of the most challenging activities for students when they first arrive at medical school. Interpretation of clinical images concerns the identification of three-dimensional anatomical features in two-dimensional cross-sectional computed tomography (CT) and magnetic resonance imaging (MRI) images in axial, sagittal and coronal planes, and the recognition of structures in ultrasound and plain radiographs. We propose that a cognitive transition occurs when initially attempting to interpret clinical images, which requires reconciling known 3D structures with previously unknown 2D visual information. Additionally, we propose that this 3D-2D transition is required when integrating an understanding of superficial 2D surface landmarks with an appreciation of underlying 3D anatomical structures during clinical examinations.Based on educational theory and research findings, we recommend that 3D and 2D approaches should be simultaneously combined within radiological and surface anatomy education. With a view to this, we have developed and utilised digital and art-based methods to support the 3D-2D transition. We outline our observations and evaluations, and describe our practical implementation of these approaches within medical curricula to serve as a guide for anatomy educators. Furthermore, we define the theoretical underpinnings and evidence supporting the integration of 3D-2D approaches and the value of our specific activities for enhancing the clinical image interpretation and surface anatomy learning of medical students.
Collapse
|
5
|
Integrating 3D Visualisation Technologies in Undergraduate Anatomy Education. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2019; 1120:39-53. [PMID: 30919293 DOI: 10.1007/978-3-030-06070-1_4] [Citation(s) in RCA: 34] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
Anatomy forms the basis of clinical examination, surgery and radiology and it is therefore essential that a fundamental understanding of the key concepts, structures and their relationships is gained by medical and healthcare students during their undergraduate training. Anatomy involves the study of three dimensional entities relating to the human body itself and its constituent structures. In our experience, the appreciation of 3D concepts is one of the most demanding areas for medical student learning of anatomy (ben Awadh et al. 2018, unpublished observations). The ability to interpret 3D anatomical features in 2D cross-sectional clinical images can be troublesome, while the dynamic nature of embryological development is a further challenge.The aim of introducing technology enhanced-learning (TEL) approaches into our practice is with a view to enhancing undergraduate medical student learning of clinically relevant anatomy. Here we will explore the importance of visualisation and visual learning in anatomy as a scholarly basis for the integration for TEL approaches. We will then describe examples of visualisation technologies that are currently being implemented within the School of Medical Education at Newcastle University based on a research informed understanding of how students learn anatomy. We will consider the available evidence that supports best practice, identify limitations where they arise, and discuss how these visual 3D learning technologies can be effectively utilised as adjuncts and self-directed resources to supplement more established approaches to undergraduate anatomy education.
Collapse
|
6
|
Scherer JJ, Reichrath S, Tschernig T, Maxeiner S. A snapshot of anatomy teaching: Peer-led anatomy boot camps — Priming for success? Ann Anat 2019; 221:141-147. [DOI: 10.1016/j.aanat.2018.10.005] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2018] [Revised: 10/01/2018] [Accepted: 10/02/2018] [Indexed: 10/28/2022]
|
7
|
Booth TC, Martins RDM, McKnight L, Courtney K, Malliwal R. The Fellowship of the Royal College of Radiologists (FRCR) examination: a review of the evidence. Clin Radiol 2018; 73:992-998. [PMID: 30322706 PMCID: PMC6506684 DOI: 10.1016/j.crad.2018.09.005] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2018] [Accepted: 09/20/2018] [Indexed: 10/29/2022]
Abstract
The Fellowship of the Royal College of Radiologists (FRCR) examination is the professional qualification that is essential for career progression in clinical radiology within the UK. It is also important for career progression in many countries internationally. The FRCR has evolved and changed over the last decade. In this systematic review we appraise and summarise the available data relating to the FRCR emphasising the published evidence regarding the validity, reliability, and acceptability of this examination. Comparison is made to other equivalent medical examinations, as well as a more recently published commissioned external review of the FRCR examinations. The Clinical Radiology Part 2B (CR2B) examination in its pre-existing format is reliable, valid, and acceptable. Recommendations from the commissioned external review are based primarily on expert opinion, with a limited evidence base comprising data from a small sample acquired during a single examination sitting and without peer review. Unlike the CR2B examination, there is little evidence regarding assessment of the CR1 and CR2 examinations. Both the CR1 and CR2 examinations are currently in the process of undergoing major changes to their formats. Blueprinting items to the curriculum might improve acceptability. Other changes may improve transparency and reliability of these assessments. Our analysis and many aspects of the external review may provide pointers regarding how the upcoming data produced by the "automated" FRCR examinations can be further analysed to provide a more robust evaluation.
Collapse
Affiliation(s)
- T C Booth
- Department of Neuroradiology, King's College Hospital NHS Foundation Trust, London, SE5 9RS, UK; School of Biomedical Engineering and Imaging, King's College London, Rayne Institute, 4th Floor, Lambeth Wing, St. Thomas' Hospital, SE1 7EH, London.
| | - R deP M Martins
- Department of Radiology, Queen Alexandra Hospital, Portsmouth PO6 3LY, UK
| | - L McKnight
- Department of Radiology, Morrison Hospital, Swansea SA6 6NL, UK
| | - K Courtney
- Department of Radiology, King's College Hospital NHS Foundation Trust, London SE5 9RS, UK
| | - R Malliwal
- Department of Radiology, King's College Hospital NHS Foundation Trust, London SE5 9RS, UK
| |
Collapse
|
8
|
Hope D, Adamson K, McManus IC, Chis L, Elder A. Using differential item functioning to evaluate potential bias in a high stakes postgraduate knowledge based assessment. BMC MEDICAL EDUCATION 2018; 18:64. [PMID: 29615016 PMCID: PMC5883583 DOI: 10.1186/s12909-018-1143-0] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/19/2017] [Accepted: 03/08/2018] [Indexed: 05/30/2023]
Abstract
BACKGROUND Fairness is a critical component of defensible assessment. Candidates should perform according to ability without influence from background characteristics such as ethnicity or sex. However, performance differs by candidate background in many assessment environments. Many potential causes of such differences exist, and examinations must be routinely analysed to ensure they do not present inappropriate progression barriers for any candidate group. By analysing the individual questions of an examination through techniques such as Differential Item Functioning (DIF), we can test whether a subset of unfair questions explains group-level differences. Such items can then be revised or removed. METHODS We used DIF to investigate fairness for 13,694 candidates sitting a major international summative postgraduate examination in internal medicine. We compared (a) ethnically white UK graduates against ethnically non-white UK graduates and (b) male UK graduates against female UK graduates. DIF was used to test 2773 questions across 14 sittings. RESULTS Across 2773 questions eight (0.29%) showed notable DIF after correcting for multiple comparisons: seven medium effects and one large effect. Blinded analysis of these questions by a panel of clinician assessors identified no plausible explanations for the differences. These questions were removed from the question bank and we present them here to share knowledge of questions with DIF. These questions did not significantly impact the overall performance of the cohort. Group-level differences in performance between the groups we studied in this examination cannot be explained by a subset of unfair questions. CONCLUSIONS DIF helps explore fairness in assessment at the question level. This is especially important in high-stakes assessment where a small number of unfair questions may adversely impact the passing rates of some groups. However, very few questions exhibited notable DIF so differences in passing rates for the groups we studied cannot be explained by unfairness at the question level.
Collapse
Affiliation(s)
- David Hope
- Centre for Medical Education, The Chancellor’s Building, College of Medicine and Veterinary Medicine, The University of Edinburgh, 49 Little France Crescent, Edinburgh, Scotland EH16 4SB UK
| | - Karen Adamson
- Medical Unit, St John’s Hospital, Livingston, Scotland EH54 6PP UK
| | - I. C. McManus
- Research Department of Medical Education, The Medical School, University College London, Gower Street, London, WC1E 6BT UK
| | - Liliana Chis
- MRCPUK Central Office, 11 St Andrew’s Place, Regent’s Park, London, NW1 4LE UK
| | - Andrew Elder
- MRCPUK Central Office, 11 St Andrew’s Place, Regent’s Park, London, NW1 4LE UK
| |
Collapse
|
9
|
Cecilio-Fernandes D, Medema H, Collares CF, Schuwirth L, Cohen-Schotanus J, Tio RA. Comparison of formula and number-right scoring in undergraduate medical training: a Rasch model analysis. BMC MEDICAL EDUCATION 2017; 17:192. [PMID: 29121888 PMCID: PMC5679154 DOI: 10.1186/s12909-017-1051-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/16/2017] [Accepted: 11/02/2017] [Indexed: 05/10/2023]
Abstract
BACKGROUND Progress testing is an assessment tool used to periodically assess all students at the end-of-curriculum level. Because students cannot know everything, it is important that they recognize their lack of knowledge. For that reason, the formula-scoring method has usually been used. However, where partial knowledge needs to be taken into account, the number-right scoring method is used. Research comparing both methods has yielded conflicting results. As far as we know, in all these studies, Classical Test Theory or Generalizability Theory was used to analyze the data. In contrast to these studies, we will explore the use of the Rasch model to compare both methods. METHODS A 2 × 2 crossover design was used in a study where 298 students from four medical schools participated. A sample of 200 previously used questions from the progress tests was selected. The data were analyzed using the Rasch model, which provides fit parameters, reliability coefficients, and response option analysis. RESULTS The fit parameters were in the optimal interval ranging from 0.50 to 1.50, and the means were around 1.00. The person and item reliability coefficients were higher in the number-right condition than in the formula-scoring condition. The response option analysis showed that the majority of dysfunctional items emerged in the formula-scoring condition. CONCLUSIONS The findings of this study support the use of number-right scoring over formula scoring. Rasch model analyses showed that tests with number-right scoring have better psychometric properties than formula scoring. However, choosing the appropriate scoring method should depend not only on psychometric properties but also on self-directed test-taking strategies and metacognitive skills.
Collapse
Affiliation(s)
- Dario Cecilio-Fernandes
- Center for Education Development and Research in Health Professions (CEDAR), University of Groningen and University Medical Center Groningen, Antonius Deusinglaan 1, FC40, 9713 AV Groningen, The Netherlands
| | - Harro Medema
- Department Business IT & Management, NHL University of Applied Sciences, Leeuwarden, Netherlands
| | - Carlos Fernando Collares
- Faculty of Health, Medicine and Life Sciences, Educational Development and Research, Maastricht University, Maastricht, Netherlands
| | - Lambert Schuwirth
- Faculty of Health, Medicine and Life Sciences, Educational Development and Research, Maastricht University, Maastricht, Netherlands
- Prideaux Centre for Research into Health Professions Education, Flinders University, Adelaide, Australia
| | - Janke Cohen-Schotanus
- Center for Education Development and Research in Health Professions (CEDAR), University of Groningen and University Medical Center Groningen, Antonius Deusinglaan 1, FC40, 9713 AV Groningen, The Netherlands
| | - René A. Tio
- Center for Education Development and Research in Health Professions (CEDAR), University of Groningen and University Medical Center Groningen, Antonius Deusinglaan 1, FC40, 9713 AV Groningen, The Netherlands
| |
Collapse
|
10
|
Backhouse M, Fitzpatrick M, Hutchinson J, Thandi CS, Keenan ID. Improvements in anatomy knowledge when utilizing a novel cyclical "Observe-Reflect-Draw-Edit-Repeat" learning process. ANATOMICAL SCIENCES EDUCATION 2017; 10:7-22. [PMID: 27164484 DOI: 10.1002/ase.1616] [Citation(s) in RCA: 44] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/04/2015] [Revised: 04/06/2016] [Accepted: 04/07/2016] [Indexed: 06/05/2023]
Abstract
Innovative educational strategies can provide variety and enhance student learning while addressing complex logistical and financial issues facing modern anatomy education. Observe-Reflect-Draw-Edit-Repeat (ORDER), a novel cyclical artistic process, has been designed based on cognitivist and constructivist learning theories, and on processes of critical observation, reflection and drawing in anatomy learning. ORDER was initially investigated in the context of a compulsory first year surface anatomy practical (ORDER-SAP) at a United Kingdom medical school in which a cross-over trial with pre-post anatomy knowledge testing was utilized and student perceptions were identified. Despite positive perceptions of ORDER-SAP, medical student (n = 154) pre-post knowledge test scores were significantly greater (P < 0.001) with standard anatomy learning methods (3.26, SD = ±2.25) than with ORDER-SAP (2.17, ±2.30). Based on these findings, ORDER was modified and evaluated in the context of an optional self-directed gross anatomy online interactive tutorial (ORDER-IT) for participating first year medical students (n = 55). Student performance was significantly greater (P < 0.001) with ORDER-IT (2.71 ± 2.17) when compared to a control tutorial (1.31 ± 2.03). Performances of students with visual and artistic preferences when using ORDER were not significantly different (P > 0.05) to those students without these characteristics. These findings will be of value to anatomy instructors seeking to engage students from diverse learning backgrounds in a research-led, innovative, time and cost-effective learning method, in the context of contrasting learning environments. Anat Sci Educ 10: 7-22. © 2016 American Association of Anatomists.
Collapse
Affiliation(s)
- Mark Backhouse
- Department of Anatomy and Clinical Skills, School of Medical Education, Newcastle University, Newcastle Upon Tyne, United Kingdom
- Department of Cardiology, Southmead Hospital, North Bristol NHS Foundation Trust, Westbury-on Trym, Bristol, United Kingdom
| | - Michael Fitzpatrick
- Department of Anatomy and Clinical Skills, School of Medical Education, Newcastle University, Newcastle Upon Tyne, United Kingdom
- Department of Trauma and Orthopedic Surgery, Bristol Royal Infirmary, University Hospitals Bristol NHS Foundation Trust, Bristol, United Kingdom
| | - Joseph Hutchinson
- Department of Anatomy and Clinical Skills, School of Medical Education, Newcastle University, Newcastle Upon Tyne, United Kingdom
| | - Charankumal S Thandi
- Department of Anatomy and Clinical Skills, School of Medical Education, Newcastle University, Newcastle Upon Tyne, United Kingdom
| | - Iain D Keenan
- Department of Anatomy and Clinical Skills, School of Medical Education, Newcastle University, Newcastle Upon Tyne, United Kingdom
| |
Collapse
|
11
|
Richens D, Graham TR, James J, Till H, Turner PG, Featherstone C. Racial and Gender Influences on Pass Rates for the UK and Ireland Specialty Board Examinations. JOURNAL OF SURGICAL EDUCATION 2016; 73:143-150. [PMID: 26319105 DOI: 10.1016/j.jsurg.2015.08.003] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/01/2015] [Revised: 07/24/2015] [Accepted: 08/05/2015] [Indexed: 06/04/2023]
Abstract
INTRODUCTION We explored effects of gender, ethnic origin, first language, and training status on scores in the Intercollegiate Specialty Board examinations in the UK and Ireland across the computer-marked written section and in the face-to-face oral and clinical section. METHODS Demographic characteristics and examination results from 9987 attempts across 177 sittings from 2009 to 2013 were analyzed in an analysis of variance by training status, gender, ethnic origin, first language, and section (computer-marked multiple-choice examination vs face-to-face oral and clinical examination). RESULTS We found increasing alignment between examiner and candidate characteristics during this period, with a 50% increase in examiners of Asian ethnic origin and a 60% increase in examiners whose first language is not English. The strongest factor in the analysis of variance was training status (F[2, 9818] = 27.67, p < 0.001), with candidates in training significantly outperforming others. Within "core candidates" (first attempt, in training), we found significant main effects for ethnic origin (F[5, 4809] = 2.36, p = 0.04), and first language (F[2, 4809] = 5.29, p = 0.003), but no interaction effects between these factors and section (both F < 1, p > 0.05). CONCLUSIONS Training status was the most important factor in candidates' results. Although the analysis showed significant effects of ethnic origin and first language within "core candidates," these differences were statistically indistinguishable between the 2 sections of the examination, suggesting that the differential attainment by these factors cannot be attributed to examiner bias in a face-to-face examination.
Collapse
Affiliation(s)
- David Richens
- Joint Committee on Intercollegiate Examinations, Edinburgh, United Kingdom.
| | - Timothy R Graham
- Joint Committee on Intercollegiate Examinations, Edinburgh, United Kingdom
| | - John James
- The Sickle Cell Society, London, United Kingdom
| | - Hettie Till
- Assessment Support and Psychometric Services, Dundee, United Kingdom
| | - Philip G Turner
- Joint Committee on Intercollegiate Examinations, Edinburgh, United Kingdom
| | - Cara Featherstone
- Joint Committee on Intercollegiate Examinations, Edinburgh, United Kingdom
| |
Collapse
|
12
|
Ndu IK, Ekwochi U, Di Osuorah C, Asinobi IN, Nwaneri MO, Uwaezuoke SN, Amadi OF, Okeke IB, Chinawa JM, Orjioke CJG. Negative Marking and the Student Physician--A Descriptive Study of Nigerian Medical Schools. JOURNAL OF MEDICAL EDUCATION AND CURRICULAR DEVELOPMENT 2016; 3:10.4137_JMECD.S40705. [PMID: 29349304 PMCID: PMC5736262 DOI: 10.4137/jmecd.s40705] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/05/2016] [Revised: 10/13/2016] [Accepted: 10/15/2016] [Indexed: 06/07/2023]
Abstract
BACKGROUND There is considerable debate about the two most commonly used scoring methods, namely, the formula scoring (popularly referred to as negative marking method in our environment) and number right scoring methods. Although the negative marking scoring system attempts to discourage students from guessing in order to increase test reliability and validity, there is the view that it is an excessive and unfair penalty that also increases anxiety. Feedback from students is part of the education process; thus, this study assessed the perception of medical students about negative marking method for multiple choice question (MCQ) examination formats and also the effect of gender and risk-taking behavior on scores obtained with this assessment method. METHODS This was a prospective multicenter survey carried out among fifth year medical students in Enugu State University and the University of Nigeria. A structured questionnaire was administered to 175 medical students from the two schools, while a class test was administered to medical students from Enugu State University. Qualitative statistical methods including frequencies, percentages, and chi square were used to analyze categorical variables. Quantitative statistics using analysis of variance was used to analyze continuous variables. RESULTS Inquiry into assessment format revealed that most of the respondents preferred MCQs (65.9%). One hundred and thirty students (74.3%) had an unfavorable perception of negative marking. Thirty-nine students (22.3%) agreed that negative marking reduces the tendency to guess and increases the validity of MCQs examination format in testing knowledge content of a subject compared to 108 (61.3%) who disagreed with this assertion (χ2 = 23.0, df = 1, P = 0.000). The median score of the students who were not graded with negative marking was significantly higher than the score of the students graded with negative marking (P = 0.001). There was no statistically significant difference in the risk-taking behavior between male and female students in their MCQ answering patterns with negative marking method (P = 0.618). CONCLUSIONS In the assessment of students, it is more desirable to adopt fair penalties for discouraging guessing rather than excessive penalties for incorrect answers, which could intimidate students in negative marking schemes. There is no consensus on the penalty for an incorrect answer. Thus, there is a need for continued research into an effective and objective assessment tool that will ensure that the students' final score in a test truly represents their level of knowledge.
Collapse
Affiliation(s)
| | - Uchenna Ekwochi
- Enugu State University Teaching Hospital, Parklane, Enugu, Nigeria
| | | | | | | | | | | | | | | | | |
Collapse
|
13
|
Ravesloot CJ, Van der Schaaf MF, Muijtjens AMM, Haaring C, Kruitwagen CLJJ, Beek FJA, Bakker J, Van Schaik JPJ, Ten Cate TJ. The don't know option in progress testing. ADVANCES IN HEALTH SCIENCES EDUCATION : THEORY AND PRACTICE 2015; 20:1325-38. [PMID: 25912621 PMCID: PMC4639571 DOI: 10.1007/s10459-015-9604-2] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/30/2014] [Accepted: 03/25/2015] [Indexed: 05/12/2023]
Abstract
Formula scoring (FS) is the use of a don't know option (DKO) with subtraction of points for wrong answers. Its effect on construct validity and reliability of progress test scores, is subject of discussion. Choosing a DKO may not only be affected by knowledge level, but also by risk taking tendency, and may thus introduce construct-irrelevant variance into the knowledge measurement. On the other hand, FS may result in more reliable test scores. To evaluate the impact of FS on construct validity and reliability of progress test scores, a progress test for radiology residents was divided into two tests of 100 parallel items (A and B). Each test had a FS and a number-right (NR) version, A-FS, B-FS, A-NR, and B-NR. Participants (337) were randomly divided into two groups. One group took test A-FS followed by B-NR, and the second group test B-FS followed by A-NR. Evidence for impaired construct validity was sought in a hierarchical regression analysis by investigating how much of the participants' FS-score variance was explained by the DKO-score, compared to the contribution of the knowledge level (NR-score), while controlling for Group, Gender, and Training length. Cronbach's alpha was used to estimate NR and FS-score reliability per year group. NR score was found to explain 27 % of the variance of FS [F(1,332) = 219.2, p < 0.0005], DKO-score, and the interaction of DKO and Gender were found to explain 8 % [F(2,330) = 41.5, p < 0.0005], and the interaction of DKO and NR 1.6 % [F(1,329) = 16.6, p < 0.0005], supporting our hypothesis that FS introduces construct-irrelevant variance into the knowledge measurement. However, NR-scores showed considerably lower reliabilities than FS-scores (mean year-test group Cronbach's alphas were 0.62 and 0.74, respectively). Decisions about FS with progress tests should be a careful trade-off between systematic and random measurement error.
Collapse
Affiliation(s)
- C J Ravesloot
- Radiology Department, University Medical Center Utrecht, Heidelberglaan 100, Room E01.132, 3508 GA, Utrecht, The Netherlands.
| | | | | | - C Haaring
- Radiology Department, University Medical Center Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - C L J J Kruitwagen
- Julius Center, University Medical Center Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - F J A Beek
- Radiology Department, University Medical Center Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - J Bakker
- Albert Schweitzer Hospital, Dordrecht, The Netherlands
| | - J P J Van Schaik
- Radiology Department, University Medical Center Utrecht, Heidelberglaan 100, Room E01.132, 3508 GA, Utrecht, The Netherlands
| | - Th J Ten Cate
- Center for Research and Development of Education, University Medical Center Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| |
Collapse
|
14
|
Does competition work as a motivating factor in e-learning? A randomized controlled trial. PLoS One 2014; 9:e85434. [PMID: 24465561 PMCID: PMC3894964 DOI: 10.1371/journal.pone.0085434] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2013] [Accepted: 11/27/2013] [Indexed: 11/19/2022] Open
Abstract
BACKGROUND AND AIMS Examinations today are often computerized and the primary motivation and curriculum is often based on the examinations. This study aims to test if competition widgets in e-learning quiz modules improve post-test and follow-up test results and self-evaluation. The secondary aim is to evaluate improvements during the training period comparing test-results and number of tests taken. METHODS Two groups were randomly assigned to either a quiz-module with competition widgets or a module without. Pre-, post- and follow up test-results were recorded. Time used within the modules was measured and students reported time studying. Students were able to choose questions from former examinations in the quiz-module. RESULTS Students from the competing group were significantly better at both post-and follow-up-test and had a significantly better overall learning efficiency than those from the non-competing group. They were also significantly better at guessing their post-test results. CONCLUSION Quiz modules with competition widgets motivate students to become more active during the module and stimulate better total efficiency. They also generate improved self-awareness regarding post-test-results.
Collapse
|
15
|
Phillips AW, Smith SG, Straus CM. Driving deeper learning by assessment: an adaptation of the Revised Bloom's Taxonomy for medical imaging in gross anatomy. Acad Radiol 2013; 20:784-9. [PMID: 23664405 DOI: 10.1016/j.acra.2013.02.001] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2012] [Revised: 01/31/2013] [Accepted: 02/01/2013] [Indexed: 11/28/2022]
Abstract
RATIONALE AND OBJECTIVES As medical imaging continues to grow as a central modality by which physicians of all specialties visualize anatomy, so, too, is its role in medical student education. However, no study to our knowledge has attempted to categorize the necessary cognitive skills. Here, we assess a tool to identify those skills and their possible hierarchical nature that reflects deeper understanding of radiological anatomy. MATERIALS AND METHODS We adapted the revision of Bloom's Taxonomy of Educational Objectives to create examination questions and teaching points for normal radiological anatomy in a medical anatomy course in 2008. All six previously established levels of cognitive processes were adapted, ranging from Remembering to Create. Reliability and validity were assessed. RESULTS Of 102 eligible students, 98 (98%) consented to participate, and 108 examination questions were assessed. Cronbach α assessing reliability ranged from poor (.197) to moderate (.571) with most categories being moderate. Score means for the levels of cognitive processes were statistically distinct [F(4, 102) = 180.63, P < .001] and tended to decrease as the level of cognitive process increased [Spearman ρ(5) = -.800, P = .104], consistent with a valid hierarchical structure. CONCLUSIONS A radiological anatomy adaptation of the revised taxonomy demonstrated generally adequate reliability and acceptable validity to establish evaluations that test different depths of cognitive processes. This is a critical first step to create a fundamental curricular tool by which medical imaging education-both normal and pathological-may be taught and assessed in the future.
Collapse
Affiliation(s)
- Andrew W Phillips
- Stanford University, Department of Emergency Medicine, Stanford, California, USA
| | | | | |
Collapse
|
16
|
Bond AE, Bodger O, Skibinski DOF, Jones DH, Restall CJ, Dudley E, van Keulen G. Negatively-marked MCQ assessments that reward partial knowledge do not introduce gender bias yet increase student performance and satisfaction and reduce anxiety. PLoS One 2013; 8:e55956. [PMID: 23437081 PMCID: PMC3577794 DOI: 10.1371/journal.pone.0055956] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2012] [Accepted: 01/07/2013] [Indexed: 11/19/2022] Open
Abstract
Multiple-choice question (MCQ) examinations are increasingly used as the assessment method of theoretical knowledge in large class-size modules in many life science degrees. MCQ-tests can be used to objectively measure factual knowledge, ability and high-level learning outcomes, but may also introduce gender bias in performance dependent on topic, instruction, scoring and difficulty. The 'Single Answer' (SA) test is often used in which students choose one correct answer, in which they are unable to demonstrate partial knowledge. Negatively marking eliminates the chance element of guessing but may be considered unfair. Elimination testing (ET) is an alternative form of MCQ, which discriminates between all levels of knowledge, while rewarding demonstration of partial knowledge. Comparisons of performance and gender bias in negatively marked SA and ET tests have not yet been performed in the life sciences. Our results show that life science students were significantly advantaged by answering the MCQ test in elimination format compared to single answer format under negative marking conditions by rewarding partial knowledge of topics. Importantly, we found no significant difference in performance between genders in either cohort for either MCQ test under negative marking conditions. Surveys showed that students generally preferred ET-style MCQ testing over SA-style testing. Students reported feeling more relaxed taking ET MCQ and more stressed when sitting SA tests, while disagreeing with being distracted by thinking about best tactics for scoring high. Students agreed ET testing improved their critical thinking skills. We conclude that appropriately-designed MCQ tests do not systematically discriminate between genders. We recommend careful consideration in choosing the type of MCQ test, and propose to apply negative scoring conditions to each test type to avoid the introduction of gender bias. The student experience could be improved through the incorporation of the elimination answering methods in MCQ tests via rewarding partial and full knowledge.
Collapse
Affiliation(s)
- A. Elizabeth Bond
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
- Advanced Professional Training in Bioscience, College of Science, Swansea University, Swansea, Wales, United Kingdom
| | - Owen Bodger
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
| | - David O. F. Skibinski
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
| | - D. Hugh Jones
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
| | - Colin J. Restall
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
| | - Edward Dudley
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
| | - Geertje van Keulen
- Institute of Life Science, College of Medicine, Swansea University, Swansea, Wales, United Kingdom
- * E-mail:
| |
Collapse
|
17
|
van Bruggen L, Manrique-van Woudenbergh M, Spierenburg E, Vos J. Preferred question types for computer-based assessment of clinical reasoning: a literature study. PERSPECTIVES ON MEDICAL EDUCATION 2012; 1:162-171. [PMID: 23205341 PMCID: PMC3508269 DOI: 10.1007/s40037-012-0024-1] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Clinical reasoning is a core competence of doctors. Therefore, the assessment of clinical reasoning of undergraduate students is an important part of medical education. Three medical universities in the Netherlands wish to develop a shared question database in order to assess clinical reasoning of undergraduate students in Computer-Based Assessments (CBA). To determine suitable question types for this purpose a literature study was carried out. Search of ERIC and PubMed and subsequent cross referencing yielded 30 articles which met the inclusion criteria of a focus on question types suitable to assess clinical reasoning of medical students and providing recommendations for their use. Script Concordance Tests, Extended Matching Questions, Comprehensive Integrative Puzzles, Modified Essay Questions/Short Answer Questions, Long Menu Questions, Multiple Choice Questions, Multiple True/False Questions and Virtual Patients meet the above-mentioned criteria, but for different reasons not all types can be used easily in CBA. A combination of Comprehensive Integrative Puzzles and Extended Matching Questions seems to assess most aspects of clinical reasoning and these question types can be adapted for use in CBA. Regardless of the question type chosen, patient vignettes should be used as a standard stimulus format to assess clinical reasoning. Further research is necessary to ensure that the combination of these question types produces valid assessments and reliable test results.
Collapse
Affiliation(s)
- Lisette van Bruggen
- Center for Research and Development of Education, University Medical Center Utrecht, P.O. Box 85500, 3508 GA, Utrecht, the Netherlands.
| | - Margreet Manrique-van Woudenbergh
- Center for Research and Development of Education, University Medical Center Utrecht, P.O. Box 85500, 3508 GA, Utrecht, the Netherlands
| | | | | |
Collapse
|
18
|
Knobe M, Holschen M, Mooij SC, Sellei RM, Münker R, Antony P, Pfeifer R, Drescher W, Pape HC. Knowledge transfer of spinal manipulation skills by student-teachers: a randomised controlled trial. EUROPEAN SPINE JOURNAL : OFFICIAL PUBLICATION OF THE EUROPEAN SPINE SOCIETY, THE EUROPEAN SPINAL DEFORMITY SOCIETY, AND THE EUROPEAN SECTION OF THE CERVICAL SPINE RESEARCH SOCIETY 2012; 21:992-8. [PMID: 22223196 DOI: 10.1007/s00586-011-2140-8] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/16/2011] [Revised: 08/02/2011] [Accepted: 12/25/2011] [Indexed: 10/14/2022]
Abstract
PURPOSE To assess the use of peer-assisted learning (PAL) of complex manipulative motor skills with respect to gender in medical students. METHODS In 2007-2010, 292 students in their 3rd and 4th years of medical school were randomly assigned to two groups [Staff group (SG), PAL group (PG)] led by either staff tutors or student-teachers (ST). The students were taught bimanual practical and diagnostic skills (course education module of eight separate lessons) as well as a general introduction to the theory of spinal manipulative therapy. In addition to qualitative data collection (Likert scale), evaluation was performed using a multiple-choice questionnaire in addition to an objective structured clinical examination (OSCE). RESULTS Complex motor skills as well as palpatory diagnostic competencies could in fact be better taught through professionals than through ST (manipulative OSCE grades/diagnostic OSCE score; SG vs. PG; male: P = 0.017/P < 0.001, female: P < 0.001/P < 0.001). The registration of theoretical knowledge showed equal results in students taught by staff or ST. In both teaching groups (SG: n = 147, PG: n = 145), no significant differences were observed between male and female students in matters of manipulative skills or theoretical knowledge. Diagnostic competencies were better in females than in males in the staff group (P = 0.041) Overall, students were more satisfied with the environment provided by professional teachers than by ST, though male students regarded the PAL system more suspiciously than their female counterparts. CONCLUSIONS The peer-assisted learning system does not seem to be generally qualified to transfer such complex spatiotemporal demands as spinal manipulative procedures.
Collapse
Affiliation(s)
- Matthias Knobe
- Department of Orthopedic and Trauma Surgery, Medical Faculty, RWTH Aachen University, Pauwelsstraße 30, 52074 Aachen, Germany.
| | | | | | | | | | | | | | | | | |
Collapse
|
19
|
Melo Prado H, Hannois Falbo G, Rodrigues Falbo A, Natal Figueirôa J. Active learning on the ward: outcomes from a comparative trial with traditional methods. MEDICAL EDUCATION 2011; 45:273-279. [PMID: 21299601 DOI: 10.1111/j.1365-2923.2010.03846.x] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
CONTEXT Academic activity during internship is essentially practical and ward rounds are traditionally considered the cornerstone of clinical education. However, the efficacy and effectiveness of ward rounds for learning purposes have been under-investigated and it is necessary to assess alternative educational paradigms for this activity. OBJECTIVES This study aimed to compare the educational effectiveness of ward rounds conducted with two different learning methodologies. METHODS Student subjects were first tested on 30 true/false questions to assess their initial degree of knowledge on pneumonia and diarrhoea. Afterwards, they attended ward rounds conducted using an active and a traditional learning methodology. The participants were submitted to a second test 48hours later in order to assess knowledge acquisition and were asked to answer two questions about self-directed learning and their opinions on the two learning methodologies used. RESULTS Seventy-two medical students taking part in a paediatric clinic rotation were enrolled. The active methodology proved to be more effective than the traditional methodology for the three outcomes considered: knowledge acquisition (33 students [45.8%] versus 21 students [29.2%]; p=0.03); self-directed learning (38 students [52.8%] versus 11 students [15.3%]; p<0.001), and student opinion on the methods (61 students [84.7%] versus 38 students [52.8%]; p<0.001). CONCLUSIONS The active methodology produced better results than the traditional methodology in a ward-based context. This study seems to be valuable in terms of the new evidence it demonstrates on learning methodologies in the context of the ward round.
Collapse
Affiliation(s)
- Hegla Melo Prado
- Instituto de Medicina Integral Professor Fernando Figueira-IMIP, Recife, Pernambuco, Brazil. ,
| | | | | | | |
Collapse
|
20
|
Foster N. Analysis of short-answer question styles versus gender in pre-clinical veterinary education. JOURNAL OF VETERINARY MEDICAL EDUCATION 2011; 38:67-73. [PMID: 21805937 DOI: 10.3138/jvme.38.1.67] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
One large study in medical education has reported that the choice of question format (or question content) could introduce a gender bias, with men outperforming women on questions with a true-false component or that required knowledge of anatomy or physiology. The purpose of our study was to ascertain whether this finding is also true in veterinary medical education. Two veterinary student cohorts were analyzed across four different modules over a three-year period (804 questions in total). The results of the study show that the women's and men's performance did not differ in any of the question types analyzed across any module or year. When students' (both women and men) overall average performance on different question types was compared with their performance on standard multiple-choice questions (MCQs), performance levels increased when students were asked to answer MCQs that contained an image-based prop (IMCQ) such as a photograph, X-ray image, or diagram. In contrast, students' performance was consistently lower when answering assertion-reason questions (ARQs), and this performance could not be explained by the demographic makeup of the two cohorts analyzed. When comparing standard MCQs with MCQs that contained a true-false question stem, no specific trend in the data could be determined. In conclusion, this study suggests that the short-answer question style does not bias against one gender in veterinary medical education, but that overall students do perform differently according to question type and, in particular, less well when ARQs are used in examinations.
Collapse
Affiliation(s)
- Neil Foster
- School of Veterinary Medicine and Science, University of Nottingham, Sutton Bonington, Leicestershire, UK.
| |
Collapse
|
21
|
Foster N, Gardner D, Kydd J, Robinson R, Roshier M. Assessing the influence of gender, learning style, and pre-entry experience on student response to delivery of a novel veterinary curriculum. JOURNAL OF VETERINARY MEDICAL EDUCATION 2010; 37:266-275. [PMID: 20847336 DOI: 10.3138/jvme.37.3.266] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
We investigated whether a novel veterinary curriculum was biased toward a particular gender, learning style, or pre-university experience (entry following undergraduate degree or direct entry from secondary school). We found no significant difference (p>0.05) in overall performance of first-year male, female, graduate-entry, or school-entry students. Students rated live-animal practical classes and facilitated problem-based learning as the most favored method of teaching, and this response was not biased by gender or pre-vet school experience. Men rated multiple-choice question (MCQ) assessment more highly than women, but there was no significant difference (p>0.05) in male or female performance on MCQ examinations. Men and women also performed comparably well in essays (both knowledge based and critical), suggesting that the retention of knowledge and depth of understanding was not gender biased. However, men performed significantly (p<0.05) better on critical essays compared with knowledge-based essays, and this trend was shown for both graduate-entry and school-entry students alike. We found no significant difference (p>0.05) in performance between groups of students with multimodal, kinesthetic, or reading-writing learning styles. Students with an auditory preference consistently performed less well in all types of assessment (p<0.05), but the number of students in this group was very small. Students whose learning style could not be specifically determined by Visual, Auditory, Read/write, Kinesthetic (VARK) tests consistently performed better than other groups, but this finding was not significant. Our results indicate that the Nottingham veterinary course does not bias for or against any of the variables we investigated.
Collapse
Affiliation(s)
- Neil Foster
- School of Veterinary Medicine and Science, University of Nottingham, Leicestershire, UK.
| | | | | | | | | |
Collapse
|