Next Article in Journal / Special Issue
Exploring Shyness among Veterinary Medical Students: Implications for Mental and Social Wellness
Previous Article in Journal
Variation in the Reported Management of Canine Prolapsed Nictitans Gland and Feline Herpetic Keratitis
Previous Article in Special Issue
Educating the Future of Science and Medicine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing Item Performance on Three- Versus Four-Option Multiple Choice Questions in a Veterinary Toxicology Course

1
Department of Clinical Sciences, North Carolina State University, Raleigh, NC 27695, USA
2
Department of Molecular and Biomedical Sciences, North Carolina State University, Raleigh, NC 27695, USA
*
Authors to whom correspondence should be addressed.
Vet. Sci. 2018, 5(2), 55; https://doi.org/10.3390/vetsci5020055
Submission received: 29 May 2018 / Revised: 6 June 2018 / Accepted: 8 June 2018 / Published: 9 June 2018
(This article belongs to the Special Issue Educating the Future of Veterinary Science and Medicine)

Abstract

:
Background: The number of answer options is an important element of multiple-choice questions (MCQs). Many MCQs contain four or more options despite the limited literature suggesting that there is little to no benefit beyond three options. The purpose of this study was to evaluate item performance on 3-option versus 4-option MCQs used in a core curriculum course in veterinary toxicology at a large veterinary medical school in the United States. Methods: A quasi-experimental, crossover design was used in which students in each class were randomly assigned to take one of two versions (A or B) of two major exams. Results: Both the 3-option and 4-option MCQs resulted in similar psychometric properties. Conclusion: The findings of our study support earlier research in other medical disciplines and settings that likewise concluded there was no significant change in the psychometric properties of three option MCQs when compared to the traditional MCQs with four or more options.

1. Introduction

Knowledge about a subject area remains an important framework for evaluating progress in the medical profession [1]. Multiple choice questions (MCQs) are widely used in medical curricula to assess the acquisition of factual knowledge in a subject area [2,3]. Other attractions to using MCQ format exams include the ability of an examiner to test large numbers of students with minimal resources, ease of grading including the use of non-expert and machine graders, high content coverage, and the ability to obtain highly reliable and accurate scores [2]. The use of MCQs, however, does have several important drawbacks since they often do not assess higher order thinking and may lack authenticity [2,3].
One defining feature of MCQs is the presence of two or more answer options. One challenge educators face in creating MCQs relates to the development of high-quality incorrect answer options, or “distractors” as they are commonly referred in the psychometrics literature [4]. In some cases, distractors represent implausible options that weaken discriminatory power and add little value to the quality of the MCQ [5,6]. The use of non-optimal distractors (e.g., distractors selected by fewer than 5% of examinees) has been identified as a common problem shared across a wide range of undergraduate MCQs [6]. One solution to the use of non-optimal distractors would be to reduce the number of incorrect answers found in a MCQ.
The optimal number of distractors provided in a MCQ remains the subject of debate. Some educators advocate for the inclusion of five or more options in an MCQ in order to reduce the influence of chance in answering the question [7]. Previous studies have shown that 3-option MCQs perform equally well as 4-option MCQs in this regard and the use of additional options often do not improve score reliability or validity [8,9,10,11]. Other studies have also shown improved student performance and increased test taker preference for the use of 3-option MCQs [8,12]. These studies, as well as a meta-analysis of the available data, have prompted some researchers to recommend the use of 3-option MCQ items because of the difficulty in writing effective distractors [13]. Despite this evidence, many MCQs used in veterinary schools include four or more options. To date, we are unaware of any research at the classroom assessment level that has investigated the item performance of 3-option versus 4-option MCQs in veterinary medicine. The purpose of the present study was to evaluate item performance on 3-option versus 4-option MCQs used in a core curriculum course in veterinary toxicology taught at North Carolina State University (NCSU).

2. Materials and Methods

2.1. Design and Instrumentation

This study evaluated the performance of examination items administered to two classes of veterinary students enrolled in a semester-long veterinary toxicology course taught at NCSU’s veterinary college in the Spring semester of 2017. As part of a curricular redesign within the college, the course uniquely was taught twice (one for second-year students and one for third-year students) within the same semester. A quasi-experimental, crossover design was used in which students in each class were randomly assigned to take one of two versions (A or B) of a midterm (Test #1) or final exam (Test #2) [14]. Each exam version had an equal number of eighteen 3-option and eighteen 4-option MCQs. Stems used for 3-option MCQs in version A were identical to the stems used in the 4-option MCQs found in version B. The course instructor (DCD) is a veterinary toxicologist and used his expert judgment to identify the least plausible distractor to remove from each 4-option MCQ. As part of an unconventional course policy, the instructor requires students to answer only 33 of the 36 questions presented in each exam. This policy is based in part on a personal philosophy that guessing will contaminate score validity, thus allowing students to delete three items in which they likely will make a guess will help improve score integrity. As part of this process, students identify the three items they would like to treat as unscored by marking “do not grade” as the answer option.
Each mid-term and final examination administered was a required component of the course. However, for the purposes of this study, data were stripped of any identifying information by the course instructor and only the anonymous response vectors were retained. Data were then analyzed by the lead author, a professional psychometrician. The institution’s IRB declared the study ‘Exempt’.

2.2. Sample

The sample frame consisted of a total of 101 second-year students and 99 third-year students. With respect to demographic characteristics, 75 (74.3%) students identified as female and 26 (25.7%) as male. Of these 101 students, 7 (6.93%) identified as Black or African American, 72 (72.28%) as White/Caucasian, and 21 (20.79%) as Other. Among third-year students, 78 (78.8%) identified as female and 21 (21.2%) as male. Of these 99 students, 81 (81.82%) identified as White/Caucasian and 18 (18.28%) identified as Other.

2.3. Analysis

Data analysis consisted of comparing item difficulty values (p-values, also commonly referred to as percent correct) and the discrimination coefficient (point biserial correlation) across test forms. Because the data were mostly normally distributed, both parametric and non-parametric procedures were performed (e.g., independent samples t-test and Mann-Whitney U test) to inferentially compare values. All significance testing was performed with alpha set to 0.05. Cohen’s d effect size estimates were also computed to determine the practical significance of the findings. Finally, correlations (Pearson’s r and Spearman’s ρ) were calculated to measure the association between item properties across forms.

3. Results

For Test #1, six items were removed from scoring, thus reducing the total number of items on this exam from 36 to 30. The reasons for the deletion included a combination of concerns regarding content and psychometric functioning. For the remaining 30 items, p-values ranged from 0.47 to 1.00 with a mean of 0.83 (SD = 0.16) for form A. p-values ranged from 0.64 to 1.00 with a mean of 0.86 (SD = 0.14) for form B. An independent samples t-test indicated the differences were not statistically significantly different (p = 0.490). The non-parametric equivalent Mann-Whitney U test also indicated the scores were not significantly different (p = 0.796). The Cohen’s d effect size was 0.199. With respect to discrimination coefficients (DI), the mean DI value of form A items was 1.04 (SD = 0.14), and the mean DI value of form B was 1.00 (SD = 0.12). These values also were not statistically significantly different based on the t-test (p = 0.280) and the Mann-Whitney U test (p = 0.605). The Cohen’s d effect size was 0.306. The Pearson’s r correlation coefficient was 0.858 for p-values and 0.231 for DI values across forms, and the Spearman’s ρ correlation coefficient was 0.847 for p-values and 0.291 for DI values across forms. Results indicate the overall mean difference between each item pair was 0.03 (SD = 0.08) for p-values and 0.04 (SD = 0.16) for DI coefficients. The Cohen’s d effect size was 0.079. A complete breakdown of results is presented in Table 1.
For Test #2, p-values ranged from 0.64 to 1.00 with a mean of 0.89 (SD = 0.11) for form A. p-values ranged from 0.58 to 0.99 with a mean of 0.90 (SD = 0.11) for form B. An independent samples t-test indicated the differences were not statistically significantly different (p = 0.821). The non-parametric equivalent Mann-Whitney U test also indicated the scores were not statistically significant different (p = 0.346). The Cohen’s d effect size was 0.090. With respect to discrimination coefficients (DI), the mean DI value of form A items was 0.99 (SD = 0.17), and the mean DI value of form B was 0.98 (SD = 0.12). These values also were not significantly different based on the t-test (p = 0.876) and the Mann-Whitney U test (p = 0.330). The Cohen’s d effect size was 0.067. The Pearson’s r correlation coefficient was 0.927 for p-values and 0.583 for DI values across forms, and the Spearman’s ρ correlation coefficient was 0.794 for p-values and 0.395 for DI values across forms. Furthermore, the differences between p-value and DI coefficients are expressed using absolute values. The results indicate that the overall mean difference between each item pair was 0.01 (SD = 0.04) for p-values and 0.01 (SD = 0.14) for DI coefficients. The Cohen’s d effect size was 0.000. A complete breakdown of results is presented in Table 2.

4. Discussion

Results indicate that both the 3-option and 4-option MCQs resulted in similar psychometric properties. We conclude, therefore, that items containing 3-options are equally effective as items with 4-options with respect to both item difficulty and item discrimination. Our findings are in general agreement with previous studies evaluating the performance of medical [15,16,17] and nursing students [18,19] given MCQs with three to five options. These studies showed no statistically significant differences in item discrimination, item difficulty, or the mean examination scores when MCQs were administered with three versus four option answer choices [15,16,17,18]. Our data suggest that the tradition of using MCQs with four or more options in veterinary education is no longer supported by empirical evidence.
In most cases, the least plausible distractor removed from the four option MCQs were chosen by less than 5% of the students. We contend that the use of this third distractor in a 4-option MCQ was an exercise in futility in many instances, as there was no evidence of any meaningful improvements to item functioning as a result of the additional distractor. This conclusion is supported by other studies that have evaluated the impact of less than optimal distractors on MCQ performance [5,6,19]. Our data suggests that veterinary educators should consider developing MCQs with three high quality options (one correct answer and two plausible distractors). This approach could be associated with multiple benefits including reduced time needed to create MCQs. For example, three option tests and examinations are easier to write and administer and studies evaluating nurse educators have shown that nursing faculty can write more questions with three options in the same time required to write four or more options [20]. We also anticipate that veterinary students will spend less time evaluating response options when fewer distractors are presented, thereby decreasing the amount of time students spend on each test item. Since the time needed to evaluate 3-option MCQs should be shorter, educators can include additional items on an assessment to further ensure adequate content coverage, reduce the size of associated standard error estimates, increase the statistical precision of scores, and increase the reliability of scores given the well-documented relationship between increases in reliability when additional items are included [21,22].
One concern veterinary educators may have relates to the increased odds that guessing has with 3-option versus 4-option MCQs (i.e., 33% vs. 25%). This concern can be alleviated in part by using exams with a larger number of 3-option MCQs. For example, an exam consisting of 30 items would require a minimum (raw) cut score of 15 (i.e., pass rate set at 50%) if the items consisted of 3 response options in order to ensure examinees could not achieve the lowest meaningful performance category by random guessing given a 5% maximum error tolerance [23]. Another concern that veterinary educators may have with respect to our study is the generalizability of our findings, since the present study involved only one veterinary medical college located in the United States. We contend that the findings from our study, however, are consistent with those of other studies performed in other health professions settings. This convergence of findings provides evidence that supports the external aspect of validity [24].

5. Conclusions

In summary, this study sought to evaluate item performance on 3-option versus 4-option MCQs used in a core curriculum course in veterinary toxicology at a large veterinary medical school. Both the 3-option and 4-option MCQs resulted in similar psychometric properties. In conclusion, the findings of our study support earlier research in other medical disciplines and settings that likewise concluded there was no significant change in the psychometric properties of three option MCQs when compared to the traditional MCQs with four or more options.

Author Contributions

Conceptualization, K.R. and D.D.; Methodology, K.R. and D.D.; Software, K.R.; Formal Analysis, K.R.; Investigation, K.R. and D.D.; Resources, K.R. and D.D.; Data Curation, D.D.; Writing-Original Draft Preparation, K.R. and D.D.; Writing-Review & Editing, K.R. and D.D.; Visualization, N/A.; Supervision, K.R. and D.D.; Project Administration, D.D.; Funding Acquisition, N/A.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Peile, E. Knowing and knowing about. BMJ 2006, 332, 645. [Google Scholar] [CrossRef] [PubMed]
  2. McCoubrie, P. Improving the fairness of multiple-choice questions: A literature review. Med. Teach. 2004, 26, 709–712. [Google Scholar] [CrossRef] [PubMed]
  3. Van Der Vleuten, C.P. The assessment of professional competence: Developments, research and practical implications. Adv. Health Sci. Educ. Theory Pract. 1996, 1, 41–67. [Google Scholar] [CrossRef] [PubMed]
  4. Dickinson, J.R. How many options do multiple-choice questions really have? Dev. Bus. Simul. Exp. Learn. 2013, 40, 171–175. [Google Scholar]
  5. Cizek, G.J.; Robinson, L.K.; O’Day, D.M. Non-functioning options: A closer look. Educ. Psychol. Meas. 1998, 58, 605–611. [Google Scholar] [CrossRef]
  6. DiBattista, D.; Kurzawa, L. Examination of the quality of multiple-choice items on classroom tests. Can. J. Scholarsh. Teach. Learn. 2011, 2, 4. [Google Scholar] [CrossRef]
  7. Haladyna, T.M.; Downing, S.M. A taxonomy of multiple-choice item-writing rules. Appl. Meas. Educ. 1989, 2, 37–50. [Google Scholar] [CrossRef]
  8. Sadeghi, K.; Masoumi, G.A. Does number of options in multiple choice tests affect item facility and discrimination? An examination of test-taker preferences. J. Engl. Lang. Teach. Learn. 2017, 19, 123–143. [Google Scholar]
  9. Shizuka, T.; Takeuchi, O.; Yashima, T.; Yoshizawa, K. A comparison of three- and four-option English tests for university entrance selection purposes in Japan. Lang. Test. 2006, 23, 35–57. [Google Scholar] [CrossRef] [Green Version]
  10. Tarrant, M.; Ware, J. A comparison of the psychometric properties of three- and four-option multiple-choice questions in nursing assessments. Nurse Educ. Today 2010, 30, 539–543. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Vyas, R.; Supe, A. Multiple choice questions: A literature review on the optimal number of options. Natl. Med. J. India 2008, 21, 130–133. [Google Scholar] [PubMed]
  12. Landrum, R.E.; Cashin, J.R.; Theis, K.S. More evidence in favor of three-option multiple choice tests. Educ. Psychol. Meas. 1993, 53, 771–778. [Google Scholar] [CrossRef]
  13. Rodriguez, M.C. Three options are optimal for multiple choice items: A meta analysis of 80 years of research. Educ. Meas. Issues Pract. 2005, 24, 3–13. [Google Scholar] [CrossRef]
  14. Royal, K.D. Robust (and ethical) education research designs. J. Vet. Med. Educ. 2018, 45, 11–15. [Google Scholar] [CrossRef] [PubMed]
  15. Deepak, K.K.; Al-Umran, K.U.; AI-Sheikh, M.H.; Dkoli, B.V.; Al-Rubaish, A. Psychometrics of multiple choice questions with non-functioning distracters: Implications to medical education. Indian J. Physiol. Pharmacol. 2015, 59, 428–435. [Google Scholar] [PubMed]
  16. Dehnad, A.; Nasser, H.; Hosseini, A.F. A comparison between three-and four-option multiple choice questions. Procedia Soc. Behav. Sci. 2014, 98, 398–403. [Google Scholar] [CrossRef]
  17. Vegada, B.; Shukla, A.; Khilnani, A.; Charan, J.; Desai, C. Comparison between three option, four option and five option multiple choice question tests for quality parameters: A randomized study. Indian J. Pharmacol. 2016, 48, 571–575. [Google Scholar] [CrossRef] [PubMed]
  18. Redmond, S.P.; Hartigan-Rogers, J.A.; Cobbett, S. High time for a change: Psychometric analysis of multiple-choice questions in nursing. Int. J. Nurs. Educ. Scholarsh. 2012, 9. [Google Scholar] [CrossRef] [PubMed]
  19. Tarrant, M.; Ware, J.; Mohammed, A.M. An assessment of functioning and non-functioning distractors in multiple-choice questions: A descriptive analysis. BMC Med. Educ. 2009, 9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Tarrant, M.; Ware, J. A framework for improving the quality of multiple-choice assessments. Nurse Educ. 2012, 37, 98–104. [Google Scholar] [CrossRef] [PubMed]
  21. Brown, W. Some experimental results in the correlation of mental abilities. Br. J. Psychol. 1910, 3, 296–322. [Google Scholar]
  22. Spearman, C. Correlation calculated from faulty data. Br. J. Psychol. 1910, 3, 271–295. [Google Scholar] [CrossRef]
  23. Royal, K.D.; Stockdale, M.R. The impact of 3-option responses to multiple-choice questions on guessing strategies and cut score determinations. J. Adv. Med. Educ. Prof. 2017, 5, 84–89. [Google Scholar] [PubMed]
  24. Messick, S. Validity of Psychological Assessment: Validation of Inferences from Persons’ Responses and Performances as Scientific Inquiry into Score Meaning. Am. Psychol. 1995, 50, 741–749. [Google Scholar] [CrossRef]
Table 1. Item Statistics for Test #1, Forms A and B (N = 30).
Table 1. Item Statistics for Test #1, Forms A and B (N = 30).
ItemForm AForm Bp-Value Δ (Absolute)DI Δ (Absolute)
# Optionsp-ValueDI# Optionsp-ValueDI
141.001.0030.961.020.040.02
240.941.0530.940.940.000.11
340.711.2130.741.170.040.04
440.921.0130.900.920.020.09
540.981.0531.001.000.020.05
640.930.9630.941.010.010.05
740.900.9730.950.980.040.01
840.791.0030.880.970.090.03
940.750.9730.871.040.120.07
1040.911.0230.901.080.010.06
1140.961.0330.981.000.020.03
1240.531.1930.420.600.110.59
1340.461.4030.791.080.330.32
1440.920.9331.001.000.080.07
1540.970.9230.990.990.020.07
1640.601.1830.821.020.230.16
1730.870.9840.900.960.030.02
1830.911.0540.911.000.000.05
1930.890.9240.920.980.030.06
2030.931.0740.861.140.070.07
2130.991.0240.950.880.040.14
2230.870.9740.861.040.010.07
2330.691.2640.671.350.020.09
2430.940.9540.971.020.030.07
2530.920.9340.960.990.030.06
2630.991.0740.980.990.010.08
2730.751.1040.751.170.000.07
2830.470.7540.570.890.100.14
2930.951.0040.901.070.050.07
3030.561.4340.550.980.010.45
Table 2. Item Statistics for Test 2, Forms A and B (N = 36).
Table 2. Item Statistics for Test 2, Forms A and B (N = 36).
ItemForm AForm Bp-Value Δ (Absolute)DI Δ (Absolute)
# Optionsp-ValueDI# Optionsp-ValueDI
140.961.1130.961.060.000.05
240.770.8530.730.970.040.12
340.940.9730.960.860.010.11
440.990.9630.991.090.000.13
540.991.0730.961.100.030.03
640.911.0630.981.000.070.06
740.871.1430.861.010.000.13
840.660.5630.600.490.060.07
940.740.7730.810.960.060.19
1040.971.0030.991.030.010.03
1140.980.9730.981.000.000.03
1240.801.3330.740.900.060.43
1340.941.0030.991.000.050.00
1440.971.1030.961.010.010.09
1540.970.9830.981.050.010.07
1640.911.0430.970.960.060.08
1740.850.9930.951.080.100.09
1840.770.8630.910.950.130.09
1930.991.0040.981.040.010.04
2031.001.0040.951.080.050.08
2130.740.9640.740.870.010.09
2231.001.0040.961.100.040.10
2330.820.9140.821.050.000.14
2430.991.0140.970.990.020.02
2530.641.5040.581.090.050.41
2630.841.1140.890.960.050.15
2730.880.8140.910.940.030.13
2830.990.9340.971.080.020.15
2930.921.1040.891.230.030.13
3031.001.0041.001.000.000.00
3131.001.0040.970.940.030.06
3230.810.9740.820.770.010.20
3330.951.0040.991.020.040.02
3430.990.9841.001.000.010.02
3530.670.5340.640.680.020.15
3630.950.9840.990.990.040.01

Share and Cite

MDPI and ACS Style

Royal, K.; Dorman, D. Comparing Item Performance on Three- Versus Four-Option Multiple Choice Questions in a Veterinary Toxicology Course. Vet. Sci. 2018, 5, 55. https://doi.org/10.3390/vetsci5020055

AMA Style

Royal K, Dorman D. Comparing Item Performance on Three- Versus Four-Option Multiple Choice Questions in a Veterinary Toxicology Course. Veterinary Sciences. 2018; 5(2):55. https://doi.org/10.3390/vetsci5020055

Chicago/Turabian Style

Royal, Kenneth, and David Dorman. 2018. "Comparing Item Performance on Three- Versus Four-Option Multiple Choice Questions in a Veterinary Toxicology Course" Veterinary Sciences 5, no. 2: 55. https://doi.org/10.3390/vetsci5020055

APA Style

Royal, K., & Dorman, D. (2018). Comparing Item Performance on Three- Versus Four-Option Multiple Choice Questions in a Veterinary Toxicology Course. Veterinary Sciences, 5(2), 55. https://doi.org/10.3390/vetsci5020055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop