مطالب مرتبط با کلیدواژه

Differential Item Functioning


۱.

Gender-based DIF across the Subject Area: A Study of the Iranian National University Entrance Exam(مقاله علمی وزارت علوم)

کلیدواژه‌ها: IRT Models Differential Item Functioning High-stakes Tests Iranian National University Entrance Exam Gender, Subject Area

حوزه‌های تخصصی:
تعداد بازدید : ۱۱۴۵ تعداد دانلود : ۴۶۰
This study aimed at investigating differential item functioning (DIF) on the Special English Test of the Iranian National University Entrance Exam (INUEE). The effect of gender and subject area was taken into account. The study utilized one-parameter IRT model with a sample of 36000 students who sat for the INUEE Special English Test in 2004 and/or 2005. The findings confirmed the presence of DIF on this test. The cloze test indicated the lowest DIF whereas language function indicated the highest DIF. The results also revealed some general gender DIF patterns across the subject area. Females were favored on the three sections of grammar, language function, and the cloze test, whereas males were favored on the vocabulary and word order sections. The reading comprehension section favored males and females equally. It was also concluded that the item format alone could not explain DIF. In other words, it is the subject area or the interaction of the subject area and item format that determines the degree and direction of DIF.
۲.

Examining the Fairness of the University Entrance Exam: A Latent Class Analysis Approach to Differential Item Functioning(مقاله علمی وزارت علوم)

کلیدواژه‌ها: Fairness Differential Item Functioning bias

حوزه‌های تخصصی:
تعداد بازدید : ۳۰۵ تعداد دانلود : ۱۶۰
Measurement has been ubiquitous in all areas of education for at least a century. Various methods have been suggested to examine the fairness of education tests especially in high-stakes contexts. The present study has adopted the newly proposed ecological approach to differential item functioning (DIF) to investigate the fairness of the Iranian nationwide university entrance exam. To this end, the actual data from an administration of the test were obtained and analyzed through both traditional logistic regression and latent class analysis (LCA) techniques. The initial DIF analysis through logistic regression revealed that 19 items (out of 70) showed either uniform or non-uniform DIF. Further examination of the sample through LCA showed that the sample is not homogeneous. LCA class enumeration revealed that three classes can be identified in the sample. DIF analysis for separate latent classes showed that three serious differences in the number of DIF items identified in each latent class ranging from zero items in latent class 3 to 43 items in latent class 2. The inclusion of the covariates in the model also showed that latent class membership could be significantly predicted from high school GPA, field of study, and acceptance quota. It is argued that the fairness of the test might be under question. The implications of the findings for the validity of the test are discussed in detail.
۳.

Native Language-Based DIF across the Subtests: A Study of the Iranian National University Entrance Exam(مقاله علمی وزارت علوم)

کلیدواژه‌ها: Iranian Undergraduate University Entrance Special English Exam (IUUESEE) Rasch Analysis Differential Item Functioning Native language

حوزه‌های تخصصی:
تعداد بازدید : ۲۶۳ تعداد دانلود : ۱۷۵
This paper reports on an investigation of native language-based differential item functioning (DIF) across the subtests of Iranian Undergraduate University Entrance Special English Exam (IUUESEE). Fourteen thousand one hundred seventy two foreign-language test takers (including four groups of Azeri, Persian, Kurdish, and Luri test takers) were chosen for the study. Uniform DIF (UDIF) and Non-uniform DIF (NUDIF) analyses were conducted on data from the four versions of IUUESEE. After establishing the unidimensionality and local independence of the data, DIF findings showed that Luri test takers were more advantaged than other native language groups across the subtests. NUDIF analysis uncovered that almost all subtests functioned in favor of low-ability test takers who haven’t been expected to outperform high-ability test takers. A probable explanation for native language-ability DIF was that Luri and low-ablity test takers were more likely to venture lucky guesses. Thoughtless errors and guessing, test-wiseness, overconfidence, stem length, unappealing distractors, and time were proposed as possible causes of DIF in IUUESEE. It was also found that the reading subtest included the large number of items with significant DIF.
۴.

Validation of a Language Center Placement Test: Differential Item Functioning

کلیدواژه‌ها: Cognitive Diagnostic Assessment Differential Item Functioning listening comprehension Placement test

حوزه‌های تخصصی:
تعداد بازدید : ۲۴۲ تعداد دانلود : ۲۴۳
The documentation of test takers’ achievements has been accomplished through large-scale assessments to find general information about students’ language ability. To remove subjectivity, Cognitive Diagnostic Assessment (CDA) has recently played a crucial role in perceiving candidates’ latent attribute patterns to find multi-diagnostic information rather than single proficiency classification. However, there are some gaps in the literature about in detail investigation of test takers’ listening comprehension language ability in responding to placement test items of a public English language center. The present study aims to validate an English placement test at a language center through a retrofitting process. In an exploratory mixed-method design, 449 participants from the same language center, including 274 females and 175 males, were selected. The performance of randomly selected participants in a language center placement test was analyzed by applying the GDINA model from R-studio packages, to detect Differential Item Functioning (DIF). Results of the study revealed DIF in some items since there is some bias in test items. The implication of this study is to provide meaningful interpretations of respondents’ attributes and improve teaching and learning by finding the strengths and weaknesses of candidates. For this purpose, the findings derived from the result of the study can raise the awareness of test developers in preparing unbiased items for the placement test, and at the same time, assist test-takers to become more critical of their English language achievements. It is also helpful for materials developers to become aware of developing materials free from bias.
۵.

Evaluating Measurement Invariance in the IELTS Listening Comprehension Test

کلیدواژه‌ها: Differential Item Functioning IELTS measurement invariance Rasch model

حوزه‌های تخصصی:
تعداد بازدید : ۱۶۵ تعداد دانلود : ۱۲۶
Measurement invariance (MI) refers to the degree to which a measurement instrument or scale produces consistent results across different groups or populations. It basically shows whether the same construct is measured in the same way across different groups, such as different cultures, genders, or age groups. If MI is established, it means that scores on the test can be compared meaningfully across different groups. To establish MI mostly confirmatory factor analysis methods are used. In this study, we aim to examine MI using the Rasch model. The responses of 211 EFL learners to the listening section of the IETLS were examined for MI across gender and randomly selected subsamples. The item difficulty measures were compared graphically using the Rasch model. Findings showed that except for a few items, the IELTS listening items exhibit MI. Therefore, score comparisons across gender and other unknown subgroups are valid with the IELTS listening scores.