Aalaa Yaseen Hassan

Aalaa Yaseen Hassan

مطالب

فیلتر های جستجو: فیلتری انتخاب نشده است.
نمایش ۱ تا ۵ مورد از کل ۵ مورد.
۱.

Validation of C-Test among Iraqi EFL University Students

کلیدواژه‌ها: C-Texts Language Ability Reliability Validity

حوزه های تخصصی:
تعداد بازدید : ۸ تعداد دانلود : ۳
This study aims to assess the performance of university students through the C-Test and to analyze the extent to which this test is valid in measuring language ability. A standardized C-Test has been created with four brief passages, each containing 20 gaps. The length of each passage varied from 95 to 109 words. Throughout each passage, only the first and last sentences were not changed. The test was taken by 100 students; 39 were male and 61 were female at Al-Nisour University/Department of English in Baghdad, Iraq. The sample consists of two groups. Both groups come from the same school and would receive similar educational input in both cases based on their grade level. The validity and reliability of the C-Test were investigated using various techniques. The study analyzed the performance of Stage 4 and Stage 3 students on the Common Language Proficiency Test in Iraq. The results showed that the test discriminates well between high-ability and low-ability examinees, with no significant difference between the two groups. The Rasch model separation reliability was relatively high, and the data were one-dimensional. The students faced difficulties in guessing the most appropriate words due to their limited English proficiency. The results suggest that developing and implementing this test could significantly improve students' academic achievements in basic foreign language classes in Iraq.
۲.

Distractor Analysis in Multiple-Choice Items Using the Rasch Model

کلیدواژه‌ها: Distractor analysis Item response theory Multiple-choice items Rasch model

حوزه های تخصصی:
تعداد بازدید : ۱۶۲ تعداد دانلود : ۱۲۴
Multiple-choice (MC) item format is commonly used in educational assessments due to its economy and effectiveness across a variety of content domains. However, numerous studies have examined the quality of MC items in high-stakes and higher education assessments and found many flawed items, especially in terms of distractors. These faulty items lead to misleading insights about the performance of students and the final decisions. The analysis of distractors is typically conducted in educational assessments with multiple-choice items to ensure high quality items are used as the basis of inference. Item response theory (IRT) and Rasch models have received little attention for analyzing distractors. For that reason, the purpose of the present study was to apply the Rasch model, to a grammar test to analyze items’ distractors of the test. To achieve this, the current study investigated the quality of 10 instructor-written MC grammar items used in an undergraduate final exam, using the items responses of 310 English as a foreign language (EFL) students who had taken part in an advanced grammar course. The results showed the acceptable fit to the Rasch model and high reliability. Malfunctioning distractors were identified.
۳.

Detecting Measurement Disturbance: Graphical Illustrations of Item Characteristic Curves

کلیدواژه‌ها: Graphical displays item characteristic curves measurement disturbances model-data fit

حوزه های تخصصی:
تعداد بازدید : ۱۵۳ تعداد دانلود : ۹۹
Measurement disturbances refer to any conditions that affect the measurement of some psychological latent variables, which result in an inaccurate interpretation of item or person estimates derived from a measurement model. Measurement disturbances are mainly attributed to the characteristics of the person, the properties of the items, and the interaction between the characteristics of the person and the features of the items. Although numerous researchers have detected measurement disturbances in different contexts, too little attention has been devoted to exploring measurement disturbances within the context of language testing and assessment, especially using graphical displays. This study aimed to show the utility of graphical displays, which surpass numeric values of infit and outfit statistics given by the Rasch model, to explore measurement disturbances in a listening comprehension test. Results of the study showed two types of outcomes for examining graphical displays and their corresponding numeric fit values: congruent and incongruent associations. It turned out that graphical displays can provide diagnostic information about the performance of test items which might not be captured through numeric values.
۴.

Multidimensional IRT Analysis of Reading Comprehension in English as a Foreign Language

کلیدواژه‌ها: Bifactor model Multidimensional IRT Reading Comprehension Unidimensional IRT

حوزه های تخصصی:
تعداد بازدید : ۱۴۸ تعداد دانلود : ۹۶
Unidimensionality is an important assumption of measurement but it is violated very often. Most of the time, tests are deliberately constructed to be multidimensional to cover all aspects of the intended construct. In such situations, the application of unidimensional item response theory (IRT) models is not justified due to poor model fit and misleading results. Multidimensional IRT (MIRT) models can handle several dimensions simultaneously and yield person ability parameters on several dimensions which is helpful for diagnostic purposes too. Furthermore, MIRT models use the correlation between the dimensions to enhance the precision of the measurement. In this study a reading comprehension test is modelled with the multidimensional Rasch model. The findings showed that a correlated 2-dimensional model has the best fit to the data. The bifactor model revealed some interesting information about the structure of reading comprehension and the reading curriculum. Implications of the study for the testing and teaching of reading comprehension are discussed.
۵.

Psychometric Modelling of Reading Aloud with the Rasch Model

کلیدواژه‌ها: Rasch partial credit model Reading aloud speaking test Validation

حوزه های تخصصی:
تعداد بازدید : ۱۴۷ تعداد دانلود : ۱۱۳
Reading aloud is recommended as a simple technique to measure speaking ability (Hughes & Hughes, 2020; Madsen, 1983). Reading aloud is currently used in the Pearson Test of English and a couple of other international English as a second language proficiency tests. Due to the simplicity of the technique, it can be used in conjunction with other techniques to measure foreign and second language learners’ speaking ability. One issue in reading aloud as a testing technique is its psychometric modelling. Because of the peculiar structure of reading aloud tasks, analysing them with item response theory models is not straightforward. In this study, the Rasch partial credit model (PCM) is suggested and used to score examinees’ reading aloud scores. The performances of 196 foreign language learners on five reading aloud passages were analysed with the PCM. Findings showed that the data fit the RPCM well and the scores are highly reliable. Implications of the study for psychometric evaluation of reading aloud or oral reading fluency are discussed.

پالایش نتایج جستجو

تعداد نتایج در یک صفحه:

درجه علمی

مجله

سال

حوزه تخصصی

زبان