مطالب مرتبط با کلیدواژه

bias


۱.

An Investigation into Cultural Representation in Interchange Textbooks بررسی نمود فرهنگی در کتاب های زبان اینترچنج(مقاله علمی وزارت علوم)

کلیدواژه‌ها: فرهنگ تبعیض کتاب culture Textbooks اینترچنج Intercultural Competence توانش ب فرهنگی Interchange bias

حوزه های تخصصی:
تعداد بازدید : ۸۴۱ تعداد دانلود : ۴۱۳
Language and culture are now deemed to be co-constitutive; hence English language teaching (ELT) textbooks should incorporate cultural aspects and promote intercultural competence. However, careful decisions should be made as regards to the cultural content of materials and the ways in which culture is represented. This study was an attempt to deconstruct the patterns of cultural representation and intercultural interactions in Interchange textbooks, an ELT textbook series taught to English as a foreign language (EFL) learners in Iran. Using content analysis of texts and images and with the focus on nationality, gender and race, it examined how different cultures were reflected in Interchange 1, Interchange 2, Interchange 3, and whether cultural bias or inequality was present. The quantitative and qualitative data analysis revealed that the white male group was dominant in all three textbooks. And, to use Kachru’s (1985) terms, inner, outer and expanding circle nationalities were represented in the textbooks, with the expanding circle constituting the major nationality, but American culture of inner circle was predominant. Additionally, the interactions were mainly limited to superficial aspects of the target culture although these textbooks sought to show various intercultural interactions. Dominance of male and white characters and the US culture indicated inequality in race and gender, and the superficial treatment of culture in the textbooks. Less attention to the hybrid culture and deeper level of intercultural aspects, i.e. critical reflections, in the textbooks suggests that the materials be supplemented by EFL teachers’ constructive discussion of the cultures that interact.
۲.

Investigating the Effect of the Training Program on Raters’ Oral Performance Assessment: A Mixed-Methods Study on Raters’ Think-Aloud Verbal Protocols(مقاله علمی وزارت علوم)

کلیدواژه‌ها: bias Oral performance assessment Rater training Think-aloud verbal protocols

حوزه های تخصصی:
تعداد بازدید : ۵۵۶ تعداد دانلود : ۲۳۲
Although the use of verbal protocols is growing in oral assessment, research on the use of raters’ verbal protocols is rather rare. Moreover, those few studies did not use a mixed-methods design. Therefore, this study investigated the possible impacts of rater training on novice and experienced raters’ application of a specified set of standards in rating. To meet this objective, the study made use of verbal protocols produced by 20 raters who scored 300 test takers’ oral performances and analyzed the data both qualitatively and quantitatively. The outcomes demonstrated that through applying the training program, the raters were able to concentrate more on linguistic, discourse, and phonological features; therefore, the extent of their agreement increased specifically among the inexperienced raters. The analysis of verbal protocols also revealed that training how to apply a well-defined rating scale can foster its use for raters both validly and reliably. Various groups of raters approach the task of rating in different ways, which cannot be explored through pure statistical analysis. Thus, think-aloud verbal protocols can shed light on the vague sides of the issue and add to the validity of oral language assessment. Moreover, since the results of this study showed that inexperienced raters can produce protocols of higher quality and quantity in the use of macro and micro strategies to evaluate test takers’ performances, there is no evidence based on which decision makers should exclude inexperienced raters solely because of their lack of adequate experience.
۳.

The Impact of Raters’ and Test Takers’ Gender on Oral Proficiency Assessment: A Case of Multifaceted Rasch Analysis(مقاله علمی وزارت علوم)

کلیدواژه‌ها: bias multifaceted Rasch measurement (MFRM) rating scale severity

حوزه های تخصصی:
تعداد بازدید : ۴۴۶ تعداد دانلود : ۲۸۸
The application of Multifaceted Rasch Measurement (MFRM) in rating test takers’ oral language proficiency has been investigated in some previous studies (e.g., Winke, Gass, & Myford, 2012). However, little research so far has ever documented the effect of test takers’ genders on their oral performances and few studies have investigated the relationship between the impact of raters’ gender on the awarded scores to male and female test takers. Thus, this study aimed to address the above-mentioned issue. Twenty English as a Foreign Language (EFL) teachers rated the oral performances of 300 test takers. The outcomes demonstrated that test takers’ gender differences did not have any significant role in their performance differences when they were rated by the raters of the same or opposite gender. The findings also reiterated that raters of different genders did not demonstrate bias in rating test takers of the opposite or same gender. Moreover, no significant difference was observed regarding male and female raters’ biases towards the rating scale categories. The outcomes of the study showed that both male and female raters assign fairly similar scores to test takers. This suggests no evidence based on which either male or female raters must be excluded from the rating process. The findings imply that there is no need to worry about the impact of gender for a more valid and reliable assessment.
۴.

Examining the Fairness of the University Entrance Exam: A Latent Class Analysis Approach to Differential Item Functioning(مقاله علمی وزارت علوم)

کلیدواژه‌ها: Fairness Differential Item Functioning bias

حوزه های تخصصی:
تعداد بازدید : ۲۹۱ تعداد دانلود : ۱۴۹
Measurement has been ubiquitous in all areas of education for at least a century. Various methods have been suggested to examine the fairness of education tests especially in high-stakes contexts. The present study has adopted the newly proposed ecological approach to differential item functioning (DIF) to investigate the fairness of the Iranian nationwide university entrance exam. To this end, the actual data from an administration of the test were obtained and analyzed through both traditional logistic regression and latent class analysis (LCA) techniques. The initial DIF analysis through logistic regression revealed that 19 items (out of 70) showed either uniform or non-uniform DIF. Further examination of the sample through LCA showed that the sample is not homogeneous. LCA class enumeration revealed that three classes can be identified in the sample. DIF analysis for separate latent classes showed that three serious differences in the number of DIF items identified in each latent class ranging from zero items in latent class 3 to 43 items in latent class 2. The inclusion of the covariates in the model also showed that latent class membership could be significantly predicted from high school GPA, field of study, and acceptance quota. It is argued that the fairness of the test might be under question. The implications of the findings for the validity of the test are discussed in detail.
۵.

Development and Validation of a Training-Embedded Speaking Assessment Rating Scale: A Multifaceted Rasch Analysis in Speaking Assessment

کلیدواژه‌ها: bias Interrater consistency Intrarater consistency multifaceted Rasch measurement (MFRM) Rater training rating scale

حوزه های تخصصی:
تعداد بازدید : ۲۱۴ تعداد دانلود : ۱۱۷
Performance testing including the use of rating scales has become widespread in the evaluation of second/foreign oral language assessment. However, no study has used Multifaceted Rasch Measurement (MFRM) including the facets of test takers’ ability, raters’ severity, group expertise, and scale category, in one study. 20 EFL teachers scored the speaking performance of 200 test-takers prior and subsequent to a rater training program using an analytic rating scale consisting of fluency, grammar, vocabulary, intelligibility, cohesion, and comprehension categories. The outcome demonstrated that the categories were at different levels of difficulty even after the training program. However, this outcome by no means indicated the uselessness of the training program since data analysis reflected the constructive influence of training in providing enough consistency in raters’ rating of each category of the rating scale at the post-training phase. Such an outcome indicated that raters could discriminate the various categories of the rating scale. The outcomes also indicated that MFRM can result in enhancement in rater training and functionality validation of the rating scale descriptors. The training helped raters use the descriptors of the rating scale more efficiently of its various band descriptors resulting in a reduced halo effect. The findings conveyed that stakeholders had better establish training programs to assist raters in better use of the rating scale categories of various levels of difficulty in an appropriate way. Further research could be done to make a comparative analysis between the outcome of this study and the one using a holistic rating scale in oral assessment.
۶.

Facet Variability in the Light of Rater Training in Measuring Oral Performance: A Multifaceted Rasch Analysis(مقاله علمی وزارت علوم)

کلیدواژه‌ها: bias Interrater consistency multifaceted Rasch measurement (MFRM) Rater training Severity/leniency

حوزه های تخصصی:
تعداد بازدید : ۱۱۴ تعداد دانلود : ۱۰۳
Due to subjectivity in oral assessment, much concentration has been put on obtaining a satisfactory measure of consistency among raters. However, obtaining consistency might not result in valid decisions. One matter that is at the core of both reliability and validity in oral performance is rater training. Recently, Multifaceted Rasch Measurement (MFRM) has been adopted to address the problem of rater bias and inconsistency; however, no research has incorporated the facets of test takers’ ability, raters’ severity, task difficulty, group expertise, scale criterion category, and test version together in a piece of research along with their two-sided impacts. Moreover, little research has investigated how long rater training effects last. Consequently, this study explored the influence of the training program and feedback by having 20 raters score the oral production, as measured by the CEP (Community English Program) test, produced by 300 test takers in three phases, i.e., before, immediately after and long after the training program. The results indicated that training can lead to higher degrees of interrater reliability and diminished measures of severity/leniency, and biasedness. However, it won't lead the raters into total unanimity, except for making them more self-consistent. Although rater training might result in higher internal consistency among raters, it cannot eradicate individual differences. That is, experienced raters, due to their idiosyncratic characteristics, did not benefit as much as inexperienced ones. This study also showed that the outcome of training might not endure in long run after training; thus, it requires ongoing training letting raters regain consistency.