ISSN 0439-755X
CN 11-1911/B
主办:中国心理学会
   中国科学院心理研究所
出版:科学出版社

心理学报 ›› 2005, Vol. 37 ›› Issue (02): 280-284.

• • 上一篇    

大学英语四、六级考试分数等值研究

朱正才   

  1. 上海交通大学外国语学院 , 上海 200030
  • 收稿日期:2004-04-08 修回日期:1900-01-01 出版日期:2005-03-30 发布日期:2005-03-30
  • 通讯作者: 朱正才

A STUDY OF SCORE EQUATING IN THE COLLEGE ENGLISH TEST: A NEW APPROACH BASED ON “ANCHOR ITEMS” AND TWO-PARAMETER IRT MODEL

Zhu Zhengcai   

  1. Institute of Foreign Language, Jiaotong University, Shanghai 200233, China
  • Received:2004-04-08 Revised:1900-01-01 Published:2005-03-30 Online:2005-03-30
  • Contact: Zhu Zhengcai

摘要: 对现有的大学英语四、六级考试分数等值模式中存在的若干问题进行了深入的分析,并提出了新的解决方案——一个基于铆题设计和两参数IRT模型的解决方案。主要包括:(1)用两参数逻辑斯蒂模型替代原来的Rasch模型,以改进题目模型的适合性;(2)用共同题目的等值设计取代原来的共同被试等值设计,解决共同被试等值设计中,等值考生的动机水平难以控制的难题;(3)建立专用的等值用题库,并且一次性完成其中铆题的预测和参数标定工作,以解决原来等值模式中存在的误差累积问题。同时,由于铆题的保密工作难度较小,因此,等值专用题库对保证等值结果的可靠性也具有重大意义;(4)本文还对新的分数等值方案进行了真实的考试数据等值计算实验,并得到了一个令人满意的分数等值结果。

关键词: 项目反应理论, 分数等值, 逻辑斯蒂模型

Abstract: In China’s College English Test (CET), Rasch model has been used in the score equating procedure for 15 years and lots of score equating data have been accumulated. This paper discusses in detail some demerits of the score equating method based on Rasch model, and introduces a new score equating approach based on “anchor items” and two-parameters IRT model (the Item Response Theory model). It is assumed that for the old score equating method based on Rasch model: 1)The students in the control group give equal attention to both the formal and the control papers. 2)There has been no leakage of the items in either paper. 3) All items have the same Discrimination Index. A failure in assumption 1) would usually occur because the students feel that the control paper test is an extra burden to them and they often do not give it the same importance as the formal paper. In this case their marks on the control paper would be lower than their true performance. If the two papers were, in fact, equally easy or difficult they would score lower marks on the control paper, thus making it appear harder. This would have the effect of making the formal paper seem to be relatively easier and in the process of equating the students’ marks would be reduced. If assumption 2) is not true and the control paper has not truly been kept confidential, the effects would be in the opposite direction. The candidates would do better than they should on the control paper, causing their marks on that paper to be relatively high in comparison with the formal test. The latter test would therefore appear to the equating algorithm to be harder than it really is and all the students’ marks would be increased. Note that this would be true even if only a few items were leaked. For example, if just one Reading passage were leaked, together with the associated items, those five items would be scored correct for students who might otherwise have failed at least in some of them. Since reading items have double weight, this could falsely increase the score of weaker students by up to 10 marks! Of course, the effect on the mean score would be smaller since many students would have scored on these items anyway. It might also be argued that, since there is evidence that the items do not all have the same Discrimination Index, the two/three-parameter IRT model should be used. It has to be accepted that any equating step will increase the standard error of measurement (SEM) of the final score because the parameters that need to be used for equating will be estimated with some standard error of their own. However, this increase will usually be small (given the sample size of several hundred used to do the model fitting) and should be more than compensated for by the reduction in the “between-forms” bias, which the equating procedure is designed to correct. In this paper, a pilot study with real CET test data is reported with satisfactory score equating results.

Key words: College English Test, Item Response Theory, score equating

中图分类号: