語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Linking Errors Introduced by Rapid G...
~
Deng, Jiayi.
FindBook
Google Book
Amazon
博客來
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling./
作者:
Deng, Jiayi.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2024,
面頁冊數:
169 p.
附註:
Source: Dissertations Abstracts International, Volume: 85-12, Section: A.
Contained By:
Dissertations Abstracts International85-12A.
標題:
Educational psychology. -
電子資源:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=31144267
ISBN:
9798383163399
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling.
Deng, Jiayi.
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling.
- Ann Arbor : ProQuest Dissertations & Theses, 2024 - 169 p.
Source: Dissertations Abstracts International, Volume: 85-12, Section: A.
Thesis (Ph.D.)--University of Minnesota, 2024.
Test score comparability in international large-scale assessments (LSA) is of utmost importance in measuring the effectiveness of education systems and understanding the impact of education on economic growth. To effectively compare test scores on an international scale, score linking is widely used to convert raw scores from different linguistic version of test forms into a common score scale. An example is the multigroup concurrent IRT calibration method, which is used for estimating item and ability parameters across multiple linguistic groups of test-takers. The method uses common item parameters to most items and groups, with a select few items allowed to have group-specific parameters. Although prior researchers used empirical data from international LSAs to demonstrate the effectiveness of multigroup concurrent IRT calibration in offering greater global comparability in score scales, it is important to note that they assumed comparable test-taking efforts across cultural and linguistic populations. This assumption may not hold true due to differential rapid guessing (RG) rates, potentially biasing item parameter estimation. To address this gap, I proposed a real data analysis and simulation to explore this area. The objective of the current study is to investigate the linking errors introduced by RG responses when employing multigroup concurrent IRT calibration.In the real data analysis, data from the Arabic and Chinese groups in the PISA 2018 Form 18 science module were linked, with RG responses flagged using response time information. Despite observed differential RG, the linking procedure proved robust to anchor identification and ability estimation. In the simulation, data was generated for two groups with varying motivation levels. These groups were administered two linguistic versions of a test form comprising multiple-choice items. Factors such as differential RG rate, association between ability and RG propensity, group impact, sample size, and model fit criteria were considered. The assessment focused on anchor identification accuracy, item parameter estimation accuracy, and ability parameter estimation accuracy and precision. The findings showed that multigroup concurrent IRT calibration was robust against differential RG, with sample size and group impact being primary factors influencing errors. However, differential RG could affect ability estimation precision and item parameter estimation accuracy.
ISBN: 9798383163399Subjects--Topical Terms:
517650
Educational psychology.
Subjects--Index Terms:
Large-scale assessments
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling.
LDR
:03713nmm a2200409 4500
001
2402213
005
20241028051457.5
006
m o d
007
cr#unu||||||||
008
251215s2024 ||||||||||||||||| ||eng d
020
$a
9798383163399
035
$a
(MiAaPQ)AAI31144267
035
$a
AAI31144267
035
$a
2402213
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Deng, Jiayi.
$3
3772436
245
1 0
$a
Linking Errors Introduced by Rapid Guessing Responses When Employing Multigroup Concurrent IRT Scaling.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2024
300
$a
169 p.
500
$a
Source: Dissertations Abstracts International, Volume: 85-12, Section: A.
500
$a
Advisor: Davison, Mark L.;Rodriguez, Michael.
502
$a
Thesis (Ph.D.)--University of Minnesota, 2024.
520
$a
Test score comparability in international large-scale assessments (LSA) is of utmost importance in measuring the effectiveness of education systems and understanding the impact of education on economic growth. To effectively compare test scores on an international scale, score linking is widely used to convert raw scores from different linguistic version of test forms into a common score scale. An example is the multigroup concurrent IRT calibration method, which is used for estimating item and ability parameters across multiple linguistic groups of test-takers. The method uses common item parameters to most items and groups, with a select few items allowed to have group-specific parameters. Although prior researchers used empirical data from international LSAs to demonstrate the effectiveness of multigroup concurrent IRT calibration in offering greater global comparability in score scales, it is important to note that they assumed comparable test-taking efforts across cultural and linguistic populations. This assumption may not hold true due to differential rapid guessing (RG) rates, potentially biasing item parameter estimation. To address this gap, I proposed a real data analysis and simulation to explore this area. The objective of the current study is to investigate the linking errors introduced by RG responses when employing multigroup concurrent IRT calibration.In the real data analysis, data from the Arabic and Chinese groups in the PISA 2018 Form 18 science module were linked, with RG responses flagged using response time information. Despite observed differential RG, the linking procedure proved robust to anchor identification and ability estimation. In the simulation, data was generated for two groups with varying motivation levels. These groups were administered two linguistic versions of a test form comprising multiple-choice items. Factors such as differential RG rate, association between ability and RG propensity, group impact, sample size, and model fit criteria were considered. The assessment focused on anchor identification accuracy, item parameter estimation accuracy, and ability parameter estimation accuracy and precision. The findings showed that multigroup concurrent IRT calibration was robust against differential RG, with sample size and group impact being primary factors influencing errors. However, differential RG could affect ability estimation precision and item parameter estimation accuracy.
590
$a
School code: 0130.
650
4
$a
Educational psychology.
$3
517650
650
4
$a
Educational tests & measurements.
$3
3168483
650
4
$a
Education policy.
$3
2191387
650
4
$a
Linguistics.
$3
524476
653
$a
Large-scale assessments
653
$a
Multigroup concurrent
653
$a
Rapid guessing
653
$a
Score linking
653
$a
IRT calibration
690
$a
0525
690
$a
0288
690
$a
0290
690
$a
0458
710
2
$a
University of Minnesota.
$b
Educational Psychology.
$3
1023204
773
0
$t
Dissertations Abstracts International
$g
85-12A.
790
$a
0130
791
$a
Ph.D.
792
$a
2024
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=31144267
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9510533
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入