Title: The North Carolina Online Computer Skills Assessment: Relationships between Item Response Times and Item Residuals
1The North Carolina Online Computer Skills
AssessmentRelationships between Item Response
Times andItem Residuals or Can Item
Response Times Tell Us Anything about the
Probability of a Correct Response?
- John Klaric1,2
- 1Department of Educational Research Methodology,
The University of North Carolina at Greensboro - 2NC Department of Public Instruction
- Accountability Services Division
- Accountability Conference, 2009
2Purpose and Significance
- Purpose
- Accountability Services at DPI faces a daunting
task - The Testing Policy and Operations team and the
Test Development team concentrate on developing
policies and tests that assess student
proficiency in a number of content areas. - Various groups analyze student test results
results that are used by many stakeholders. These
results are also used for federal and state
reports as officials make data-driven policy
decisions. - What if the tempo, or pace, with which students
respond correctly or incorrectly is a stable
student characteristic that is informative about
that students proficiency?
3Nuisance Variables
- These factors can influence the variable of most
interest and DPI is most interested in measures
of student proficiency. - Speed-accuracy tradeoff (van der Linden, 2005)
- Proficient students who are slow to respond
during a test can be penalized, compared to
those at the same proficiency who respond
quickly. - Potential Significance consider the chemical
viscosity of O-rings, such as those used in
mechanical applications. Temperature is here a
nuisance variable it isnt of much interest.
But differences in temperature can have
devastating impacts.
4Research MethodsThe NC Online Computer Skills
Assessment (OCSA, 3rd Edition Administered Fall,
2005)
- Because it is a computer-based assessment,
examinee actions with the mouse and/or keyboard
can be captured accurately. - Response times length of item presentation vs
time to overt response - 2 datasets built from the Fall 2005 data
- Complete dataset, containing item responses and
times from all students taking the exam. - Time Truncated dataset some slower-responding
examinees were systematically excluded (about
2000, taking longer than roughly 2 hours to
complete the test)
5Research MethodsThe NC Online Computer Skills
Assessment (OCSA, 3rd Edition Administered Fall,
2005)
- Test Description
- Computer-based (non-adaptive) assessment
- 54 items in length approximately half,
multiple-choice (MC) items with 4 response
options remainder, performance-based arranged in
problem-based item sets - 0/1 scoring procedures
- Fall 2005 Administration
- 8 forms spiralled within schools, administered to
over 100,000 8th graders - Motivation NC graduation requirement
- Source North Carolina Department of Public
Instruction, 2008.
6Research MethodsThe NC Online Computer Skills
Assessment (OCSA, 3rd Edition Administered Fall,
2005)
- 6 content-related strands
- Societal/Ethical Issues (12-14)
- Spreadsheet (22-25)
- Multimedia Presentation (10-12)
- Database (22-25)
- Keyboarding/Word Processing/
- Desktop Publishing (18-20)
- Telecommunications and Internet (10-12)
7Figure 1. Total Test Response Times, Complete
Dataset (N105917) Fall 2005, NC Online Computer
Skills Assessment
8Figure 2. Total Test Response Times,
Time-Truncated Dataset (N103751) Fall 2005, NC
Online Computer Skills Assessment
9Figure 3. Total Test Score, Time Truncated
Dataset Fall 2005, NC Online Computer Skills
Assessment
10Comparison of Score Distributions from Complete
vs Time Truncated Data
11Classical Item Statistics from Edition 3 of the
NC Computer Skills Assessment (Time Truncated
Data) Fall 2005 Administration
12Item Summary Statistics from a 3-Parameter
Logistic IRT Model Fall 05 NC Computer Skills
Assessment, Time Truncated Data
13Figure 4. Response Times by Score, Item 7 (Time
Truncated Data) Fall 2005, NC Online Computer
Skills Assessment
14Figure 5. Response Times by Score, Item 9 (Time
Truncated Data) Fall 2005, NC Online Computer
Skills Assessment
15Figure 6. Response Times by Score, Item 14 (Time
Truncated Data) Fall 2005, NC Online Computer
Skills Assessment
16Figure 7. Response Times by Score, Item 23 (Time
Truncated Data) Fall 2005, Online Computer
Skills Assessment
17Figure 8. Response Times by Score, Item 45 (Time
Truncated Data) Fall 2005, NC Online Computer
Skills Assessment
18A Relationship between Error andResponse Time?
Item 1 Components of Potential Interest
A
B
Variability in item error when estimating
probability of a correct response.
Variability in item response time.
19Non-zero Correlation between Error and Response
Time
A
B
A Variance in Item Residual Error B Variance in
Item Response Time Intersection Portion of A
explained by B quantified by a semi-partial
correlation
20Summary of the NC Online Computer Skills
Assessment and Ongoing Studies
- Apparently, little intersection between residual
error and response time (see table on next slide) - Good news for the NC testing program the OCSA
appears to primarily measure student proficiency.
Proficiency estimates are not highly influenced
by response time measures when - 0/1 responses are calibrated with a
unidimensional IRT model, and - Calibration is performed with sufficiently
informative priors on the IRT c-parameter.
21Correlations expressing relationships between
residual errors and item response times
22Ongoing Studies
- Simulation studies are being conducted to see if
this intersection can be detected, and under what
circumstances. Whether possible intersections
impact estimates of student proficiency as shown
by bias and RMSE statistics is also being
studied.
23Selected References
- Lord, F.M. (1980). Applications of Item Response
Theory to Practical Testing Problems. Hillsdale,
NJ Lawrence Erlbaum Associates, Inc. - Luecht, R.M. (2008). MIRTGEN 2.0 with Response
Times. Greensboro NC University of North
Carolina at Greensboro. - North Carolina Department of Public Instruction.
(2008). Test of computer skills (Graduation
Requirement) Electronic Version. Retrieved
September 12, 2008 from http//www.dpi.state.nc.us
/accountability/testing/computerskills/ - Thissen, D. (1983). Timed testing An approach
using item response theory. In D.J. Weiss (ed.),
New Horizons in Testing Latent Trait Test Theory
and Computerized Adaptive Testing. New York, NY
Academic Press. - van der Linden, W.J. (2006). A lognormal model
for response times on test items. Journal of
Educational and Behavioral Statistics, 31,
181-204. - Wise, S.L., Kong, X. (2005). Response time
effort A new measure of examinee motivation in
computer-based tests. Applied Measurement in
Education, 18, 163-183.
24Acknowledgements
- Many thanks are owed to many people. Here are a
few - Dr. Ric Luecht (UNCG, ERM)
- Dr. Terry Ackerman (UNCG, ERM)
- Dr. Lou Fabrizio (NCDPI, Accountability Services)
- Dr. Gary Williamson (NCDPI, Accountability
Services) - Dr. Laura Kramer (NCDPI, Test Development)
- Dr. Wim van der Linden (U. Twente)