CRT Development - PowerPoint PPT Presentation

About This Presentation
Title:

CRT Development

Description:

CRT Development Item specifications and analysis Considerations for CRTs Unlike NRTs, individual CRT items are not expendable because they have been written to ... – PowerPoint PPT presentation

Number of Views:85
Avg rating:3.0/5.0
Slides: 18
Provided by: Martyn85
Category:
Tags: crt | development

less

Transcript and Presenter's Notes

Title: CRT Development


1
CRT Development
  • Item specifications and analysis

2
Considerations for CRTs
  • Unlike NRTs, individual CRT items are not
    expendable because they have been written to
    assess specific areas of interest
  • If a criterion-referenced test doesnt
    unambiguously describe just what its measuring,
    it offers no advantage over norm-referenced
    measures. (Popham, 1984, p. 29)

Popham, W. J. (1984). Specifying the domain of
content or behaviors. In R. A. Berk (Ed.), A
guide to criterion-referenced test construction
(pp. 29-48). Baltimore, MD The Johns Hopkins
University Press.
3
CRT score interpretation
Good CRT
Bad CRT
(Popham, 1984, p. 31)
4
Test Specifications
  • Blueprints for creating test items
  • Ensure that item content matches objectives (or
    criteria) to be assessed
  • Though usually associated with CRTs, can also be
    useful in NRT development (Davidson Lynch,
    2002)
  • Recent criticism Many CRT specs (and resulting
    tests) are too tied to specific item types and
    lead to narrow learning

5
Specification components
  • General Description (GD) brief statement of the
    focus of the assessment
  • Prompt Attributes (PA) details what will be
    given to the test taker
  • Response Attributes (RA) describes what should
    happen when the test-taker responds to the prompt
  • Sample Item (SI)
  • Specification Supplement (SS) other useful
    information regarding the item or scoring

(Davidson Lynch, 2002)
6
Item specification congruence
(Brown, 1996, p. 78)
7
CRT Statistical Item Analysis
  • Based on criterion groups
  • To select groups, ask Who should be able to
    master the objectives and who should not?
  • Logical group comparisons
  • Pre-instruction / post-instruction
  • Uninstructed / instructed
  • Contrasting groups
  • The interpretation of the analysis will depend in
    part on the groups chosen

8
Pre-instruction / post-instruction
  • Advantages
  • Individual as well as group gains can be measured
  • Can give diagnostic information about progress
    and program
  • Disadvantages
  • Requires post-test
  • Potential for test effect

Berk, R. A. (1984). Conducting the item analysis.
In R. A. Berk (Ed.), A guide to
criterion-referenced test construction (pp.
97-143). Baltimore, MD The Johns Hopkins
University Press.
9
Uninstructed / instructed
  • Advantages
  • Analysis can be conducted at one point in time
  • Test can be used immediately for mastery /
    non-mastery decisions
  • Disadvantages
  • Group identification might be difficult
  • Group performance might be affected by a variety
    of factors (i.e., age, background, etc.)

10
Contrasting groups
  • Advantages
  • Does not equate instruction with mastery
  • Sample of masters is proportional to population
  • Disadvantages
  • Defining mastery can be difficult
  • Individually creating each group is time
    consuming
  • Extraneous variables

11
Guidelines for selecting CRT items
Item Characteristic Criterion Index value
Item-spec congruence Matches objective being tested
IF (difficulty) Hard for UG Easy for IG IF less than .5 IF greater than .7
Discrimination Positively discriminates between criterion groups High positive
12
Item discrimination for groups
DI IF (master) IF (non-master) Sometimes
called DIFF (difference score)
(Berk, 1984, p. 194)
13
(Brown, 1996, p. 81)
14
Item analysis interpretation
(Berk, 1984, p. 125)
15
Distractor efficiency analysis
  • Each distractor should be selected by more
    students in the uninstructed (or incompetent)
    group than in the instructed (or competent)
    group.
  • At least a few uninstructed (or incompetent)
    students (5 10) should choose each distractor.
  • No distractor should receive as many responses by
    the instructed (or competent) group as the
    correct answer.

(Berk, 1984, p. 127)
16
The B-index
  • Difficulty index calculated from one test
    administration
  • The criterion groups are defined by their passing
    or failing the test
  • Failing is defined as falling below a
    predetermined cut score
  • The validity of the cut score decision will
    affect the validity of the B-index

17
(Brown, 1996, p. 83)
Write a Comment
User Comments (0)
About PowerShow.com