Title: CRT Development
1CRT Development
- Item specifications and analysis
2Considerations for CRTs
- Unlike NRTs, individual CRT items are not
expendable because they have been written to
assess specific areas of interest - If a criterion-referenced test doesnt
unambiguously describe just what its measuring,
it offers no advantage over norm-referenced
measures. (Popham, 1984, p. 29)
Popham, W. J. (1984). Specifying the domain of
content or behaviors. In R. A. Berk (Ed.), A
guide to criterion-referenced test construction
(pp. 29-48). Baltimore, MD The Johns Hopkins
University Press.
3CRT score interpretation
Good CRT
Bad CRT
(Popham, 1984, p. 31)
4Test Specifications
- Blueprints for creating test items
- Ensure that item content matches objectives (or
criteria) to be assessed - Though usually associated with CRTs, can also be
useful in NRT development (Davidson Lynch,
2002) - Recent criticism Many CRT specs (and resulting
tests) are too tied to specific item types and
lead to narrow learning
5Specification components
- General Description (GD) brief statement of the
focus of the assessment - Prompt Attributes (PA) details what will be
given to the test taker - Response Attributes (RA) describes what should
happen when the test-taker responds to the prompt - Sample Item (SI)
- Specification Supplement (SS) other useful
information regarding the item or scoring
(Davidson Lynch, 2002)
6Item specification congruence
(Brown, 1996, p. 78)
7CRT Statistical Item Analysis
- Based on criterion groups
- To select groups, ask Who should be able to
master the objectives and who should not? - Logical group comparisons
- Pre-instruction / post-instruction
- Uninstructed / instructed
- Contrasting groups
- The interpretation of the analysis will depend in
part on the groups chosen
8Pre-instruction / post-instruction
- Advantages
- Individual as well as group gains can be measured
- Can give diagnostic information about progress
and program
- Disadvantages
- Requires post-test
- Potential for test effect
Berk, R. A. (1984). Conducting the item analysis.
In R. A. Berk (Ed.), A guide to
criterion-referenced test construction (pp.
97-143). Baltimore, MD The Johns Hopkins
University Press.
9Uninstructed / instructed
- Advantages
- Analysis can be conducted at one point in time
- Test can be used immediately for mastery /
non-mastery decisions
- Disadvantages
- Group identification might be difficult
- Group performance might be affected by a variety
of factors (i.e., age, background, etc.)
10Contrasting groups
- Advantages
- Does not equate instruction with mastery
- Sample of masters is proportional to population
- Disadvantages
- Defining mastery can be difficult
- Individually creating each group is time
consuming - Extraneous variables
11Guidelines for selecting CRT items
Item Characteristic Criterion Index value
Item-spec congruence Matches objective being tested
IF (difficulty) Hard for UG Easy for IG IF less than .5 IF greater than .7
Discrimination Positively discriminates between criterion groups High positive
12Item discrimination for groups
DI IF (master) IF (non-master) Sometimes
called DIFF (difference score)
(Berk, 1984, p. 194)
13(Brown, 1996, p. 81)
14Item analysis interpretation
(Berk, 1984, p. 125)
15Distractor efficiency analysis
- Each distractor should be selected by more
students in the uninstructed (or incompetent)
group than in the instructed (or competent)
group. - At least a few uninstructed (or incompetent)
students (5 10) should choose each distractor. - No distractor should receive as many responses by
the instructed (or competent) group as the
correct answer.
(Berk, 1984, p. 127)
16The B-index
- Difficulty index calculated from one test
administration - The criterion groups are defined by their passing
or failing the test - Failing is defined as falling below a
predetermined cut score - The validity of the cut score decision will
affect the validity of the B-index
17(Brown, 1996, p. 83)