CRT Development Item specifications and analysis.

17
CRT Development Item specifications and analysis

Transcript of CRT Development Item specifications and analysis.

Page 1: CRT Development Item specifications and analysis.

CRT Development

Item specifications and analysis

Page 2: CRT Development Item specifications and analysis.

Considerations for CRTs

• Unlike NRTs, individual CRT items are not ‘expendable’ because they have been written to assess specific areas of interest

• “If a criterion-referenced test doesn’t unambiguously describe just what it’s measuring, it offers no advantage over norm-referenced measures.” (Popham, 1984, p. 29)

Popham, W. J. (1984). Specifying the domain of content or behaviors. In R. A. Berk (Ed.), A guide to criterion-referenced test construction (pp. 29-48). Baltimore, MD: The Johns Hopkins University Press.

Page 3: CRT Development Item specifications and analysis.

CRT score interpretation

(Popham, 1984, p. 31)

‘Good’ CRT

‘Bad’ CRT

Page 4: CRT Development Item specifications and analysis.

Test Specifications

• ‘Blueprints’ for creating test items• Ensure that item content matches

objectives (or criteria) to be assessed• Though usually associated with CRTs, can

also be useful in NRT development (Davidson & Lynch, 2002)

• Recent criticism: Many CRT specs (and resulting tests) are too tied to specific item types and lead to ‘narrow’ learning

Page 5: CRT Development Item specifications and analysis.

Specification components

• General Description (GD) – brief statement of the focus of the assessment

• Prompt Attributes (PA) – details what will be given to the test taker

• Response Attributes (RA) – describes what should happen when the test-taker responds to the prompt

• Sample Item (SI) • Specification Supplement (SS) – other useful

information regarding the item or scoring

(Davidson & Lynch, 2002)

Page 6: CRT Development Item specifications and analysis.

Item – specification congruence

(Brown, 1996, p. 78)

Page 7: CRT Development Item specifications and analysis.

CRT Statistical Item Analysis

• Based on criterion groups• To select groups, ask: Who should be able to

master the objectives and who should not?• Logical group comparisons

– Pre-instruction / post-instruction– Uninstructed / instructed– Contrasting groups

• The interpretation of the analysis will depend in part on the groups chosen

Page 8: CRT Development Item specifications and analysis.

Pre-instruction / post-instruction

Advantages • Individual as well as

group gains can be measured

• Can give diagnostic information about progress and program

Disadvantages• Requires post-test• Potential for test

effect

Berk, R. A. (1984). Conducting the item analysis. In R. A. Berk (Ed.), A guide to criterion-referenced test construction (pp. 97-143). Baltimore, MD: The Johns Hopkins University Press.

Page 9: CRT Development Item specifications and analysis.

Uninstructed / instructed

Advantages• Analysis can be

conducted at one point in time

• Test can be used immediately for mastery / non-mastery decisions

Disadvantages• Group identification

might be difficult • Group performance

might be affected by a variety of factors (i.e., age, background, etc.)

Page 10: CRT Development Item specifications and analysis.

Contrasting groups

Advantages• Does not equate

instruction with mastery

• Sample of masters is proportional to population

Disadvantages• Defining ‘mastery’ can

be difficult• Individually creating

each group is time consuming

• Extraneous variables

Page 11: CRT Development Item specifications and analysis.

Guidelines for selecting CRT items

Item Characteristic

Criterion Index value

Item-spec congruence

Matches objective being tested

IF (difficulty) Hard for UG

Easy for IG

IF less than .5

IF greater than .7

Discrimination Positively discriminates between criterion groups

High positive

Page 12: CRT Development Item specifications and analysis.

Item discrimination for groups

Uninstructed

Non-masters

Instructed

Masters

DI = IF (‘master’) – IF (‘non-master’)

Sometimes called DIFF (difference score)

(Berk, 1984, p. 194)

Page 13: CRT Development Item specifications and analysis.

(Brown, 1996, p. 81)

Page 14: CRT Development Item specifications and analysis.

Item analysis interpretation

(Berk, 1984, p. 125)

Page 15: CRT Development Item specifications and analysis.

Distractor efficiency analysis

• Each distractor should be selected by more students in the uninstructed (or incompetent) group than in the instructed (or competent) group.

• At least a few uninstructed (or incompetent) students (5 – 10%) should choose each distractor.

• No distractor should receive as many responses by the instructed (or competent) group as the correct answer.

(Berk, 1984, p. 127)

Page 16: CRT Development Item specifications and analysis.

The B-index

• Difficulty index calculated from one test administration

• The criterion groups are defined by their passing or failing the test

• Failing is defined as falling below a predetermined cut score

• The validity of the cut score decision will affect the validity of the B-index

Page 17: CRT Development Item specifications and analysis.

(Brown, 1996, p. 83)