Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community...
-
Upload
sheila-nicholson -
Category
Documents
-
view
213 -
download
1
Transcript of Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community...
![Page 1: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/1.jpg)
Item AnalysisUrsula Waln, Director of Student Learning Assessment
Central New Mexico Community College
![Page 2: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/2.jpg)
Item Analysis Used with Objective Assessment
• Looks at frequency of correct responses (or behaviors) in connection with overall performance
• Used to examine item reliability • How consistently a question or performance criterion discriminates
between high and low performers
• Can be useful in improving validity of measures
• Can help instructors decide whether to eliminate certain items from the grade calculations
• Can reveal specific strengths and gaps in student learning
![Page 3: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/3.jpg)
How Item Analysis Works
• Groups students by the highest, mid-range, and lowest overall scores and examines item responses by group
• Assumes that higher-scoring students have a higher probability of getting any given item correct than do lower-scoring students• May have studied and/or practiced more and understood the material
better• May have greater test-taking savvy, less anxiety, etc.
• Produces a calculation for each item• Do it yourself to easily calculate a group difference or discrimination index• Use EAC Outcomes (a Blackboard plug-in made available to all CNM faculty
by the Nursing program) to generate a point-biserial correlation coefficient
• Gives the instructor a way to analyze performance on each item
![Page 4: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/4.jpg)
One Way to Do Item Analysis by HandShared by Linda Suskie at the NMHEAR Conference, 2015
Item
Tally of those in Top 27%
who missed item*
Tally of those in
the Middle 46% who missed
item
Tally of those in
the Lower 27% who missed item*
Total % Who
Missed Item
Group Difference(# in Lower minus # in
Top)
1 ||||| ||||| ||||| ||||| ||
||||| ||||| ||||| ||
34% 17
2 ||||| || ||||| ||||| ||||| |||||
||||| ||||| ||||| ||||
40% 12
3 ||| | || 5% -14 ||||| |||| ||||| |||||
|17% 11
* You can use whatever portion you want for the top and lower groups, but they need to be equal. Using 27% is accepted convention (Truman Kelley, 1939).
An
unreliable
question
Good
discriminatio
n
![Page 5: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/5.jpg)
Another Way to Do Item Analysis by HandRasch Item Discrimination Index (D)
N=31 because the upper and lower group each contain 31 students (115 students tested)
Item
# in Upper Group who answered correctly
(#UG)
Portion of UG who
answered correctly
(pUG)
# in Lower Group who answered correctly
(#LG)
Portion of LG who
answered correctly
(pLG)
Discrimination Index (D)D = pUGpLG
orD
1 31 1.00 (100%) 14 0.45 (45%) 0.552 24 0.77 (77%) 12 0.39 (39%) 0.38
3 28 0.90 (90%) 29 0.93 (93%) -0.034 31 1.00 (100%) 20 0.65 (65%) 0.35
A discrimination index of 0.4 or greater is generally regarded as high and anything less than 0.2 as low (R.L. Ebel, 1954).
An
unreliable
question
Good
discriminatio
n
![Page 6: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/6.jpg)
The Same Thing but Less ComplicatedRasch Item Discrimination Index (D)
N in Upper and Lower Groups is 31 (27% of 115 students)
Item
# in Upper Group who answered correctly
(#UG)
# in Lower Group who answered correctly
(#LG)
Discrimination Index (D)
D
1 31 14 0.552 24 12 0.38
3 28 29 -0.034 31 20 0.35
It isn’t necessary to calculate the portions of correct responses in each group if you use the formula shown here.
This is really easy to do!
= 0.55 = 0.38 = -0.03 = 0.35
N=.27
![Page 7: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/7.jpg)
Example of an EAC Outcomes Report
A point-biserial correlation is the Pearson correlation between responses to a particular item and scores on the total test (with or without that item).
Correlation coefficients range from -1 to 1.
This is available to CNM faculty through Blackboard course tools.
An
unreliable
question
Good
discriminatio
n
![Page 8: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/8.jpg)
Identifying Key Questions
• A key (a.k.a. signature) question is one that provides information about student learning in relation to a specific instructional objective (or student learning outcome statement).
• The item analysis methods shown in the preceding slides can help you identify and improve the reliability of key questions.• A low level of discrimination may indicate a need to tweak the
wording.• Improving discrimination value also improves question validity.• The more valid an assessment measure, the more useful it is in
gauging student learning.
![Page 9: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/9.jpg)
Detailed Multiple-Choice Item Analysis
• The detailed item analysis method shown on the next slide is for use with key multiple-choice items.
• This type of analysis can provide clues to the nature of students’ misunderstanding, provided:• The item is a valid measure of the instructional objective• Incorrect options (distractors) are written to be diagnostic (i.e., to
reveal misconceptions or breakdowns in understanding)
![Page 10: Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.](https://reader036.fdocuments.in/reader036/viewer/2022072007/56649d345503460f94a0b36d/html5/thumbnails/10.jpg)
Example of a Detailed Item Analysis
Item 2 of 4. The correct option is E. (115 students tested)
Item Response Pattern A B C D E Row
TotalUpper 27%
|| ||||| ||||| ||||| ||||| ||||| |||| 31
6.5% 16% 77.5%
Middle 46%
||| ||||| ||||| |||| || | ||||| ||||| ||||| ||||| ||||| ||||| |||
53
6% 26% 4% 2% 62%
Lower 27%
||||| ||||| || ||||| || ||||| ||||| || 31 16% 23% 16% 6% 39%
Grand Total
10 26 7 3 69 1158.5% 23% 6% 2.5% 60%
These results suggest that distractor B might provide the greatest clue about breakdown in students’ understanding, followed by distractor A, then C.