SlideShare uma empresa Scribd logo
1 de 47
Starting Windows Especially Designed
Basics of
Item
Analysis
QUESTIONS?
Click to end
Contact Me…!!!
©James L. Paglinawan TM
® e-mail me @:
JLPaglinawan@cmu.edu.ph
JLPaglinawan@teacher.deped.gov.ph
Item Analysis
Techniques to
improve test items
and instruction
11. Administer
Tests
12. Conduct
Item
Analysis
1. Review
National
and
Profession
al
Standards
7. Develop New Test
Questions
8. Review Test Questions
13. Standard Setting
Study
14. Set Passing
Standard
9. Assemble Operational
Test Forms
10. Produce Printed Tests
Mat.
2. Convene
National
Advisory
Committee
3. Develop Domain,
Knowledge and
Skills Statements
4. Conduct Job
Analysis
5. Establish Test
Specifications
6. Develop Test
Design
Test Development Process
What is Item Analysis ?
• process that examines student
responses to individual test items
assessing quality items and the test as
a whole
• valuable in improving items which will
be used again in later tests and
eliminate ambiguous or misleading
items
• valuable for increasing instructors'
skills in test construction, and
• identifying specific areas of course
content which need greater emphasis
or clarity.
Several Purposes
1. More diagnostic information on
students
– Classroom level:
• determine questions most found very
difficult/ guessing on, -
– reteach that concept
• questions all got right –
– don't waste more time on this area
• find wrong answers students are
choosing-
– identify common misconceptions
– Individual level:
• isolate specific errors this student
made
2.Build future tests,
revise test items to make
them better
• know how much work in writing good
questions
• SHOULD NOT REUSE WHOLE TESTS -->
diagnostic teaching means responding
to needs of students, so after a few years
a test bank is build up and choose a
tests for the class
• can spread difficulty levels across your
blueprint (TOS)
3. Part of continuing
professional development
– doing occasional item analysis will
help become a better test writer
– documenting just how good your
evaluation is
– useful for dealing with parents or
administrators if there's ever a dispute
– once you start bringing out all these
impressive looking stats, parents and
administrators will believe why some
students failed.
Classical ITEM Analysis
Statistics
• Reliability (test level statistic)
• Difficulty (item level statistic)
• Discrimination (item level statistic)
Test level statistic
Quality of the Test
• Reliability and Validity
–Reliability
Consistency of measurement
–Validity
Truthfulness of response
Overall Test Quality
Individual Item Quality
Reliability
refers to the extent to which the
test is likely to produce consistent
scores.
Characteristics:
1. The intercorrelations among the items --
the greater/stronger the relative number
of positive relationships are, the greater
the reliability.
2. The length of the test –
a test with more items will have a higher
reliability, all other things being equal.
3. The content of the test --
generally, the more diverse
the subject matter tested
and the testing techniques
used, the lower the
reliability.
4. Heterogeneous groups of
test takers
Types of reliability
• Stability
1. Test – Retest
• Stability
2. Inter – rater / Observer/ Scorer
• applicable for mostly essay questions
• Use Cohen’s Kappa Statistic
• Equivalence
3. Parallel-Forms/ Equivalent
Used to assess the consistency of
the results of two tests
constructed in the same way from
the same content domain.
• Internal Consistency
• Used to assess the consistency of
results across items within a test.
4. Split – Half
• 5. Kuder-Richardson
Formula 20 / 21
Correlation is
determined from a
single administration of
a test through a study
of score variances
• 6. Cronbach's Alpha (a)
Reliability
Indices
Interpretation
.91 and above
Excellent reliability; at the level of the best standardized
tests
.81 - .90 Very good for a classroom test
.71 - .80
Good for a classroom test; in the range of most. There
are probably a few items which could be improved.
.61 - .70
Somewhat low. This test needs to be supplemented by
other measures (e.g., more tests) to determine
grades. There are probably some items which could
be improved.
.51 - .60
Suggests need for revision of test, unless it is quite
short (ten or fewer items). The test definitely needs to
be supplemented by other measures (e.g., more tests)
for grading.
.50 or below
Questionable reliability. This test should not contribute
heavily to the course grade, and it needs revision.
Test Item statistic
 Item Difficulty
Percent answering correctly
 Item Discrimination
How well the item "functions“
How “valid” the item is based
on the total test score criterion
WHAT IS A WELL-
FUNCTIONING
TEST ITEM?
• how many students got it correct?
(DIFFICULTY)
• which students got it correct?
(DISCRIMINATION)
Three important information
on quality of test items
• Item difficulty: measure whether an
item was too easy or too hard.
• Item discrimination: measure whether
an item discriminated between students
who knew the material well and students
who did not.
• Effectiveness of alternatives:
Determination whether distracters
(incorrect but plausible answers) tend to
be marked by the less able students and
not by the more able students.
Item Difficulty
• Item difficulty is simply the percentage
of students who answer an item
correctly. In this case, it is also equal to
the item mean.
Diff = # of students choosing correctly
total # of students
• The item difficulty index ranges from 0
to 100; the higher the value, the easier
the question.
Item Difficulty Level:
Definition
The percentage of students who
answered the item correctly.
High
(Difficult)
Medium
(Moderate)
Low
(Easy
)
<= 30% > 30% AND < 80%
>=80
%
0 10 20 30 40 50 60 70 80 90 100
Item Difficulty Level:
Sample
Item
No.
No. Correct
Answers
%
Correct
Difficulty
Level
1 15
2 25
3 35
4 45
Number of students who answered each item = 50
30 High
50 Medium
70 Medium
90 Low
Item Difficulty Level:
Questions/Discussion
• Is a test that nobody failed
too easy?
• Is a test on which nobody got
100% too difficult?
• Should items that are “too
easy” or “too difficult” be
thrown out?
Item Discrimination
• Traditionally, using high and low scoring
groups (upper 27 % and lower 27%)
• Computerized analyses provide more
accurate assessment of the
discrimination power of items since it
accounts all responses rather than just
high and low scoring groups.
• Equivalent to point-biserial correlation. It
provides estimate the degree an
individual item is measuring the same
thing as the rest of the items.
What is Item Discrimination?
• Generally, students who did well on
the exam should select the correct
answer to any given item on the exam.
• The Discrimination Index
distinguishes for each item between
the performance of students who did
well on the exam and students who
did poorly.
Indices of Difficulty and
Discrimination
(by Hopkins and Antes)
Index Difficulty Discrimination
0.86 above Very Easy To be discarded
0.71 – 0.85 Easy To be revised
0.30 – 0.70 Moderate Very Good items
0.15 – 0.29 Difficult To be revised
0.14 below Very Difficult To be discarded
Item Discrimination:
Questions / Discussion
• What factors could contribute to
low item discrimination between
the two groups of students?
• What is a likely cause for a
negative discrimination index?
ITEM
ANALYSIS
PROCESS
Sample TOS
Remember Understand Apply Total
Section
A
4
(1,3,7,9)
6 10 20
Section
B
5
(2,5,8,11,15)
5 4 14
Section
C
3
(6,17,21)
7 6 16
Total 12 18 20 50
Steps in Item
analysis
1. Code the test items:
- 1 for correct and 0 for
incorrect
- Vertical – columns (item
numbers)
- Horizontal – rows
(respondents/students)
TEST ITEMS
No. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 . . . . 50
1 1 0 1 1 1 0 0 0 0 1 1 1 0 0 0 0 0 1 1
2 1 1 0 1 1 1 0 1 1 1 0 1 1 1 1 0 1 1 1
3 0 0 0 1 0 0 0 1 0 0 0 1 1 1 1 1 1 1 0
4 0 1 0 0 0 1 0 0 0 1 0 0 1 0 0 0 1 0 0
5 1 0 1 1 1 0 1 1 1 0 1 1 0 1 1 1 0 1 0
6 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 1 1 0 1
7 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1
8 1 1 0 1 1 1 0 1 1 1 0 1 1 0 0 0 1 0 0
2. IN SPSS:
Analyze  Scale 
Reliability analysis 
(drag/place variables to
Item box)  Statistics 
Scale if item deleted 
ok.
• ****** Method 1 (space saver) will be used for this analysis ******
• R E L I A B I L I T Y A N A L Y S I S - S C A L E (A L P H A)
• Item-total Statistics
• Scale Scale Corrected
• Mean Variance Item- Alpha
• if Item if Item Total if Item
• Deleted Deleted Correlation Deleted
• VAR00001 14.4211 127.1053 .9401 .9502
• VAR00002 14.6316 136.8440 .7332 .9542
• VAR00003 14.4211 141.5695 .4774 .9574
• VAR00004 14.4737 128.6109 .6511 .9508
• VAR00005 14.4737 128.8252 .8274 .9509
• VAR00006 14.0526 130.6579 .2236 .9525
• VAR00007 14.2105 127.8835 .2533 .9511
• VAR00008 14.1053 128.6673 .1906 .9515
• VAR00009 14.4211 129.1410 .7311 .9513
• .....................
• VAR00022 14.4211 129.1410 .7311 .9513
• VAR00023 14.4211 127.1053 .4401 .9502
• VAR00024 14.6316 136.8440 -.0332 .9542
• VAR00047 14.4737 128.6109 .8511 .9508
• VAR00048 14.4737 128.8252 .8274 .9509
• VAR00049 14.0526 130.6579 .5236 .9525
• VAR00050 14.2105 127.8835 .7533 .9511
• Reliability Coefficients
• N of Cases = 57.0 N of Items = 50
• Alpha = .9533
3. In the output dialog box:
• Alpha placed at the bottom
• the corrected item total
correlation is the point
biserial correlation as bases
for index of test reliability
4. Count the number of
items discarded and fill
up summary item analysis
table.
Test Item Reliability Analysis
Summary (sample)
Test Level of
Difficulty
Number
of Items
% Item Number
Math Very Easy 1 2 1
(50 items) Easy 2 4 2,5
Moderate 10 20 3,4,10,15…
Difficult 30 60 6,7,8,9,11,…
Very
Difficult
7 14 16,24,32…
5. Count the number of
items retained based
on the cognitive
domains in the TOS.
Compute the
percentage per level
of difficulty.
Remember Understand Apply
N Ret N Ret N Ret
A 4 1 6 3 10 3
B 5 3 5 3 4 2
C 3 2 7 4 6 3
Total 12 6 18 10 20 8
% 50% 56% 40%
Over
all
24/50 = 48%
• Realistically: Do item analysis
to your most important tests
–end of unit tests, final exams -->
summative evaluation
–common exams with other
teachers
(departmentalized exam)
• common exams gives bigger
sample to work with, which is
good
• makes sure that questions other
teacher wrote are working for
YOUR class
ITEM ANALYSIS is one area
where even a lot of otherwise
very good classroom teachers
fall down:
–they think they're doing a good
job;
–they think they've doing good
evaluation;
–but without doing item analysis,
–they can't really know
ITEM ANALYSIS is not
an end in itself,
–no point unless you
use it to revise items,
and
–help students on
basis of information
you get out of it.
END OF PRESENTATION…
THANK U FOR LISTENING…
HAVE A RELIABLE AND
ENJOYABLE DAY….
L/O/G/O

Mais conteúdo relacionado

Mais procurados

Development of classroom assessment tools
Development of classroom assessment toolsDevelopment of classroom assessment tools
Development of classroom assessment tools
Eaicz12
 
performance based -product oriented assessment
performance based -product oriented assessmentperformance based -product oriented assessment
performance based -product oriented assessment
echem101
 

Mais procurados (20)

Authentic Assessment Tools
Authentic Assessment ToolsAuthentic Assessment Tools
Authentic Assessment Tools
 
Grading and reporting
Grading and reportingGrading and reporting
Grading and reporting
 
Analyzing and using test item data
Analyzing and using test item dataAnalyzing and using test item data
Analyzing and using test item data
 
Constructing Test Questions and the Table of Specifications (TOS)
Constructing Test Questions and the Table of Specifications (TOS)Constructing Test Questions and the Table of Specifications (TOS)
Constructing Test Questions and the Table of Specifications (TOS)
 
Test Construction
Test ConstructionTest Construction
Test Construction
 
Development of classroom assessment tools
Development of classroom assessment toolsDevelopment of classroom assessment tools
Development of classroom assessment tools
 
Completion type of test
Completion type of testCompletion type of test
Completion type of test
 
Assessment of student learning 1
Assessment of student learning 1Assessment of student learning 1
Assessment of student learning 1
 
K to 12 Basic Education Curriculum
K to 12 Basic Education CurriculumK to 12 Basic Education Curriculum
K to 12 Basic Education Curriculum
 
Matching type
Matching typeMatching type
Matching type
 
Table of specifications
Table of specificationsTable of specifications
Table of specifications
 
preparing a TOS
preparing a TOSpreparing a TOS
preparing a TOS
 
Types of test questions
Types of test questionsTypes of test questions
Types of test questions
 
performance based -product oriented assessment
performance based -product oriented assessmentperformance based -product oriented assessment
performance based -product oriented assessment
 
Assessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice TestAssessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice Test
 
The teacher and the School Curriculum
The teacher and the School CurriculumThe teacher and the School Curriculum
The teacher and the School Curriculum
 
Type of Test
Type of TestType of Test
Type of Test
 
TOS in Science III and Sample Test
TOS in Science III and Sample TestTOS in Science III and Sample Test
TOS in Science III and Sample Test
 
GRASP PERFORMANCE ASSESSMENT
GRASP PERFORMANCE ASSESSMENTGRASP PERFORMANCE ASSESSMENT
GRASP PERFORMANCE ASSESSMENT
 
Teaching Strategies, Approaches and Methods Under DepEd K-12 Curriculum
Teaching Strategies, Approaches and Methods Under DepEd K-12 CurriculumTeaching Strategies, Approaches and Methods Under DepEd K-12 Curriculum
Teaching Strategies, Approaches and Methods Under DepEd K-12 Curriculum
 

Destaque (10)

Problems and Issues in the Philippine Educational System
Problems and Issues in the Philippine Educational SystemProblems and Issues in the Philippine Educational System
Problems and Issues in the Philippine Educational System
 
R Activity in Biostatistics
R Activity in BiostatisticsR Activity in Biostatistics
R Activity in Biostatistics
 
Research methods in education_dr. paglinawan
Research methods in education_dr. paglinawanResearch methods in education_dr. paglinawan
Research methods in education_dr. paglinawan
 
Action research - Basic concepts
Action research - Basic conceptsAction research - Basic concepts
Action research - Basic concepts
 
Edgar Dale's Cone of Experience
Edgar Dale's Cone of ExperienceEdgar Dale's Cone of Experience
Edgar Dale's Cone of Experience
 
Handicrafts (2)
Handicrafts (2)Handicrafts (2)
Handicrafts (2)
 
ARALING PANLIPUNAN
ARALING PANLIPUNANARALING PANLIPUNAN
ARALING PANLIPUNAN
 
Handicraft (2)
Handicraft (2)Handicraft (2)
Handicraft (2)
 
Printed Media
Printed MediaPrinted Media
Printed Media
 
Item analysis
Item analysis Item analysis
Item analysis
 

Semelhante a Item analysis

item analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysisitem analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysis
swatisheth8
 
Lesson 21 designing the questionaire and establishing validity and reliabilty
Lesson 21 designing the questionaire and establishing validity and reliabiltyLesson 21 designing the questionaire and establishing validity and reliabilty
Lesson 21 designing the questionaire and establishing validity and reliabilty
mjlobetos
 

Semelhante a Item analysis (20)

Item analysis with spss software
Item analysis with spss softwareItem analysis with spss software
Item analysis with spss software
 
DEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRY
DEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRYDEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRY
DEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRY
 
item analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysisitem analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysis
 
Teaching technology2
Teaching technology2Teaching technology2
Teaching technology2
 
Chapter24
Chapter24Chapter24
Chapter24
 
Lesson 21 designing the questionaire and establishing validity and reliabilty
Lesson 21 designing the questionaire and establishing validity and reliabiltyLesson 21 designing the questionaire and establishing validity and reliabilty
Lesson 21 designing the questionaire and establishing validity and reliabilty
 
Test appraisal
Test appraisalTest appraisal
Test appraisal
 
Item analysis2
Item analysis2Item analysis2
Item analysis2
 
Item_Analysis.pptx
Item_Analysis.pptxItem_Analysis.pptx
Item_Analysis.pptx
 
Item analysis
Item analysisItem analysis
Item analysis
 
New item analysis
New item analysisNew item analysis
New item analysis
 
Analyzing and using test item data
Analyzing and using test item dataAnalyzing and using test item data
Analyzing and using test item data
 
Analyzing and using test item data
Analyzing and using test item dataAnalyzing and using test item data
Analyzing and using test item data
 
Analyzing and using test item data
Analyzing and using test item dataAnalyzing and using test item data
Analyzing and using test item data
 
Evaluation of Measurement Instruments.ppt
Evaluation of Measurement Instruments.pptEvaluation of Measurement Instruments.ppt
Evaluation of Measurement Instruments.ppt
 
MCQ test item analysis
MCQ test item analysisMCQ test item analysis
MCQ test item analysis
 
Characteristics of a Good Test
Characteristics of a Good TestCharacteristics of a Good Test
Characteristics of a Good Test
 
Testing and evaluation
Testing and evaluationTesting and evaluation
Testing and evaluation
 
TEST CONSTRUCTION in Psychology to measure different traits
TEST CONSTRUCTION in Psychology to measure different traitsTEST CONSTRUCTION in Psychology to measure different traits
TEST CONSTRUCTION in Psychology to measure different traits
 
Test Blueprinting and Exam Revision
Test Blueprinting and Exam Revision Test Blueprinting and Exam Revision
Test Blueprinting and Exam Revision
 

Último

Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
QucHHunhnh
 

Último (20)

FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptx
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and Modifications
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Dyslexia AI Workshop for Slideshare.pptx
Dyslexia AI Workshop for Slideshare.pptxDyslexia AI Workshop for Slideshare.pptx
Dyslexia AI Workshop for Slideshare.pptx
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 

Item analysis

  • 3. QUESTIONS? Click to end Contact Me…!!! ©James L. Paglinawan TM ® e-mail me @: JLPaglinawan@cmu.edu.ph JLPaglinawan@teacher.deped.gov.ph
  • 4. Item Analysis Techniques to improve test items and instruction
  • 5. 11. Administer Tests 12. Conduct Item Analysis 1. Review National and Profession al Standards 7. Develop New Test Questions 8. Review Test Questions 13. Standard Setting Study 14. Set Passing Standard 9. Assemble Operational Test Forms 10. Produce Printed Tests Mat. 2. Convene National Advisory Committee 3. Develop Domain, Knowledge and Skills Statements 4. Conduct Job Analysis 5. Establish Test Specifications 6. Develop Test Design Test Development Process
  • 6. What is Item Analysis ? • process that examines student responses to individual test items assessing quality items and the test as a whole • valuable in improving items which will be used again in later tests and eliminate ambiguous or misleading items • valuable for increasing instructors' skills in test construction, and • identifying specific areas of course content which need greater emphasis or clarity.
  • 7. Several Purposes 1. More diagnostic information on students – Classroom level: • determine questions most found very difficult/ guessing on, - – reteach that concept • questions all got right – – don't waste more time on this area • find wrong answers students are choosing- – identify common misconceptions – Individual level: • isolate specific errors this student made
  • 8. 2.Build future tests, revise test items to make them better • know how much work in writing good questions • SHOULD NOT REUSE WHOLE TESTS --> diagnostic teaching means responding to needs of students, so after a few years a test bank is build up and choose a tests for the class • can spread difficulty levels across your blueprint (TOS)
  • 9. 3. Part of continuing professional development – doing occasional item analysis will help become a better test writer – documenting just how good your evaluation is – useful for dealing with parents or administrators if there's ever a dispute – once you start bringing out all these impressive looking stats, parents and administrators will believe why some students failed.
  • 10. Classical ITEM Analysis Statistics • Reliability (test level statistic) • Difficulty (item level statistic) • Discrimination (item level statistic)
  • 11. Test level statistic Quality of the Test • Reliability and Validity –Reliability Consistency of measurement –Validity Truthfulness of response Overall Test Quality Individual Item Quality
  • 12. Reliability refers to the extent to which the test is likely to produce consistent scores. Characteristics: 1. The intercorrelations among the items -- the greater/stronger the relative number of positive relationships are, the greater the reliability. 2. The length of the test – a test with more items will have a higher reliability, all other things being equal.
  • 13. 3. The content of the test -- generally, the more diverse the subject matter tested and the testing techniques used, the lower the reliability. 4. Heterogeneous groups of test takers
  • 14. Types of reliability • Stability 1. Test – Retest
  • 15. • Stability 2. Inter – rater / Observer/ Scorer • applicable for mostly essay questions • Use Cohen’s Kappa Statistic
  • 16. • Equivalence 3. Parallel-Forms/ Equivalent Used to assess the consistency of the results of two tests constructed in the same way from the same content domain.
  • 17. • Internal Consistency • Used to assess the consistency of results across items within a test. 4. Split – Half
  • 18. • 5. Kuder-Richardson Formula 20 / 21 Correlation is determined from a single administration of a test through a study of score variances
  • 19. • 6. Cronbach's Alpha (a)
  • 20. Reliability Indices Interpretation .91 and above Excellent reliability; at the level of the best standardized tests .81 - .90 Very good for a classroom test .71 - .80 Good for a classroom test; in the range of most. There are probably a few items which could be improved. .61 - .70 Somewhat low. This test needs to be supplemented by other measures (e.g., more tests) to determine grades. There are probably some items which could be improved. .51 - .60 Suggests need for revision of test, unless it is quite short (ten or fewer items). The test definitely needs to be supplemented by other measures (e.g., more tests) for grading. .50 or below Questionable reliability. This test should not contribute heavily to the course grade, and it needs revision.
  • 21. Test Item statistic  Item Difficulty Percent answering correctly  Item Discrimination How well the item "functions“ How “valid” the item is based on the total test score criterion
  • 22. WHAT IS A WELL- FUNCTIONING TEST ITEM? • how many students got it correct? (DIFFICULTY) • which students got it correct? (DISCRIMINATION)
  • 23. Three important information on quality of test items • Item difficulty: measure whether an item was too easy or too hard. • Item discrimination: measure whether an item discriminated between students who knew the material well and students who did not. • Effectiveness of alternatives: Determination whether distracters (incorrect but plausible answers) tend to be marked by the less able students and not by the more able students.
  • 24. Item Difficulty • Item difficulty is simply the percentage of students who answer an item correctly. In this case, it is also equal to the item mean. Diff = # of students choosing correctly total # of students • The item difficulty index ranges from 0 to 100; the higher the value, the easier the question.
  • 25. Item Difficulty Level: Definition The percentage of students who answered the item correctly. High (Difficult) Medium (Moderate) Low (Easy ) <= 30% > 30% AND < 80% >=80 % 0 10 20 30 40 50 60 70 80 90 100
  • 26. Item Difficulty Level: Sample Item No. No. Correct Answers % Correct Difficulty Level 1 15 2 25 3 35 4 45 Number of students who answered each item = 50 30 High 50 Medium 70 Medium 90 Low
  • 27. Item Difficulty Level: Questions/Discussion • Is a test that nobody failed too easy? • Is a test on which nobody got 100% too difficult? • Should items that are “too easy” or “too difficult” be thrown out?
  • 28. Item Discrimination • Traditionally, using high and low scoring groups (upper 27 % and lower 27%) • Computerized analyses provide more accurate assessment of the discrimination power of items since it accounts all responses rather than just high and low scoring groups. • Equivalent to point-biserial correlation. It provides estimate the degree an individual item is measuring the same thing as the rest of the items.
  • 29. What is Item Discrimination? • Generally, students who did well on the exam should select the correct answer to any given item on the exam. • The Discrimination Index distinguishes for each item between the performance of students who did well on the exam and students who did poorly.
  • 30. Indices of Difficulty and Discrimination (by Hopkins and Antes) Index Difficulty Discrimination 0.86 above Very Easy To be discarded 0.71 – 0.85 Easy To be revised 0.30 – 0.70 Moderate Very Good items 0.15 – 0.29 Difficult To be revised 0.14 below Very Difficult To be discarded
  • 31. Item Discrimination: Questions / Discussion • What factors could contribute to low item discrimination between the two groups of students? • What is a likely cause for a negative discrimination index?
  • 33. Sample TOS Remember Understand Apply Total Section A 4 (1,3,7,9) 6 10 20 Section B 5 (2,5,8,11,15) 5 4 14 Section C 3 (6,17,21) 7 6 16 Total 12 18 20 50
  • 34. Steps in Item analysis 1. Code the test items: - 1 for correct and 0 for incorrect - Vertical – columns (item numbers) - Horizontal – rows (respondents/students)
  • 35. TEST ITEMS No. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 . . . . 50 1 1 0 1 1 1 0 0 0 0 1 1 1 0 0 0 0 0 1 1 2 1 1 0 1 1 1 0 1 1 1 0 1 1 1 1 0 1 1 1 3 0 0 0 1 0 0 0 1 0 0 0 1 1 1 1 1 1 1 0 4 0 1 0 0 0 1 0 0 0 1 0 0 1 0 0 0 1 0 0 5 1 0 1 1 1 0 1 1 1 0 1 1 0 1 1 1 0 1 0 6 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 1 1 0 1 7 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 8 1 1 0 1 1 1 0 1 1 1 0 1 1 0 0 0 1 0 0
  • 36. 2. IN SPSS: Analyze  Scale  Reliability analysis  (drag/place variables to Item box)  Statistics  Scale if item deleted  ok.
  • 37. • ****** Method 1 (space saver) will be used for this analysis ****** • R E L I A B I L I T Y A N A L Y S I S - S C A L E (A L P H A) • Item-total Statistics • Scale Scale Corrected • Mean Variance Item- Alpha • if Item if Item Total if Item • Deleted Deleted Correlation Deleted • VAR00001 14.4211 127.1053 .9401 .9502 • VAR00002 14.6316 136.8440 .7332 .9542 • VAR00003 14.4211 141.5695 .4774 .9574 • VAR00004 14.4737 128.6109 .6511 .9508 • VAR00005 14.4737 128.8252 .8274 .9509 • VAR00006 14.0526 130.6579 .2236 .9525 • VAR00007 14.2105 127.8835 .2533 .9511 • VAR00008 14.1053 128.6673 .1906 .9515 • VAR00009 14.4211 129.1410 .7311 .9513 • ..................... • VAR00022 14.4211 129.1410 .7311 .9513 • VAR00023 14.4211 127.1053 .4401 .9502 • VAR00024 14.6316 136.8440 -.0332 .9542 • VAR00047 14.4737 128.6109 .8511 .9508 • VAR00048 14.4737 128.8252 .8274 .9509 • VAR00049 14.0526 130.6579 .5236 .9525 • VAR00050 14.2105 127.8835 .7533 .9511 • Reliability Coefficients • N of Cases = 57.0 N of Items = 50 • Alpha = .9533
  • 38. 3. In the output dialog box: • Alpha placed at the bottom • the corrected item total correlation is the point biserial correlation as bases for index of test reliability
  • 39. 4. Count the number of items discarded and fill up summary item analysis table.
  • 40. Test Item Reliability Analysis Summary (sample) Test Level of Difficulty Number of Items % Item Number Math Very Easy 1 2 1 (50 items) Easy 2 4 2,5 Moderate 10 20 3,4,10,15… Difficult 30 60 6,7,8,9,11,… Very Difficult 7 14 16,24,32…
  • 41. 5. Count the number of items retained based on the cognitive domains in the TOS. Compute the percentage per level of difficulty.
  • 42. Remember Understand Apply N Ret N Ret N Ret A 4 1 6 3 10 3 B 5 3 5 3 4 2 C 3 2 7 4 6 3 Total 12 6 18 10 20 8 % 50% 56% 40% Over all 24/50 = 48%
  • 43. • Realistically: Do item analysis to your most important tests –end of unit tests, final exams --> summative evaluation –common exams with other teachers (departmentalized exam) • common exams gives bigger sample to work with, which is good • makes sure that questions other teacher wrote are working for YOUR class
  • 44. ITEM ANALYSIS is one area where even a lot of otherwise very good classroom teachers fall down: –they think they're doing a good job; –they think they've doing good evaluation; –but without doing item analysis, –they can't really know
  • 45. ITEM ANALYSIS is not an end in itself, –no point unless you use it to revise items, and –help students on basis of information you get out of it.
  • 46. END OF PRESENTATION… THANK U FOR LISTENING… HAVE A RELIABLE AND ENJOYABLE DAY….