SlideShare a Scribd company logo
1 of 29
COLLABORATIVE EXAMINATION ITEM
REVIEW PROCESS IN A TEAM-TAUGHT
COURSE
Adam Pate, PharmD; David J. Caldwell,
PharmD; Laurel Sampognaro, PharmD
Objectives
1. Describe how to implement a collaborative item review process
2. Identify potential barriers to implementation and success of a
collaborative item writing process
3. Generate ideas to establish a collaborative process at your
respective institution
Baseline information about us
A review process was created to limit variability of
content and quality of items in the course
The first meeting is crucial in setting the tone
for the semester
All instructors relied on principles from
Haladyna et. al. to define “good” items
Real quick item statistics
• Point biserial correlation (rpb)
• Difficulty (p)
Item classification guide
Item Class Item Difficulty
Item Discrimination
(point biserial)
Description
Level I 0.45 to 0.75 +0.20 or higher
Best item statistics; use most
items in this range if possible
Level II 0.76 to 0.91 +0.15 or higher Easy; use sparingly
Level III 0.25 to 0.44 +0.10 or higher
Difficulty; use very sparingly and
only if content is essential--rewrite
if possible
Level IV <0.24 or >0.91 Any discrimination
Extremely difficult or easy; do not
use unless content is essential
Faculty discussed these guidelines and
came to group consensus on usage
Guideline For (%) Uncited (%) Against (%)
Use positives, no negatives 63 19 18
Write as many plausible
distractors as you can
70 26 4
Use carefully None of the Above 44 7 48
Avoid All of the Above 70 7 22
Use humor sparingly 0 85 15
Item Review Process
• Who is involved?
• Self Care I – 11 faculty
• Self Care II – 9 faculty
• Participation
Item Review Process
Deadline for
exam items
Test creation
and
dissemination
Meeting
preparation
Item Review Process
• Meeting details
• Environment
• Discussion
• Agreement
• Follow-up after meeting
Item Review Process
• What worked well?
Item Review Process
• Barriers to implementation
Self Care 1
(without training or review)
Self Care 2
(without training or review)
Spring and Fall 2012, control sequence
6 exams, 272 items
(without training or review)
Spring and Fall 2013, intervention sequence
Self Care 1
(with training or review)
Self Care 2
(with training or review)
6 exams, 264 items
(without training or review)
Interventions
1. Pre-semester survey
2. Presentation of item-writing
guidelines at semester start
3. Guideline review and discussion
at each exam review meeting
4. Review and editing of exam
items per guidelines
5. Post-semester survey
Comparisons
item difficulty, discrimination, and
classification by these factors, and
student performance
All interventions were completed in both
Self Care 1 and 2; instructors teaching in
both only completed surveys in
Self Care 1
Comparisons
Pre- versus post-survey
About the participating faculty
NOTABLE BASELINE REPORTS
Which of the following factors affect your
sense of success in item writing?
• Item statistics (n=9)
• Previous training in item writing (n=4)
• Student challenges to exam items (n=3)
How often have you participated in peer-
review of exam items?
• Half of the time (n=4)
• A minority of the time (n=1)
• Never (n=5)
Results
GOALS
1. To improve examination quality through a faculty development program,
followed by a longitudinal item review occurring before examination
administration
2. To improve faculty members’ self-rated confidence and success
3. To measure changes in their opinions regarding item-writing guidelines
and review.
Goal 1: What happened to item-quality?
Results
• No significant difference between
the control and intervention items,
respectively
• Mean student scores (% ± SD)
did change (p<0.001):
• Control sequence, 88.3 ± 4.5
• Intervention sequence 85.6 ± 6.0
Distribution by level
CONTROL ITEMS INTERVENTION ITEMS
Distribution by level
ITEM CLASS WITH REVIEW, n (%) WITHOUT REVIEW, n (%)
Level 1 31 (11.4) 52 (19.7)
Level 2 70 (27.5) 76 (28.8)
Level 3 7 (2.6) 3 (1.1)
Level 4 142 (52.2) 122 (46.2)
Uncategorizable 22 (8.1) 11 (4.2)
Goal 2: How did participants’ self-rated
confidence and success change?
Survey opinions: self-focused
ITEM Pre (mean) Post (mean) p-value
How would you rate your confidence at writing effective
multiple-choice test items?
(0 – very unconfident, 10 – very confident)
6.0 8.1 0.002
How would you rate your success at writing multiple-
choice test items?
(0 – very unsuccessful, 10 – very successful)
6.4 7.9 <0.001
To what degree do you feel confident that you can
properly evaluate your and your colleagues' test
questions?
(0 – very unconfident, 10 – very confident)
6.7 8.4 0.005
To what degree do you feel confident that you could
implement a formal exam item evaluation process
as a coordinator of another course?
(0 – very unconfident, 10 – very confident)
5.5 7.1 0.008
Goal 3: How did participants’ opinions of
item guidelines and review change?
Survey opinions: item-focused
ITEM Pre (mean) Post (mean) p-value
In your opinion, to what degree will peer-review
of exam items affect item quality?
(0 – very negatively, 10 – very positively)
7.9 8.5 0.14
Do you plan to modify future multiple-choice
items based on item-writing guidelines?
(1 – Yes, 2 – No)
9 Y
1 N
10 Y
0 N
1.00
In your opinion, to what degree will voluntary
application of item-writing guidelines affect
item quality?
(0 – very negatively, 10 – very positively)
7.9 8.4 0.24
Other findings
Top 5 item flaws
GUIDELINES n (% of total changes)
Include the central idea in the stem instead of the choices. 37 (33.6)
Use correct grammar, punctuation, capitalization, and spelling. 17 (15.5)
Minimize the amount of reading in each stem. 13 (11.8)
Use the question, completion, and best answer versions of the conventional
multiple choice (MC), the alternate choice, true-false, multiple true-false,
matching, and the context-dependent item and item set formats, but avoid the
complex MC (Type K) format.
10 (9.1)
Keep choices independent; choices should not be overlapping. 7 (6.4)
Comparison of undesirable format frequency
Item Review Process
• Experience in other courses

More Related Content

What's hot

Administering,scoring and reporting a test ppt
Administering,scoring and reporting a test pptAdministering,scoring and reporting a test ppt
Administering,scoring and reporting a test ppt
Manali Solanki
 
Designing and conducting summative evaluations
Designing and conducting summative evaluationsDesigning and conducting summative evaluations
Designing and conducting summative evaluations
Larry Cobb
 
Assessment and individual differences
Assessment and individual differencesAssessment and individual differences
Assessment and individual differences
Sullivan Turner
 
Design Chapter 7 - Testing and Evaluation Techniques
Design Chapter 7 - Testing and Evaluation TechniquesDesign Chapter 7 - Testing and Evaluation Techniques
Design Chapter 7 - Testing and Evaluation Techniques
guest01bdf1
 
Developing assessment instruments
Developing assessment instrumentsDeveloping assessment instruments
Developing assessment instruments
Larry Cobb
 

What's hot (20)

The Evaluation Checklist
The Evaluation ChecklistThe Evaluation Checklist
The Evaluation Checklist
 
Developing Assessment Instruments
Developing Assessment InstrumentsDeveloping Assessment Instruments
Developing Assessment Instruments
 
Administering,scoring and reporting a test ppt
Administering,scoring and reporting a test pptAdministering,scoring and reporting a test ppt
Administering,scoring and reporting a test ppt
 
Src Voc
Src VocSrc Voc
Src Voc
 
Developing assessment instruments
Developing assessment instrumentsDeveloping assessment instruments
Developing assessment instruments
 
Rubric design workshop
Rubric design workshopRubric design workshop
Rubric design workshop
 
Designing and conducting summative evaluations
Designing and conducting summative evaluationsDesigning and conducting summative evaluations
Designing and conducting summative evaluations
 
Assessment and individual differences
Assessment and individual differencesAssessment and individual differences
Assessment and individual differences
 
Rubrics for performance assessment
Rubrics for performance assessmentRubrics for performance assessment
Rubrics for performance assessment
 
Design Chapter 7 - Testing and Evaluation Techniques
Design Chapter 7 - Testing and Evaluation TechniquesDesign Chapter 7 - Testing and Evaluation Techniques
Design Chapter 7 - Testing and Evaluation Techniques
 
Test Administration, Test administration, Test-taking Strategies
Test Administration, Test administration, Test-taking StrategiesTest Administration, Test administration, Test-taking Strategies
Test Administration, Test administration, Test-taking Strategies
 
Properties of-assessment-methods
Properties of-assessment-methodsProperties of-assessment-methods
Properties of-assessment-methods
 
Developing assessment instruments
Developing assessment instrumentsDeveloping assessment instruments
Developing assessment instruments
 
Chapter 12
Chapter 12Chapter 12
Chapter 12
 
Designing and Conducting Formative Evaluations
Designing and Conducting Formative EvaluationsDesigning and Conducting Formative Evaluations
Designing and Conducting Formative Evaluations
 
Rubric Development for Teachers
Rubric Development for TeachersRubric Development for Teachers
Rubric Development for Teachers
 
Improve your test item writing skills to help create better nursing exams
Improve your test item writing skills to help create better nursing examsImprove your test item writing skills to help create better nursing exams
Improve your test item writing skills to help create better nursing exams
 
Test Blueprinting and Exam Revision
Test Blueprinting and Exam Revision Test Blueprinting and Exam Revision
Test Blueprinting and Exam Revision
 
Interrogating evaluation 2015 induction
Interrogating evaluation 2015 inductionInterrogating evaluation 2015 induction
Interrogating evaluation 2015 induction
 
Colorado assessment summit_teacher_eval
Colorado assessment summit_teacher_evalColorado assessment summit_teacher_eval
Colorado assessment summit_teacher_eval
 

Similar to Collaborative Examination Item Review Process in a Team-Taught Course

NED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & RubricsNED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & Rubrics
Carmina Gurrea
 
Faculty assessment presentation april 2014
Faculty assessment presentation april 2014Faculty assessment presentation april 2014
Faculty assessment presentation april 2014
Rita Ndagire Kizito
 
Tqf day 2 - assessment and feedback
Tqf   day 2 - assessment and feedbackTqf   day 2 - assessment and feedback
Tqf day 2 - assessment and feedback
RMIT
 
Training &n Development Studies & Evaluation
Training &n Development Studies & EvaluationTraining &n Development Studies & Evaluation
Training &n Development Studies & Evaluation
ElsaCherian1
 

Similar to Collaborative Examination Item Review Process in a Team-Taught Course (20)

Criterion-Referenced Assessment Review
Criterion-Referenced Assessment ReviewCriterion-Referenced Assessment Review
Criterion-Referenced Assessment Review
 
NED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & RubricsNED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & Rubrics
 
Rubric
Rubric Rubric
Rubric
 
Evaluation the many faces
Evaluation   the many facesEvaluation   the many faces
Evaluation the many faces
 
Faculty assessment presentation april 2014
Faculty assessment presentation april 2014Faculty assessment presentation april 2014
Faculty assessment presentation april 2014
 
Evaluation
EvaluationEvaluation
Evaluation
 
Chapter 4: Evaluating the curriculum
Chapter 4: Evaluating the curriculumChapter 4: Evaluating the curriculum
Chapter 4: Evaluating the curriculum
 
Student Growth Measures
Student Growth MeasuresStudent Growth Measures
Student Growth Measures
 
Teacher evaluation and goal setting connecticut
Teacher evaluation and goal setting   connecticutTeacher evaluation and goal setting   connecticut
Teacher evaluation and goal setting connecticut
 
Developing effective safety training تطوير التدريب في مجال السلامة
Developing effective safety training تطوير التدريب في مجال السلامةDeveloping effective safety training تطوير التدريب في مجال السلامة
Developing effective safety training تطوير التدريب في مجال السلامة
 
Self-, peer-, and instructor-assessment from Bloom’s perspective
Self-, peer-, and instructor-assessment from Bloom’s perspective Self-, peer-, and instructor-assessment from Bloom’s perspective
Self-, peer-, and instructor-assessment from Bloom’s perspective
 
Quality, Who Says
Quality, Who SaysQuality, Who Says
Quality, Who Says
 
Using Nursing Exam Data Effectively in Preparing Nursing Accreditation
Using Nursing Exam Data Effectively in Preparing Nursing AccreditationUsing Nursing Exam Data Effectively in Preparing Nursing Accreditation
Using Nursing Exam Data Effectively in Preparing Nursing Accreditation
 
Test construction
Test constructionTest construction
Test construction
 
Tqf day 2 - assessment and feedback
Tqf   day 2 - assessment and feedbackTqf   day 2 - assessment and feedback
Tqf day 2 - assessment and feedback
 
Test appraisal
Test appraisalTest appraisal
Test appraisal
 
Check, Check, Check in the Simulation Lab
Check, Check, Check in the Simulation LabCheck, Check, Check in the Simulation Lab
Check, Check, Check in the Simulation Lab
 
Training &n Development Studies & Evaluation
Training &n Development Studies & EvaluationTraining &n Development Studies & Evaluation
Training &n Development Studies & Evaluation
 
Rubric
RubricRubric
Rubric
 
Revising Instructional Materials
Revising Instructional MaterialsRevising Instructional Materials
Revising Instructional Materials
 

More from ExamSoft

Programs Coming Together Using ExamSoft to assess interprofessional education...
Programs Coming Together Using ExamSoft to assess interprofessional education...Programs Coming Together Using ExamSoft to assess interprofessional education...
Programs Coming Together Using ExamSoft to assess interprofessional education...
ExamSoft
 

More from ExamSoft (20)

More Than Assessment: Using computer-based testing software to deliver instru...
More Than Assessment: Using computer-based testing software to deliver instru...More Than Assessment: Using computer-based testing software to deliver instru...
More Than Assessment: Using computer-based testing software to deliver instru...
 
Creating Test Items Using NCLEX® Alternative Item Types
Creating Test Items Using NCLEX® Alternative Item TypesCreating Test Items Using NCLEX® Alternative Item Types
Creating Test Items Using NCLEX® Alternative Item Types
 
Psychometrics 201: Putting assessment data into action
Psychometrics 201: Putting assessment data into actionPsychometrics 201: Putting assessment data into action
Psychometrics 201: Putting assessment data into action
 
Programs Coming Together Using ExamSoft to assess interprofessional education...
Programs Coming Together Using ExamSoft to assess interprofessional education...Programs Coming Together Using ExamSoft to assess interprofessional education...
Programs Coming Together Using ExamSoft to assess interprofessional education...
 
5 Tips for Course Alignment: Improve student outcomes while mapping your curr...
5 Tips for Course Alignment: Improve student outcomes while mapping your curr...5 Tips for Course Alignment: Improve student outcomes while mapping your curr...
5 Tips for Course Alignment: Improve student outcomes while mapping your curr...
 
Last Minute Tips to Positively Impact Your Students This Semester
Last Minute Tips to Positively Impact Your Students This SemesterLast Minute Tips to Positively Impact Your Students This Semester
Last Minute Tips to Positively Impact Your Students This Semester
 
Using the NCLEX RN Blueprint as a Guide for Testing in a Nursing Baccalaureat...
Using the NCLEX RN Blueprint as a Guide for Testing in a Nursing Baccalaureat...Using the NCLEX RN Blueprint as a Guide for Testing in a Nursing Baccalaureat...
Using the NCLEX RN Blueprint as a Guide for Testing in a Nursing Baccalaureat...
 
Retiring Exam Questions? How to Use These Items in Formative Assessments
Retiring Exam Questions? How to Use These Items in Formative AssessmentsRetiring Exam Questions? How to Use These Items in Formative Assessments
Retiring Exam Questions? How to Use These Items in Formative Assessments
 
Using ExamSoft to Evaluate NCLEX Test Plan Success
Using ExamSoft to Evaluate NCLEX Test Plan SuccessUsing ExamSoft to Evaluate NCLEX Test Plan Success
Using ExamSoft to Evaluate NCLEX Test Plan Success
 
Using ExamSoft Rubrics to Assess Student Medical Research
Using ExamSoft Rubrics to Assess Student Medical Research Using ExamSoft Rubrics to Assess Student Medical Research
Using ExamSoft Rubrics to Assess Student Medical Research
 
Psychometrics 101: Know What Your Assessment Data is Telling You
Psychometrics 101: Know What Your Assessment Data is Telling YouPsychometrics 101: Know What Your Assessment Data is Telling You
Psychometrics 101: Know What Your Assessment Data is Telling You
 
From Conception to Execution: Strategies for designing and implementing a com...
From Conception to Execution: Strategies for designing and implementing a com...From Conception to Execution: Strategies for designing and implementing a com...
From Conception to Execution: Strategies for designing and implementing a com...
 
What's in it for me-- said the student, faculty, and curriculum
 What's in it for me-- said the student, faculty, and curriculum What's in it for me-- said the student, faculty, and curriculum
What's in it for me-- said the student, faculty, and curriculum
 
"What's in it for me?" said the Student, Faculty, and Curriculum
"What's in it for me?" said the Student, Faculty, and Curriculum"What's in it for me?" said the Student, Faculty, and Curriculum
"What's in it for me?" said the Student, Faculty, and Curriculum
 
Closing the Loop on Clinical Competency Based Assessments
 Closing the Loop on Clinical Competency Based Assessments Closing the Loop on Clinical Competency Based Assessments
Closing the Loop on Clinical Competency Based Assessments
 
Using Exam Data for Scholarly Activities
Using Exam Data for Scholarly Activities Using Exam Data for Scholarly Activities
Using Exam Data for Scholarly Activities
 
Communication is Key! Using ExamSoft to Keep Everyone Involved In the Teachin...
Communication is Key! Using ExamSoft to Keep Everyone Involved In the Teachin...Communication is Key! Using ExamSoft to Keep Everyone Involved In the Teachin...
Communication is Key! Using ExamSoft to Keep Everyone Involved In the Teachin...
 
Stop the cheating! best practices to minimize security risks on exams
Stop the cheating! best practices to minimize security risks on examsStop the cheating! best practices to minimize security risks on exams
Stop the cheating! best practices to minimize security risks on exams
 
Using ExamSoft Data for Item Revision and Faculty Development
Using ExamSoft Data for Item Revision and Faculty DevelopmentUsing ExamSoft Data for Item Revision and Faculty Development
Using ExamSoft Data for Item Revision and Faculty Development
 
Telling Your Assessment Story with ExamSoft Data
Telling Your Assessment Story with ExamSoft DataTelling Your Assessment Story with ExamSoft Data
Telling Your Assessment Story with ExamSoft Data
 

Recently uploaded

Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
kauryashika82
 

Recently uploaded (20)

Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdf
 

Collaborative Examination Item Review Process in a Team-Taught Course

  • 1. COLLABORATIVE EXAMINATION ITEM REVIEW PROCESS IN A TEAM-TAUGHT COURSE Adam Pate, PharmD; David J. Caldwell, PharmD; Laurel Sampognaro, PharmD
  • 2. Objectives 1. Describe how to implement a collaborative item review process 2. Identify potential barriers to implementation and success of a collaborative item writing process 3. Generate ideas to establish a collaborative process at your respective institution
  • 4. A review process was created to limit variability of content and quality of items in the course
  • 5. The first meeting is crucial in setting the tone for the semester
  • 6. All instructors relied on principles from Haladyna et. al. to define “good” items
  • 7. Real quick item statistics • Point biserial correlation (rpb) • Difficulty (p)
  • 8. Item classification guide Item Class Item Difficulty Item Discrimination (point biserial) Description Level I 0.45 to 0.75 +0.20 or higher Best item statistics; use most items in this range if possible Level II 0.76 to 0.91 +0.15 or higher Easy; use sparingly Level III 0.25 to 0.44 +0.10 or higher Difficulty; use very sparingly and only if content is essential--rewrite if possible Level IV <0.24 or >0.91 Any discrimination Extremely difficult or easy; do not use unless content is essential
  • 9. Faculty discussed these guidelines and came to group consensus on usage Guideline For (%) Uncited (%) Against (%) Use positives, no negatives 63 19 18 Write as many plausible distractors as you can 70 26 4 Use carefully None of the Above 44 7 48 Avoid All of the Above 70 7 22 Use humor sparingly 0 85 15
  • 10. Item Review Process • Who is involved? • Self Care I – 11 faculty • Self Care II – 9 faculty • Participation
  • 11. Item Review Process Deadline for exam items Test creation and dissemination Meeting preparation
  • 12. Item Review Process • Meeting details • Environment • Discussion • Agreement • Follow-up after meeting
  • 13. Item Review Process • What worked well?
  • 14. Item Review Process • Barriers to implementation
  • 15. Self Care 1 (without training or review) Self Care 2 (without training or review) Spring and Fall 2012, control sequence 6 exams, 272 items (without training or review) Spring and Fall 2013, intervention sequence Self Care 1 (with training or review) Self Care 2 (with training or review) 6 exams, 264 items (without training or review) Interventions 1. Pre-semester survey 2. Presentation of item-writing guidelines at semester start 3. Guideline review and discussion at each exam review meeting 4. Review and editing of exam items per guidelines 5. Post-semester survey Comparisons item difficulty, discrimination, and classification by these factors, and student performance All interventions were completed in both Self Care 1 and 2; instructors teaching in both only completed surveys in Self Care 1 Comparisons Pre- versus post-survey
  • 16. About the participating faculty NOTABLE BASELINE REPORTS Which of the following factors affect your sense of success in item writing? • Item statistics (n=9) • Previous training in item writing (n=4) • Student challenges to exam items (n=3) How often have you participated in peer- review of exam items? • Half of the time (n=4) • A minority of the time (n=1) • Never (n=5)
  • 17. Results GOALS 1. To improve examination quality through a faculty development program, followed by a longitudinal item review occurring before examination administration 2. To improve faculty members’ self-rated confidence and success 3. To measure changes in their opinions regarding item-writing guidelines and review.
  • 18. Goal 1: What happened to item-quality?
  • 19. Results • No significant difference between the control and intervention items, respectively • Mean student scores (% ± SD) did change (p<0.001): • Control sequence, 88.3 ± 4.5 • Intervention sequence 85.6 ± 6.0
  • 20. Distribution by level CONTROL ITEMS INTERVENTION ITEMS
  • 21. Distribution by level ITEM CLASS WITH REVIEW, n (%) WITHOUT REVIEW, n (%) Level 1 31 (11.4) 52 (19.7) Level 2 70 (27.5) 76 (28.8) Level 3 7 (2.6) 3 (1.1) Level 4 142 (52.2) 122 (46.2) Uncategorizable 22 (8.1) 11 (4.2)
  • 22. Goal 2: How did participants’ self-rated confidence and success change?
  • 23. Survey opinions: self-focused ITEM Pre (mean) Post (mean) p-value How would you rate your confidence at writing effective multiple-choice test items? (0 – very unconfident, 10 – very confident) 6.0 8.1 0.002 How would you rate your success at writing multiple- choice test items? (0 – very unsuccessful, 10 – very successful) 6.4 7.9 <0.001 To what degree do you feel confident that you can properly evaluate your and your colleagues' test questions? (0 – very unconfident, 10 – very confident) 6.7 8.4 0.005 To what degree do you feel confident that you could implement a formal exam item evaluation process as a coordinator of another course? (0 – very unconfident, 10 – very confident) 5.5 7.1 0.008
  • 24. Goal 3: How did participants’ opinions of item guidelines and review change?
  • 25. Survey opinions: item-focused ITEM Pre (mean) Post (mean) p-value In your opinion, to what degree will peer-review of exam items affect item quality? (0 – very negatively, 10 – very positively) 7.9 8.5 0.14 Do you plan to modify future multiple-choice items based on item-writing guidelines? (1 – Yes, 2 – No) 9 Y 1 N 10 Y 0 N 1.00 In your opinion, to what degree will voluntary application of item-writing guidelines affect item quality? (0 – very negatively, 10 – very positively) 7.9 8.4 0.24
  • 27. Top 5 item flaws GUIDELINES n (% of total changes) Include the central idea in the stem instead of the choices. 37 (33.6) Use correct grammar, punctuation, capitalization, and spelling. 17 (15.5) Minimize the amount of reading in each stem. 13 (11.8) Use the question, completion, and best answer versions of the conventional multiple choice (MC), the alternate choice, true-false, multiple true-false, matching, and the context-dependent item and item set formats, but avoid the complex MC (Type K) format. 10 (9.1) Keep choices independent; choices should not be overlapping. 7 (6.4)
  • 28. Comparison of undesirable format frequency
  • 29. Item Review Process • Experience in other courses

Editor's Notes

  1. We decided to start an exam item review process because: we had a wide variability in what we collectively thought was a “good” or a “bad” question. We had a relatively young faculty members who like myself had no clue how to write questions if we’re being honest. Lastly we wanted to limit any grade variability that may have been due to poorly written questions. Am I describing that this was only clinical faculty here too?
  2. So how do you make the first meeting and all the meeting less like this and more like this? First we got ALL faculty members involved in the course into the room for meeting 1. We had faculty members with experience in item writing present a mini- “faculty development” presentation of haladyna’s item writing guidelines.
  3. Developed from an analysis of 27 textbooks and 27 research studies Purpose was to validate each guideline based on agreement in studied sources Haladyna and Downing examined 46 measurement textbook passages dealing with how to write multiple choice items. They produced a set of 43 item-writing guidelines. They found that some guidelines had a strong consensus from these testing specialists. Some guidelines were given lesser attention. Several guidelines were controversial. Coverage of these guidelines in these books varied from very comprehensive to very limited. Commonly, authors did not logically or empirically justify the guidelines they presented.
  4. Point biserial correlates student scores on one particular question with their scores on the test as a whole. The driving assumption is, students who score well on the test as a whole should on average score well on the question under review. And vice versa. If a question deviates from this assumption the rpb lets us know. Rpb ranges from -1.0 to +1.0 The closer the rpb to 1.0 means the more reliable the question is considered because it “discriminates” well between students who mastered the material and those who did not. P-value simple measure of question difficulty ranges from 0 to 1 with lower numbers meaning more difficult the question. For example if p value is 0 then no students got the question correct or if p value is 1 then everyone got it correct.
  5. Let's look at the handout. The majority of the guidelines are just common sense and are universally endorsed (when mentioned). I don't think we need to spend any time discussing these since the consensus is already there, but do read over them all at some point just to bring them into active consideration. What I think we should spend time discussing are the five that received mixed recommendations, both from textbooks and empirical research.
  6. So now that you know why we felt a peer review process was important, I am going to discuss the implementation and logistics of the process. We have two, sequential Self Care modules that occur over one year. 11 faculty teach in SC I, and 9 faculty teach in SC II. Of 11 faculty in SC I, 6 (Fall 2012) and 5 (Fall 2013) also teach in SC II Important that all faculty members are willing to participate in the process.
  7. Deadlines for exam items are posted in the original schedule at the beginning of each semester. We usually only require them one week in advance, but due to this process, we asked that they be submitted at least 2 weeks in advance. Once all TQs were submitted, the course coordinator created the assessment in ES and downloaded it as a PDF. This draft was then sent to all course instructors as quickly as possible. We tried to give them several days to prepare for the meeting. In preparation for the meeting, we asked that all faculty review each item and be ready to present suggested revisions at the meeting. They should keep in mind all guidelines for exam writing that were covered at the beginning of the course and use these as a basis for recommendations.
  8. The faculty involved in the course were distributed over all three of our campuses. The meeting were face-to-face with distance connection to the off-site campuses. We fostered an open, friendly environment for people to feel comfortable making their suggestions. Each meeting included great discussion about several items ending in agreement of how to make the items better. When the exam was sent out to the faculty, the author of each item was not noted. However, we all have access to the syllabus and know what each other teach. During the meetings, the authors of items were often asked for clarification and/or rationale behind certain questions. Once everyone agreed upon suggested revisions, the course coordinators revised the exam and resent it to the group to double check that all changes accurately reflected the group’s decision. Once, this process was complete, the exam was posted for students to download.
  9. All faculty volunteered to participate and bought into the process Meetings were well attended Faculty were prepared for the meeting All faculty were members of the same department Faculty adhered to the agreed upon “do not use” formats from the orientation No one took revisions personally
  10. Resistant faculty Courses that involve more than one department/discipline Communication Buy-in
  11. The pre- and post-survey assessed: Faculty confidence and success at writing exam items Past experience with test question writing guidelines and peer review processes How they think this process will affect item quality Confidence in incorporating the process into other courses
  12. Ten of 12 instructors completed both the pre- and post-surveys. Survey questions and responses are summarized in Table 1. Six participating faculty reported that they have been teaching in a professional pharmacy curriculum for ≤5 years and four for 6-10 years. Seventy percent reported previous training in item writing with faculty development programs (n=8) and credentialing board training (n=3) being the most commonly reported experiences. When asked, “which of the following factors affect your sense of success in item writing,” faculty responded as follows: item statistics (n=9), previous training in item writing (n=4), and student challenges to exam items (n=3). At baseline, only 5 faculty members (50%) had ever participated in item peer-review, four reporting using peer-review “half of the time” and one “a minority of the time”. Similarly, only five faculty members had ever modified exam items based on item-writing guidelines at baseline.
  13. Ten of 12 instructors completed both the pre- and post-surveys. Survey questions and responses are summarized in Table 1. Six participating faculty reported that they have been teaching in a professional pharmacy curriculum for ≤5 years and four for 6-10 years. Seventy percent reported previous training in item writing with faculty development programs (n=8) and credentialing board training (n=3) being the most commonly reported experiences. When asked, “which of the following factors affect your sense of success in item writing,” faculty responded as follows: item statistics (n=9), previous training in item writing (n=4), and student challenges to exam items (n=3). At baseline, only 5 faculty members (50%) had ever participated in item peer-review, four reporting using peer-review “half of the time” and one “a minority of the time”. Similarly, only five faculty members had ever modified exam items based on item-writing guidelines at baseline.
  14. Ten of 12 instructors completed both the pre- and post-surveys. Survey questions and responses are summarized in Table 1. Six participating faculty reported that they have been teaching in a professional pharmacy curriculum for ≤5 years and four for 6-10 years. Seventy percent reported previous training in item writing with faculty development programs (n=8) and credentialing board training (n=3) being the most commonly reported experiences. When asked, “which of the following factors affect your sense of success in item writing,” faculty responded as follows: item statistics (n=9), previous training in item writing (n=4), and student challenges to exam items (n=3). At baseline, only 5 faculty members (50%) had ever participated in item peer-review, four reporting using peer-review “half of the time” and one “a minority of the time”. Similarly, only five faculty members had ever modified exam items based on item-writing guidelines at baseline.
  15. Word the stem positively, avoid negatives such as NOT or EXCEPT. If negative words are used, 
use the word cautiously and always ensure that the word appears capitalized and boldface. 7 (6.4) Avoid window dressing. 6 (5.5) Avoid all-of-the-above. 3 (2.7) Place choices in logical or numerical order. 2 (1.8) Develop as many effective choices as you can, but research suggests three is adequate. 2 (1.8) Ensure that the directions in the stem are very clear. 2 (1.8) Avoid giving clues to the right answer, such as grammatical inconsistencies that cue the test-taker to the correct choice. 1 (0.9)