Evaluating Student Motivation in Field Testing Concepts Implementation

  • Slides: 57
Download presentation
Evaluating Student Motivation in Field Testing: Concepts, Implementation, Analysis, and Interpretation

Evaluating Student Motivation in Field Testing: Concepts, Implementation, Analysis, and Interpretation

Background Information • • • SB 288 Legislation Impact on High School Testing Eligibility

Background Information • • • SB 288 Legislation Impact on High School Testing Eligibility Rule for End of Course Examinations State Board of Education and NDE Leadership Guidelines Nevada Test Advisory Committee Recommendations Collaborative Approach with Nevada Department of Education, West. Ed and Measured Progress

EVALUATING STUDENT MOTIVATION IN FIELD TESTING: CONCEPTS, IMPLEMENTATION, ANALYSIS, AND INTERPRETATION Chad W. Buckendahl

EVALUATING STUDENT MOTIVATION IN FIELD TESTING: CONCEPTS, IMPLEMENTATION, ANALYSIS, AND INTERPRETATION Chad W. Buckendahl

Overview Impact of student motivation on assessment development and validation Strategies for evaluating motivation

Overview Impact of student motivation on assessment development and validation Strategies for evaluating motivation – Embedded, unscored items – Stand-alone field tests Cross-disciplinary research opportunity 1/3/2022 4

Defining malingering Medicine – fabricating or exaggerating mental or physical disorders for secondary gain

Defining malingering Medicine – fabricating or exaggerating mental or physical disorders for secondary gain motives (e. g. , financial, avoiding work, obtaining drugs) Education – demonstrating suboptimal effort or performance for secondary gain (e. g. , financial, attention) Distinct from apathy 1/3/2022 5

Clinical instruments Test of Memory Malingering (TOMM; Tombaugh, 1996) – 50 -item test that

Clinical instruments Test of Memory Malingering (TOMM; Tombaugh, 1996) – 50 -item test that distinguishes between malingering and memory impairment Victoria Symptom Validity Test (VSVT; Slick, Hopp, Strauss, & Thompson, 1997) – 48 item test to evaluate possible exaggeration or feigning cognitive impairments 1/3/2022 6

Generalizing to EOCs New high school assessments Anticipated low student motivation Consequences for some,

Generalizing to EOCs New high school assessments Anticipated low student motivation Consequences for some, but not all students Additional strategy for evaluating credibility of field test data 1/3/2022 7

Methodology Embedding easier, unscored items within the content domain – Prioritization of assessment real

Methodology Embedding easier, unscored items within the content domain – Prioritization of assessment real estate Evaluating data with and without students identified as “unmotivated” Using results to inform standard setting, form construction, future field tests Footer Text 1/3/2022 8

Summary Commend NV Department of Education for supporting applied research Implications for practice –

Summary Commend NV Department of Education for supporting applied research Implications for practice – Adjunct approach for empirically evaluating motivation – Stability of item level data – Informing development and validation activities 1/3/2022 9

Evaluating Student Motivation in Field Testing: Design and Development of the Field Test Joanne

Evaluating Student Motivation in Field Testing: Design and Development of the Field Test Joanne L. Jensen, West. Ed Paper presented at the National Conference on Student Assessment June 23, 2015 San Diego

Assessment Program in Transition —Similarities and Differences Similarities… • Active collaboration with NDE staff

Assessment Program in Transition —Similarities and Differences Similarities… • Active collaboration with NDE staff • Active collaboration with Nevada educators • Stakes for students

Assessment Program in Transition— Similarities and Differences… • Proficiency examinations were based on a

Assessment Program in Transition— Similarities and Differences… • Proficiency examinations were based on a select number of content standards reflecting a breadth but not a depth of content • Writing prompt versus writing task based on a reading passage • Calculator use versus no calculator • Reporting structured by claims rather than content domains • New item types (multiple select, two-part items, gridded response) • Rigor of the assessment

Math I

Math I

Math I

Math I

Math I

Math I

Math I

Math I

Math I

Math I

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

Math II

ELA I

ELA I

ELA I

ELA I

ELA I

ELA I

ELA II

ELA II

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math I

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items Math II

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Malingering items ELA I

Feedback from the field… • Students report that English Language Arts did not seem

Feedback from the field… • Students report that English Language Arts did not seem to be more difficult than the previous HSPEs. • The mathematics assessments were too long (60 items) and too hard. • Need more practice items.

Jennifer L. Dunn Malingering Items: Some preliminary ideas around identification, evaluation and use. 45

Jennifer L. Dunn Malingering Items: Some preliminary ideas around identification, evaluation and use. 45 Measured Progress © 2015

How should malingering items be used? The context: § In stand alone field tests,

How should malingering items be used? The context: § In stand alone field tests, students often have very little incentive to make a significant effort in responding to items § The potential lack of valid responses may compromise evaluations of the items The initial idea: § Identify students with malingering behaviors § Remove these students from the data used to evaluate the items 46 Measured Progress © 2015

Goal Use malingering items to identify malingering students and evaluate the impact of removing

Goal Use malingering items to identify malingering students and evaluate the impact of removing these students on item evaluations. § Step 1: Identify malingering students § Step 2: Establish an understanding of malingering behaviors § Step 3: Compare the influence of these students on the item evaluations 47 Measured Progress © 2015

Identifying malingering students Identify students that are performing differently on the malingering items than

Identifying malingering students Identify students that are performing differently on the malingering items than expected Assumption: § Students should be getting the malingering items correct. § Students who get malingering items wrong are demonstrating malingering behavior Key Difference: § Other cognitive assessments: test takers are trying to be different. § Educational Assessment: test takers are not trying. 48 Measured Progress © 2015

Identifying malingering students Option 1: § Calculate total scores for all students using only

Identifying malingering students Option 1: § Calculate total scores for all students using only the malingering items § Examine the malingering total score distribution § If bimodal – the lower mode represents malingering students § If not bimodal – determine a reasonable cut point and carefully evaluate the response patterns of those students 49 Measured Progress © 2015

Identifying Malingering Students: Option 1 250 200 Malingering Students 150 100 50 0 1

Identifying Malingering Students: Option 1 250 200 Malingering Students 150 100 50 0 1 2 3 4 5 6 50 7 8 9 10 11 Measured Progress © 2015

Identifying malingering students Option 2: § Identify low scoring students (using all items on

Identifying malingering students Option 2: § Identify low scoring students (using all items on the test) § Any student who scores at or below the total score that corresponds to random guessing § Calculate a total score using the malingering items § Remove any student who got a high score on the malingering items from the population of low scoring students § Remaining students are the malingerers 51 Measured Progress © 2015

Verifying the identification of malingering students Use multiple pieces of evidence § Larger than

Verifying the identification of malingering students Use multiple pieces of evidence § Larger than typical declines in performance § Inconsistent performance across content areas Compare the total score distribution of a randomly selected group of items to the total score distribution of the malingering items § The distribution of the randomly selected items should not be bimodal 52 Measured Progress © 2015

Understanding malingering students What are the characteristics of malingering students? Are malingering students simply

Understanding malingering students What are the characteristics of malingering students? Are malingering students simply random guessers? § Identify students who score below the chance level on the assessment. Compare the selected students and response patterns Do malingering behaviors change as the test progresses? § Do response the response patterns of the malingering students vary by test position? 53 Measured Progress © 2015

Understanding malingering items Evaluate the malingering items § Do malingering items differ from the

Understanding malingering items Evaluate the malingering items § Do malingering items differ from the other items on the test? § Are they easier? § Are they statistically identifiable? § Does test position impact the effectiveness of the malingering items? § Are the item characteristics of the malingering items consistent across test positions? How many malingering items are necessary? 54 Measured Progress © 2015

Assessing the impact on item evaluation Evaluate the items statistics with and without the

Assessing the impact on item evaluation Evaluate the items statistics with and without the malingering students § Item Statistics § P-value, Item total correlations, IRT § Distractor statistics § Model fit § Test Statistics § Reliability, Test Information How many malingering students does it take to make a difference? 55 Measured Progress © 2015

Additional Questions Are there other uses for malingering items? § Generating student scores §

Additional Questions Are there other uses for malingering items? § Generating student scores § Linking § Standard Setting How can we verify that the malingering items are functioning as intended? Is including these items worthwhile? 56 Measured Progress © 2015

Thank You! Questions? Comments? Suggestions? 57 Measured Progress © 2015

Thank You! Questions? Comments? Suggestions? 57 Measured Progress © 2015