Assessment in Online Courses Online Courseware I see

  • Slides: 11
Download presentation
Assessment in Online Courses

Assessment in Online Courses

Online Courseware I see two fundamental aspects: • Content • Assessment (Exercises and activities)

Online Courseware I see two fundamental aspects: • Content • Assessment (Exercises and activities)

Assessment Example: Evaluating Student Essays • Manual: Calibrated Peer Review (CPR): UCLA (Coursera) •

Assessment Example: Evaluating Student Essays • Manual: Calibrated Peer Review (CPR): UCLA (Coursera) • Automated: Automated Essay Scoring (AES): Ed. X

Automated Grading: AES • Long history • There are formal competitions for automated essay

Automated Grading: AES • Long history • There are formal competitions for automated essay scoring systems! • Machine learning algorithms • Indistinguishable in quality from human graders • Ed. X: Grade 100 sample essays • Breaks down for longer assignments, things where style counts (poetry, humor), individual topics

Manual Grading: CPR • Requires a strong rubric (multiple choice) • Students must be

Manual Grading: CPR • Requires a strong rubric (multiple choice) • Students must be “trained”: • Score 3+ essays, then get a competency rating • Competency rating affects weighting on reviews of others • Your reviewer rating depends on how close you are to weighted average • Gives students experience with reviewing • At end, students can self-rate own essays, and be graded on that exercise

CPR Issues • Some fraction of papers get all low-quality evaluators • Only works

CPR Issues • Some fraction of papers get all low-quality evaluators • Only works for an assignment for which a rubric can be generated (not so individual) • Length limits on what students can bear since they evaluate a number of essays • “Synchronous” vs. “Asynchronous” courses • Is there a peer group?

Comparison • Type: CPR has a little more flexibility, greater length restrictions • Consistency

Comparison • Type: CPR has a little more flexibility, greater length restrictions • Consistency of scoring: AES more consistent • Feedback: More consistent with AES, but CPR allows potentially better feedback from (good) human evaluators • Instructor workload: AES needs a training pool (100 essays), CPR needs rubric, intervention for some fraction of poorly evaluated essays (not scaleable? ) • Student learning: AES has rapid feedback, is good for catching mechanical writing problem. CPR teaches evaluation skills (for a time cost)

Assessment: Automated • Data interchange for those and custom exercises (all we really need

Assessment: Automated • Data interchange for those and custom exercises (all we really need is scoring and perhaps interaction data) • Formats and systems for “standard” question types (MCQ, etc) • How rich can we make the activities?

Automated Assessment: Learning to Program • We will study this because it is a

Automated Assessment: Learning to Program • We will study this because it is a fairly well developed area (many systems), and it provides and interesting example of a “non-trivial” interactive exercise type. • Categories: • Programming Exercises: Write actual code • Proficiency exercises of some sort

Problets • Not programming exercises, but has other interesting interactive exercises. • Implementation: Java

Problets • Not programming exercises, but has other interesting interactive exercises. • Implementation: Java Applets • http: //www. problets. org/

Open. DSA Proficiency Exercises • Algorithm “simulation” • Manipulate a view of a data

Open. DSA Proficiency Exercises • Algorithm “simulation” • Manipulate a view of a data structure to reproduce the behavior of an algorithm. • Model answer generated by instumented implementation of algorithm: Store a series of states • Each (appropriate) manipulation by user also generates a state • Simply verify that the states match (by some definition of match) • Various behaviors possible when states do not match. • http: //algoviz. org/Open. DSA/Books/CS 3114/html/BST. html