1 / 19

Classroom Assessment

Classroom Assessment. Reliability. Classroom Assessment Reliability. Reliability = Assessment Consistency. Consistency within teachers across students. Consistency within teachers over multiple occasions for students. Consistency across teachers for the same students.

hali
Télécharger la présentation

Classroom Assessment

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Classroom Assessment Reliability

  2. Classroom Assessment Reliability • Reliability = Assessment Consistency. • Consistency within teachers across students. • Consistency within teachers over multiple occasions for students. • Consistency across teachers for the same students. • Consistency across teachers across students.

  3. Three Types of Reliability • Stability reliability. • Alternate form reliability. • Internal consistency reliability.

  4. Stability Reliability • Stability Reliability • Concerned with the question: Are assessment results consistent over time (over occasions). Think of some examples where stability reliability might be important. Why might test results NOT be consistent over time?

  5. Evaluating Stability Reliability • Test-Retest Reliability. Compute the correlation between a first and later administration of the same test. • Classification-consistency. Compute the percentage of consistent student classifications over time. (Example on next slide). • Main concern is with the stability of the assessment over time.

  6. Example of Classification Consistency

  7. Example of Classification Consistency (Good Reliability)

  8. Example of Classification Consistency (Poor Reliability)

  9. Alternate-form Reliability • Are two, supposedly equivalent, forms of an assessment in fact actually equivalent? • The two forms do not have to yield identical scores. • The correlation between two or more forms of the assessment should be reasonably substantial.

  10. Evaluating Alternate-form Reliability Administer two forms of the assessment to the same individuals and correlate the results. Determine the extent to which the same students are classified the same way by the two forms. Alternate-form reliability is established by evidence, not by proclamation.

  11. Example of Using a Classification Table to Assess Alternate-Form Reliability

  12. Example of Using a Classification Table to Assess Alternate-Form Reliability

  13. Internal Consistency Reliability Concerned with the extent to which the items (or components) of an assessment function consistently. To what extent do the items in an assessment measure a single attribute? For example, consider a math problem-solving test. To what extent does reading comprehension play a role? What is being measured?

  14. Evaluating Internal Consistency Reliability • Split-Half Correlations. • Kuder-Richardson Formua (KR20). • Used with binary-scored (dichotomous) items. • Average of all possible split-half correlations. • Cronbach’s Coefficient Alpha. • Similar to KR20, except used with non-binary scored (polytomous) items (e.g., items that measure attitude.

  15. ReliabilityComponents of an Observation O = T + E Observation = True Status + Error.

  16. Standard Error of Measurement • Provides an index of the reliability of an individual’s score. • The standard deviation of the theoretical distribution of errors (i.e. the E’s). • The more reliable a test, the smaller the SEM.

  17. Individual characteristics Anxiety Motivation Health Fatigue Understanding (of task) “Bad hair day” External characteristics Directions Environmental disturbances Scoring errors Observer differences/biases Sampling of items Sources of Error in Measurement

  18. Things to Do toImprove Reliability • Use more items or tasks. • Use items or tasks that differentiate among students. • Use items or tasks that measure within a single content domain. • Keep scoring objective. • Eliminate (or reduce) extraneous influences • Use shorter assessments more frequently.

  19. End

More Related