Measurement: - PowerPoint PPT Presentation

1 / 18
About This Presentation
Title:

Measurement:

Description:

... of the same tests for same examinees. ... Are assessment results consistent over time (over occasions) ... For example, consider a math problem-solving test. ... – PowerPoint PPT presentation

Number of Views:45
Avg rating:3.0/5.0
Slides: 19
Provided by: george123
Category:

less

Transcript and Presenter's Notes

Title: Measurement:


1
Measurement
  • Reliability

2
Classroom Assessment Reliability
  • Reliability Assessment Consistency.
  • Consistency within tests across examinees.
  • Consistency within tests over multiple
    administrations to the same examinees.
  • Consistency across alternative forms of the same
    tests for same examinees.

3
Three Types of Reliability
  • Stability reliability.
  • Alternate form reliability.
  • Internal consistency reliability.

4
Stability Reliability
  • Stability Reliability
  • Concerned with the question
  • Are assessment results consistent over time (over
    occasions).
  • Think of some examples where stability
    reliability might be important.
  • Why might test results NOT be consistent over
    time?

5
Evaluating Stability Reliability
  • Test-Retest Reliability.
  • Compute the correlation between a first and later
    administration of same test.
  • Classification-consistency.
  • Compute the percentage of consistent student
    classifications over time.
  • Main concern is with the stability of the
    assessment over time.

6
Example of Classification Consistency
7
Example of Classification Consistency (Good
Reliability)
8
Example of Classification Consistency (Poor
Reliability)
9
Alternate-form Reliability
  • Are two, supposedly equivalent, forms of an
    assessment in fact actually equivalent?
  • The two forms do not have to yield identical
    scores.
  • The correlation between two or more forms of the
    assessment should be reasonably substantial.

10
Evaluating Alternate-form Reliability
  • Administer two forms of the assessment to the
    same individuals and correlate the results.
  • Determine the extent to which the same students
    are classified the same way by the two forms.
  • Alternate-form reliability is established by
    evidence, not by proclamation.

11
Example of Using a Classification Table to Assess
Alternate-Form Reliability
12
Example of Using a Classification Table to Assess
Alternate-Form Reliability
13
Internal Consistency Reliability
  • Concerned with the extent to which the items (or
    components) of an assessment function
    consistently.
  • To what extent do the items in an assessment
    measure a single attribute?
  • For example, consider a math problem-solving
    test. To what extent does reading comprehension
    play a role? What is being measured?

14
Evaluating Internal Consistency Reliability
  • Split-Half Correlations.
  • Kuder-Richardson Formua (KR20).
  • Used with binary-scored (dichotomous) items.
  • Average of all possible split-half correlations.
  • Cronbachs Coefficient Alpha.
  • Similar to KR20, except used with non-binary
    scored (polytomous) items (e.g., items that
    measure attitude.

15
ReliabilityComponents of an Observation
  • O T E
  • Observation True Status Measurement Error.

16
Standard Error of Measurement
  • Provides an index of the reliability of an
    individuals score.
  • The standard deviation of the theoretical
    distribution of errors (i.e. the Es).
  • The more reliable a test, the smaller the SEM.
  • The SEM is smallest near the average score on a
    test.

17
Things to Do toImprove Reliability
  • Use more items or tasks.
  • Use items or tasks that differentiate among
    students.
  • Use items or tasks that measure within a single
    content domain.
  • Keep scoring objective.
  • Eliminate (or reduce) extraneous influences
  • Use shorter assessments more frequently.

18
  • End
Write a Comment
User Comments (0)
About PowerShow.com