The reliability (𝛒tt) is the accuracy of a test

A test is reliable if the scores of a person do not differ substantially when tested repeatedly

If this holds true for all/most participants, we expect a high reliability of the test scores

Implication: We estimate reliability based on the variance of the observed test scores

Because the test score variance consists of a systematic (i.e., due to systematic differences among participants; differences in their ability/personality) and an unsystematic part (i.e., variance due to a random measurement error)

Reliability (illustration)

Reliability

(4+3)

A test's true score variance is not known, however, and

reliability must be estimated rather than calculated directly

There are several ways to estimate a test's reliability based on the implications of CTT

In general, reliability can be measured by the correlation between two strictly parallel tests

– Strictly parallel: Two tests with

(a) identical true scores

(b) same means

(c) same variances, and

(d) same error variances

Each method involves assessing the consistency of an examinee's scores over time (retest reliability), across different content samples (parallel test reliability), or across different items (internal consistency)

Reliabiltiy

Definiton (with formula)

Reliability Methods (3)

The interrelatedness of items measuring the same trait is computed (internal consistency)

used to assess the consistency of item responses across items within a test

Cronbach‘s alpha, McDonald’s Omega and others

The correlation of two (parallel) tests which measure the same latent variable (e.g., intelligence) is computed

A test is administered twice with a certain interval and the correlation is computed (retest reliability)

Reliability: Consistency

Situation: a single administration of one test form

Procedure: Divide test into comparable halves and correlate sum scores from both halves

Split Half with Spearman Brown adjustment

Kuder Richardson #20 and #21

Cronbach’sAlpha

Meaning: consistency of item responses across the parts of a measuring instrument (“parts” = individual items or subgroups of items)

Correct interpretation: Amount of systematic variance in test scores. NOT: Degree of homogeneity, whereby homogeneity = unidimensionality

A measure is said to be unidimensional if a single latent trait accounts for the correlations among a set of items

Retest Reliability

Situation: Same people taking two administrations of the same test

Procedure: Correlate scores on the two tests which yields the coefficient of stability

Meaning: the extent to which scores on a test can be generalized over different occasions (temporal stability).

Appropriate use: Information about the stability of the trait over time

– Stable rank-ordering of test takers

Parallel Test Reliability

Situation: Testing of same people on different but comparable/parallel forms of the test

Procedure: correlate the scores from the two tests which yields a coefficient of equivalence

Meaning: the consistency of responses to different item samples (where testing is immediate) and across occasions (where testing is delayed)

Providing information about the equivalence of forms

Last changed11 days ago