• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/51

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

51 Cards in this Set

  • Front
  • Back

Refers to the consistency of scores obtained by the same persons when they are reexamined by the same test on different occassions.

Reliability

Variance from true differences.

True variance

Variance from irrelevant, random sources.

Error variance

Ratio of true score variance and the total variance.

Reliability coefficient

The most common way of computing a correlation.

Pearson product moment correlation coefficient

Expresses the degree of correspondence or a relationship between two sets of scores.

Correlation coefficient

Sources of error variance:

Test construction


Test administration


Test scoring and interpretation

Reliability estimates:

Test retest reliability


Alternate forms reliability estimate


Split half reliability estimate


Kuder richardson reliability and coefficient alpha


Interscorer reliability

Estimate of reliability obtained by correlating pairs of scores from the same people on two different administrations of the same test.

Test retest reliability

If interval of test retest administration is more than 6 months.

Coefficient of stability

Estimate of the extent to which item sampling and other errors have affected scores on versions of the same test.

Alternate forms reliability estimate

Independently constructed test designed to meet same specifications.

Alternate

The degree of relationship between various forms of a test.

Coefficient of equivalence

Is obtained by correlating two pairs of scores obtained from equivalent halves of a single test administered once.

Split half reliability estimate

Other term for split half reliability

Coefficient of internal consistency

Estimates the effect of lengthening and shortening of a test.

Spearman brown formula

Two sources of error variance that influence inter item consistency:

Content sampling


Heterogeneity of the behavior being sampled.

The most common formula for finding inter item consistency.

Kuder Richardson formula 20

Is the mean of all split half coefficients from different splitting of the test.

Kuder richardson reliability coefficient

Is applicable to tests whose items are scored as right or wrong, or according to some other all-or-none systems.

Kuder richardson formula

A generalized formula for tests that have multiple-scored items.

Coefficient alpha

Degree of agreement or consistency between two or more scorers with regards to a particular measure.

Interscorer reliability

Index of measurement for interscorer reliability

Coefficient of interscorer reliability

Errors due to examiner bias:

Error of central tendency


Leniency/generosity error


Severity error


Halo effect


Horn effect


Contrast error


Recency bias

Less than accurate rating or evaluation by a rater or judge due that rater's tendency to make ratings near the midpoint of the scale.

Error of central tendency

Rater's tendency to be too forgiving or insufficiently critical.

Leniency/generosity error

Rater's tendency to be overly critical.

Severity error

Tendency of the leader to judge all aspects of an individual using a general impression that was formed on only one or few of the individual's characteristics .

Halo effect

Refers to the tendency to let one poor rating influence all other ratings, resulting in a lower overall evaluation than deserved.

Horn effect

Happens when raters compares examinees with one another instead of against performance standards.

Contrast error

Occurs when leader assigns ratings based only on the employee's most recent performance.

Recency bias

This measure is suited to the interpretation of individual scores. Is independent of the variability of the group on which it is computed.

Standard error of measurement

Is an estimate of how well a test measures what it purports to measure.

Validity

3 categories of validity:

Content validity


Criterion related validity


Construct validity

Two major trends in validity:

Strengthened theoretical orientation


Close linkage between psychological theory and verification through empirical and experimental hypothesis testing

Evaluation of subjects, topics, or contenys in a test.

Content validity

Evaluation of the relationship of scores to scores on other tests or instruments.

Criterion-related validity

Comprehensive analysis of theoretical framework+scores on other tests.

Construct validity

Describes a judgement of how adequately a test samples behavior representative of the universe of behavior that the test is designed to sample.

Content validity

Judgment of how a test score can be used to infer an individual's most probable standing based on some measure of interest.

Criterion-related validity

May be broadly defined as the standard against whichba test or a test score is evaluated.

Criterion

Criterion has to be:

Relevant


Valid


Uncontaminated

Error where ratings become influenced based on rater's knowledge of test scores.

Criterion contamination

2 types of criterion-related validity:

Concurrent validity


Predictive validity

The extent scores on a new measure relate to scores from a criterion measure administered at the same time.

Concurrent validity

Uses the scores from the new measure to predict performance on a criterion measure administered at a later time.

Predictive validity

A correlation coefficient that provides a measure of the relationship between test scores and scores on the criterion measure.

Validity coefficient

Judgment about the appropriateness of inferences drawn from test scores regarding individual standings on a variable called construct.

Construct validity

He pointed out that in order to demonstrate construct validity, a test must correlate highly with variables it should theoretically correlate, and vice versa.

DT Campbell

High relationship with measures construct is supposed to be related to.

Convergent evidence

Low relationship with measures construct is not supposed to be related to.

Discriminate evidence