0% found this document useful (0 votes)
72 views29 pages

Chapter 3 Part 1 Validity Reliability 1

The document discusses various characteristics of effective assessment methods, including validity, reliability, practicability, comprehensiveness, and relevance. It defines validity as the extent to which an assessment measures what it intends to measure. There are four types of validity: face validity, content validity, construct validity, and criterion validity. Reliability refers to the consistency and stability of measurement and includes four types: inter-rater reliability, test-retest reliability, parallel-forms reliability, and internal consistency reliability. Effective assessments should not only be valid and reliable but also practical, comprehensive, and relevant to the learning outcomes being assessed.

Uploaded by

anjeesahiri91589
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
72 views29 pages

Chapter 3 Part 1 Validity Reliability 1

The document discusses various characteristics of effective assessment methods, including validity, reliability, practicability, comprehensiveness, and relevance. It defines validity as the extent to which an assessment measures what it intends to measure. There are four types of validity: face validity, content validity, construct validity, and criterion validity. Reliability refers to the consistency and stability of measurement and includes four types: inter-rater reliability, test-retest reliability, parallel-forms reliability, and internal consistency reliability. Effective assessments should not only be valid and reliable but also practical, comprehensive, and relevant to the learning outcomes being assessed.

Uploaded by

anjeesahiri91589
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 29

by: LLOYD PSYCHE T.

BALTAZAR
Characteristics of Assessment Methods
1 VALIDITY is the extent to which an instrument
measures what it intends to measure.

2 RELIABILITY is the degree to which a test is


consistent and stable in measuring what it is
intended to measure.

3 PRACTICABILITY considers that the test must have


practical value from time, economy, scorability, and
administration point of view.
Characteristics of Assessment Methods

4 COMPREHENSIVE considers that the test must


cover the lesson taught in which it should assess
knowledge, skills, and values as adequately as
possible.

5 RELEVANT means that the test measures the


desired learning outcomes.
Validity

Validity - is the extent to which a measurement tool


measures what it's supposed to measure.

Remember your thermometer?

It's measuring the room temperature, not


your body temperature. Since it's supposed
to be measuring your body temperature, the
thermometer is not valid.
Validity
VALIDITY is an indication of how sound your
assessment is.

▪ Validity also refers to the usefulness of the


instrument for a given purpose.

▪ A test is valid when it is aligned to the learning


outcome.
4 Types of Validity
Face Validity
Content Validity
Construct Validity
Criterion Validity
4 Types of Validity
Face validity is the degree to which a test appears
to be related to a specific construct, in the judgment
of non-experts such as test takers.

▪ A test has face validity if its content simply looks relevant


to the person taking the test.

▪ It evaluates the appearance of the questionnaire in terms


of feasibility, readability, consistency of style and
formatting, and the clarity of the language used.
4 Types of Validity

Content validity is the degree to which


elements of an assessment instrument are
relevant to and representative of the targeted
construct for a particular assessment purpose
(Haynes, Richard, & Kubany, 1995).
4 Types of Validity
Content Validity
▪ involves evaluation of a test in order to ensure that it
includes all the items that are essential and eliminates
undesirable items

▪ the teacher writes out the test based on the TOS

▪ subject matter experts evaluate the extent to which our


test adequately captures the content domain
4 Types of Validity
Content Validity

Examples of measurements that are content valid:

➢If we want to test knowledge on Philippine History (construct),


then the test must cover everything from the beginning to the
present.

➢AP Physics knowledge (construct) measured by the AP exam


(measurement).
4 Types of Validity
Content Validity

Examples of measurements that have debatable content


validity:

➢The Bar Exam is not a good measure of ability to practice law.

➢IQ tests are not a good way to measure intelligence.


4 Types of Validity
Construct validity defined as the experimental
demonstration that a test is measuring the
construct it claims to be measuring.

A construct, or psychological construct as it is also called, is an


attribute, proficiency, ability, or skill that happens in the human
brain and is defined by established theories.

Examples: English proficiency, Math ability, Anxiety, Intelligence


4 Types of Validity
Construct Validity
▪ Construct validity is established when relationships between
our test and other variables confirm what is predicted by
theory.

▪ For example, theory might indicate that the personality traits


of conscientiousness and neuroticism should be negatively
related. If we develop a test of conscientiousness and then
demonstrate that scores on our test correlate negatively with
scores on a test of neuroticism.
4 Types of Validity
Construct Validity

▪ Furthermore, theory might indicate that conscientiousness


contains three specific dimensions. Statistical analysis of the
items within our test could show that the items tend to cluster,
or perform similarly, in three specific groups.
4 Types of Validity
Construct Validity

1. Convergent validity - refers to the degree to which two


measures of constructs that theoretically should be related,
are in fact related.

2. Discriminant validity - tests that constructs that should have


no relationship do, in fact, not have any relationship.
4 Types of Validity
Criterion validity is the extent to which a measure
is related to an outcome. It measures how well one
measure predicts an outcome for another measure.

▪ It is useful for predicting performance or behavior


in another situation (past, present, or future).
4 Types of Validity
Examples of Criterion validity:

➢A job applicant takes a performance test during the interview


process. If this test accurately predicts how well the employee
will perform on the job, the test is said to have criterion validity.

➢A graduate student takes the GRE (Graduate Record Exam). The


GRE has been shown as an effective tool (i.e. it has criterion
validity) for predicting how well a student will perform in graduate
studies.
4 Types of Validity
Criterion Validity

1. Predictive validity – occurs when a test accurately


predicts what it is supposed to predict.
4 Types of Validity
Criterion Validity
2. Concurrent validity - refers to the extent to which
the results of a particular test, or measurement,
correspond to those of a previously established
measurement for the same construct.
Reliability
Reliability - is the degree to which an assessment
tool produces stable and consistent results.

➢ It is a measure of the stability or consistency of test


scores. You can also think of it as the ability for a
test or research findings to be repeatable.

▪ For example, a medical thermometer is a reliable tool


that would measure the correct temperature each
time it is used.
4 Types of Reliability
Inter-rater Reliability

Test-retest Reliability

Parallel-Forms Reliability

Internal Consistency Reliability


4 Types of Reliability
Inter-rater reliability is a measure of reliability used to assess the degree
to which different judges or raters agree in their assessment decisions.

An Example of What
Type of Reliability When to Use How to Use You can Say When
You’re Done
When you want to know Examine the percent of The inter-rater reliability
whether there is agreement between for the best-dressed
consistency in the rating raters. Football player judging
Inter-rater Reliability of some outcome. was 0.9, which indicates
a high degree of
agreement among
judges.
4 Types of Reliability
Test-retest reliability is a measure of reliability obtained by administering
the same test twice over a period of time to a group of individuals. The
scores from Time 1 and Time 2 can then be correlated in order to
evaluate the test for stability over time.

An Example of What
Type of Reliability When to Use How to Use You can Say When
You’re Done
When you want to know Correlate the scores The Bonzo test of
whether a test is reliable from a test given in Time identity formation for
Test-retest Reliability over time. 1 with the same test adolescence is reliable
given in Time 2. over time.
4 Types of Reliability
Parallel-Forms reliability is a measure of reliability obtained by
administering different versions of an assessment tool (both versions
must contain items that probe the same construct, skill, knowledge base,
etc.) to the same group of individuals.
An Example of What
Type of Reliability When to Use How to Use You can Say When
You’re Done
When you want to know Correlate the scores Set A and Set B of the
if several different forms from one form of the test Math Exams are
of a test are reliable or with the scores from a equivalent to one
Parallel Forms equivalent. second, different form of another.
Reliability the same test of the
same content.
4 Types of Reliability
Internal Consistency reliability is a measure of reliability used to
evaluate the degree to which different test items that probe the same
construct produce similar results. Usually Cronbach Alpha (α) is used.

An Example of What
Type of Reliability When to Use How to Use You can Say When
You’re Done
When you want to know Correlate each individual All of the items on the
if the items on a test item score with the total emotional intelligence
assess one, and only score test assess the same
Internal Consistency one dimension. construct.
Reliability
4 Types of Reliability
Type of Reliability Description

Inter-rater Reliability Different people, same test

Test-retest Reliability Same people, different times

Different people, same time,


Parallel Forms Reliability
different/equivalent test

Internal Consistency Reliability Different questions, same construct


Other Good Qualities of Assessment
Instruments
1. Scorability
- ease of scoring
- test should have clear directions for scoring

2. Administrability
- clear provisions for directions and test rules for the
students to follow
Other Good Qualities of Assessment
Instruments
3. Objectivity
- agreement of two or more raters with regards to scoring
- test should not be influenced by personal bias

4. Fairness
- absence of discrimination in the test due to race, skin color,
gender, religion, etc.
Other Good Qualities of Assessment
Instruments

5. Adequacy
- An adequate test is a test with representativeness of test
items to the concept to be measured.

You might also like