Educ 6 M2-Midterm
Educ 6 M2-Midterm
Educ 6 M2-Midterm
INTRODUCTION
This module will introduce you to the principles and characteristics of a high-quality
assessment. The first part will discuss the productive and unproductive uses of tests. The second
part will focus on the classifications of tests and other types.
PRE-ASSESSMENT
Instruction: Complete the table below with the desired information.
What I Know about this What I want to know about What do I Learned from this
lesson? this lesson? lesson?
MODULE MAP
CONTENT
When designing good assessment, start by asking if the learning targets are on the right level of
difficulty to be able to motivate students and if there is an adequate balance among the different types of
learning targets.
A learning target is a clear description of what students know and able are to do. Learning targets are
categorized by Stiggins and Conklin (1992) into five.
a. Knowledge learning target is the ability of the student master a substantive subject matter.
b. Reasoning learning target is the ability to use knowledge and to solve problems.
c. Skill learning target is the ability to demonstrate achievement related skills like conducting
experiments, playing basketball, and operating computers.
d. Product learning target is the ability to create achievement related products such as written
reports, oral presentations, and art products.
e. Affective learning target is the attainment of affective traits such as attitudes, values, interests,
and self-efficacy.
Once the learning targets have been identified, match them with their corresponding methods by
considering the strengths of various methods in measuring different targets.
3. Validity
This refers to the extent to which the test serves its purpose or the efficiency with which it intends to
measure. Validity is a characteristic that pertains to the appropriateness of the inferences, uses, and results
of the test or any other method utilized to gather data.
There are factors that influence the validity of the test; namely, appropriateness of test items,
directions, reading vocabulary and sentence structures, pattern of answers, and arrangement of items.
Validity is always determined by professional judgment. However, there are different types of
evidence to use in determining validity. The following major sources of information can be used to
establish validity:
iv. Face validity is determined on the basis of the appearance of an assessment, whether
based on the superficial examination of the test, there seems to be a reasonable measure of the
objectives of a domain.
v. Instructional-related validity determines to what extent the domain of content in the test is
taught in class.
The following are suggestions for enhancing the validity of classroom assessments:
iv. Consider the reading vocabulary of the examinees. The test should not be made up of
jargons.
4. Reliability
This refers to the consistency with which a student may be expected to perform on a given test. It
means the extent to which a test is dependable, self-consistent, and stable.
There are factors that affect test reliability. These include the (1) scorer's inconsistency because of
his/her subjectivity, (2) limited sampling because of incidental inclusion and accidental exclusion of some
materials in the test, (3) changes in the individual examinee himself/herself and his/her instability during the
examination, and (4) testing environment.
There are various ways of establishing test reliability. These are the length of the test, difficulty of the
test, and objectivity of the scorer. There are also four methods in estimating the reliability of a good
measuring instrument.
i. Test-Retest Method or Test of Stability. The same measuring instrument is administered to the
same group of subjects. The scores of the first and second administrations of the test are determined
by correlation coefficient. The limitations of this method are: However, memory effects may operate
when the time interval is short. Likewise, factors such as unlearning and forgetting may occur when
the time interval is long resulting in low correlation of the test. Another limitation of the method is that
other varying environmental conditions may affect the correlation of the test regardless of the time
interval separating the two administrations.
ii. Parallel-Forms Method or Test of Equivalence. Parallel or equivalent forms of a test may be
administered to the group of subjects and the paired observations correlated. The two forms of the
test must be constructed in a manner that the content, type of item, difficulty, instructions for
administration, and several others, should be similar but not identical
iii. Split-Half Method. The test in this method may only be administered once, but the test items are
divided into two halves. The common procedure is to divide a test into odd and even items. The two
halves of the test must be similar but not identical in content, difficulty, means and standard
deviations.
iv. Assessment of Internal Consistency Method. This method is used with psychological tests,
which are constructed as dichotomously scored items. The testee either passes or fails in an item.
The method of obtaining reliability coefficients in this method is determined by the Kuder Richardson
formula.
Thus, an observed score can be higher or lower than the true score, depending on the nature of error.
The sources of error are reflected in Table 2.2
ii. Use independent raters or observers who can provide similar scores to the same performances.
iii. Make sure the assessment procedures and scoring are objective.
5. Fairness
This pertains to the intent that each question should be made as clear as possible to the examinees
and the test is absent of any biases. An example of a bias in an intelligence test is an item about a person or
object that has not been part of the cultural and educational context of the test taker. In mathematical tests
for instance, the reading difficulty level of an item can be a source of unfairness. Identified elements of
fairness are the student's knowledge of learning targets before instruction, the opportunity to learn, the
attainment of pre-requisite knowledge and skills, unbiased assessment tasks and procedures, and teachers
who avoid stereotypes.
6. Positive Consequences
These enhance the overall quality of assessment, particularly the effect of assessments on the
students' motivation and study habits.
Assessments need to take into consideration the teacher's familiarity with the method, the time
required, the complexity of administration, the ease of scoring and interpretation, and the cost to be able to
determine an assessment's practicality and efficiency. Administrability requires that a test must be
administered with ease. clarity, and uniformity. Directions must be specific so that students and teachers will
understand what they must do exactly. Scorability demands that a good test should be easy to score. Test
results should readily be available to both students and teachers for remedial and follow-up measures.
1. Learning Analysis. Tests are used to identify the reasons or causes why students do not leam and the
solutions to help them learn. Ideally, a test should be designed to determine what students do not know so
that the teachers can take appropriate actions.
2. Improvement of Curriculum. Poor performance in a test may indicate that the teacher is not explaining
the material effectively, the textbook is not clear, the students are not properly taught, and the students do not
see the meaningfulness of the materials. When only a few students have difficulties, the teacher can address
them separately and extend special help. If the entire class does poorly, the curriculum needs to be revised
or special units need to be developed for the class to continue.
3. Improvement of Teacher. In a reliable grading system, the class average is the grade the teacher has
earned.
4. Improvement of Instructional Materials. Tests measure how effective instructional materials are in
bringing about intended changes.
5. Individualization. Effective tests always indicate differences in students' learning. These can serve as
bases for individual help.
6. Selection. When enrollment opportunity or any other opportunity is limited, a test can be used to screen
those who are more qualified.
8. Guidance and Counseling. Results from appropriate tests, particularly standardized tests, can help
teachers and counselors guide students in assessing future academic and career possibilities.
9. Research. Tests can be feedback tools to find effective methods of teaching and learn more about
students, their interests, goals and achievements.
10. Selling and Interpreting the School to the Community. Effective tests help the community understand
what the students are learning, since test items are representative of the content of instruction. Tests can
also be used to diagnose general schoolwide weaknesses and strengths that require community or
government support.
11. Identification of Exceptional Children. Tests can reveal exceptional students inside the classroom.
More often than not, these students are overlooked and left unattended.
12. Evaluation of Learning Program. Ideally, tests should evaluate the effectiveness of each element in a
learning program, not just blanket the information of the total learning environment.
1. Grading. Tests should not be used as the only determinants in grading a student. Most tests do not
accurately reflect a student's performance or true abilities. Poor performance on a certain task may not only
indicate failure but lack or absence of the needed foundations as well.
2. Labeling. It is often a serious disservice to label a student, even if the label is positive. Negative labels
may lead the students to believe the label and act accordingly. Positive labels, on the other hand, may lead
the students to underachieve or avoid standing out as different or become overconfident and not exert effort
anymore.
4. Unannounced Testing. Surprise tests are generally not recommended. More often than not, they are the
scapegoats of teachers who are unprepared, upset by an unruly class or reprimanded by superiors. Studies
reveal that students perform at a slightly higher level when tests are announced; unannounced tests create
anxiety on the part of the students, particularly those who are already fearful of tests; unannounced tests do
not give students the adequate time to prepare, and surprise tests do not promote efficient learning or higher
achievement.
5. Ridiculing. This means using tests to deride students. Tracking Students are grouped according to
deficiencies as revealed by tests without continuous reevaluation, thus locking them into categories.
Classifications of Tests
Throughout the years, psychologists and educators have cooperatively produced new and better tests
and scales that measure the students' performance with greater accuracy. These tests may be classified
according
1. Administration
a. Individual - given orally and requires the examinees' constant attention since the manner of
answering may be as important as the score. An example of this is the Wechsler Adult Intelligence
Scale, one of the three individually administered intelligence scales. Another is a PowerPoint
presentation used as a performance test in a speech class.
b. Group - for measuring cognitive skills to measure achievement. Most tests in schools are
considered group tests where different test takers can take the tests as a group.
2. Scoring
a. Objective - independent scorers agree on the number of points the answer should receive, e.g.,
multiple choice and true or false.
b. Subjective - answers can be scored through various ways. These are then given different values
by scorers, e.g., essays and performance tests.
a. Power - allows examinees a generous time limit to be able to answer every item. The questions
are difficult and this difficulty is what is emphasized
b. Speed - with severely limited time constraints but the items are easy and only a few examinees are
expected to make errors.
a. Performance - requires students to perform a task. This is usually administered individually so that
the examiner can count the errors and measure the time the examinee has performed in each task
5. What is Measured
a. Sample - limited representative test designed to measure the total behavior of the examinee,
although no test can exhaustively measure all the knowledge of an individual.
b. Sign test -diagnostic test designed to obtain diagnostic signs suggest that some form of
remediation is needed.
a. Teacher - made test for use within the classroom and contains the subject being taught by the
same teacher who constructed the test.
b. Standardized test - constructed by test specialists working with curriculum experts and teachers.
1. Mastery tests measure the level of learning of a given set of materials and the level attained.
2. Discriminatory tests distinguish the differences between students or groups of students. It indicates the
areas where students need help.
3. Recognition tests require students to choose the right answer from a given set of responses.
4. Recall tests require students to supply the correct answer from their memory.
5. Specific recall tests require short responses that are fairly objective.
6. Free recall tests require students to construct their own complex responses. There are no right answers
but a given answer might be better than the other.
7. Maximum performance tests require students to obtain the best score possible.
9. Written tests depend on the ability of the students to understand, read and write
10. Oral examinations depend on the examinees' ability to speak. Logic is also required.
13. Structured tests have very specific, well-defined instructions and expected outcomes.
14. Projective tests present ambiguous stimulus or questions designed to elicit highly individualized
responses.
16. Process tests focus on how the examinees attack, solve, or work out a problem
17. External reports are tests where a ratee is evaluated by another person.
19. Open book tests depend on one's understanding and ability to express one's ideas and evaluate
concepts.
20. Closed book tests depend heavily on the memory of the examinees.
21. Non-learning format tests determine how much information the students know.
22. Learning format tests require the students to apply previously learned materials
23. Convergent format tests purposely lead the examinees to one best answer.
24. Divergent format tests lead the examinees to several possible answers.
27. Pretests measure how much is known about a material before it is presented
28. Posttests measure how much has been learned after a learning material has been given
29. Sociometric reveals the interrelationship among members or the social structure of a group
30. Anecdotal records reveal episodes of behavior that may indicate a profile of the students.
POST-ASSESSMENT
What I Know about this What I want to know about What do I Learned from this
lesson? this lesson? lesson?
REFERENCES