490 14
490 14
490 14
HS 490
Chapter 14
Evaluation Framework
The evaluation framework can be thought of as the “skeleton” of a plan that can
be used to conduct an evaluation. It puts in order the steps to be followed.
Evaluation Design
An evaluation design is used to organize the evaluation and to provide for
planned, systematic data collection, analysis, and reporting. A well-planned evaluation
design helps ensure that the conclusions drawn about the program will be as accurate as
possible.
Evaluation Approaches
Categorizing the different evaluation approaches. A brief description of each is
presented here:
– Systems analysis uses output measures, such as test scores, to determine if the
program has demonstrated the desired change. It also determines whether funds have
been efficiently used, as in cost analysis.
– Behavioral objectives, or goal-based evaluation, uses the program goals and
collects evidence to determine whether the goals have been reached.
– Decision Making focuses on the decision to be made and presents evidence about
the effectiveness of the program to the decision maker (manager or administrator).
Evaluation Approaches
Goal-free evaluation does not base the evaluation on program goals; instead, the
evaluator searches for all outcomes, often finding unintended side effects.
Art criticism uses the judgment of an expert in the area to increase awareness and
appreciation of the program in order to lead to improved standards and better
performance.
Professional (accreditation) review uses professionals to judge the work of other
professionals; the source of standards and criteria is the professionals conducting the
review.
Evaluation Approaches
Quasi-legal evaluation uses a panel to hear evidence considering the arguments
for and against the program; a quasi-legal procedure is used for both evaluating and
policy making.
Case study uses techniques such as interviews and observations to examine how
people view the program.
Economic Evaluations
Economic evaluations are typical strategies used in systems analysis approaches.
They have been defined as the comparison of alternative courses of action in terms of
both costs and outcomes. Control over rising health care costs has forced many
administrators and planners to be concerned about the cost of health promotion
programs.
Decision-Making Approach
There are three steps to the evaluation process: delineating (focusing of
information), obtaining (collecting, organizing, and analyzing information), and
providing (synthesizing information so it will be useful).
The decision maker, usually a manager or administrator, wants and needs
information to help answer relevant questions regarding a program.
The four types of evaluation in this approach include 1) context, 2) input, 3)
process, and 4) product (CIPP), with each providing information to the decision maker.
Context evaluation describes the conditions in the environment, identifies unmet
needs and unused opportunities, and determines why these occur.
Input evaluation is to determine how to use resources to meet program goals.
Process evaluation provides feedback to those responsible for program
implementation.
Product evaluation is to measure and interpret attainments during and after the
program.
Goal-Free Approach
Suggests that evaluation should not be based on goals in order to enable the
evaluator to remain unbiased. The evaluator must search for all outcomes, including
unintended positive or negative side effects. Thus, the evaluator does not base the
evaluation on reaching goals and remains unaware of the program goals.
Goal-Free Approach
The goal-free approach is not often used in evaluation. It is difficult for
evaluators to determine what to evaluate when program objectives are not to be used.
One concern is that evaluators will substitute their own goals, since there is a lack of
clear methodology as to how to proceed.
Management-Oriented Approaches
Management-oriented approaches focus “on identifying and meeting the
informational needs of managerial decision makers.” That is, good decision making is
best made on good evaluative information. In this approach, the evaluators and managers
work closely together to identify the decisions that must be made and the information
needed to make them. The evaluators then collect the necessary data “about the
advantages and disadvantages of each decision alternative to allow for fair judgment
based on specified criteria. The success of the evaluation rests on the quality of the
teamwork between evaluators and decision makers.”
CIPP
The acronym CIPP stands for the four type decisions facing managers, context,
input, process, and product. Context evaluation describes the conditions in the
environment, identifies unmet needs and unused opportunities, and determines why these
occur. The purpose of input evaluation is to determine how to use resources to meet
program goals. Process evaluation provides feedback to those responsible for program
implementation. The purpose of product evaluation is to measure and interpret
attainments during and after the program. It is the decision maker, not the evaluator, who
uses this information to determine the worth of the program.
Consumer-Oriented Approaches
Consumer-oriented approaches focus on “developing evaluative information on
‘products,’ broadly defined, and accountability, for use by consumers in choosing among
competing products.” This approach gets its “label” of consumer-oriented, in part, from
the fact that it’s an evaluation approach that helps “protect” the consumer by evaluating
“products” used by the consumer. The consumer-oriented approach, which is summative
in nature, primarily uses checklists and criteria to allow the evaluator to collect data that
can be used to rate the “product.” This is the approach used by: Consumer Reports when
evaluating various consumer products, principals when evaluating their teachers, and
instructors when they are evaluating the skill of their students to perform cardio-
respiratory resuscitation (CPR). It is an approach that has been used extensively in
evaluating educational materials and personnel.
The highest level of checklist in the hierarchy is a COMlist. A COMlist is a
checklist comprised of the criteria that essentially define the merit of the “product.” For
example, what are the criteria that define an excellent health promotion program, or an
outstanding program facilitator, or exemplary instructional materials for a program? The
criteria of merit (COM) are identified by being able to answer the question: “What
properties are parts of the concept (the meaning) of ‘a good X’?” Thus if we were to
apply this to program planning, we would ask the question “What are the criteria that
define an excellent health promotion program?” Or, “What are the qualities of an
outstanding facilitator?” Or, “What must be included for an instructional material to be
considered exemplary?”
Expertise-Oriented Approaches
Expertise-oriented approaches, which are probably the oldest of the approaches to
evaluation, rely “primarily on the direct application of professional expertise to judge the
quality of whatever endeavor is evaluated.” Most of these approaches can be placed in
one of three categories, formal professional review systems, or informal professional
reviews, and individual reviews. Formal professional reviews are characterized by
having:
(1) structure or organization established to conduct a periodic review; (2)
published standards (and possibly instruments) for use in such reviews; (3) a
prespecified schedule (for example, every five years) for when reviews will be
conducted; (4) opinions of several experts combining to reach the overall
judgments of value; and (5) an impact on the status of that which is reviewed,
depending on the outcome.
The most common formal professional review system is that of accreditation.
Accreditation is a process by which a recognized professional body evaluates the work
of an organization (i.e., school, universities, and hospitals) to determine if such work
meets prespecified standards. If they do, then the organization is approved or accredited.
Examples of accreditation processes with which readers may be familiar are those of the
National Council for the Accreditation of Teacher Education (NCATE) which accredits
teacher education programs, including health education programs, and the Joint
Commission on Accreditation of Healthcare Organizations (JCAHO) which accredits
various healthcare facilities.
Expertise-Oriented Approaches
In all of the approaches presented so far in this chapter, the primary focus of each has
been on something other than serving the needs of the priority population. It is not that
those who use the previous approaches are unconcerned about the priority population, but
the valuation process does not begin with the priority population. The participant-
oriented approaches are different. They focus on a process “in which involvement of
participants (stakeholders in that which is evaluated) are central in determining the
values, criteria, needs, data, and conclusions for the evaluation” In addition, their
characteristics of less structure and fewer constraints, informal communication and
reporting, and less attention to goals and objectives may be a drawback for those who
want more formal, objective-type evaluation.
Fitzpatrick and colleagues (2004) have identified the following common elements of
participant-oriented approaches:
The early steps provide the foundation, and all steps should be
finalized before moving to the next step:
Step 1- Engaging stakeholders
This step begins the evaluation cycle. Stakeholders must be engaged to insure
that their perspectives are understood. The three primary groups of stakeholders are 1)
those involved in the program operations, 2) those served of affected by the program, and
3) the primary users of the evaluation results. The scope and level of stakeholder
involvement will vary with each program being evaluated.
Step 2- Describing the program:
This step sets the frame of reference for all subsequent decisions in the evaluation
process. At a minimum, the program should be described in enough detail that the
mission, goals, and objectives are known. Also, the program’s capacity to effect change,
its stage of development, and how it fits into the larger organization and community
should be known.
Step 3- Focusing the evaluation design:
This step entails making sure that the interests of the stakeholders are addressed
while using time and resources efficiently. Among the items to consider at this step are
articulating the purpose of the evaluation (i.e., gain insight, change practice, assess
effects, affect participants), determining the users and uses of the evaluation results,
formulating the questions to be asked, determining which specific design type will be
used, and finalizing any agreements about the process.
Step 4- Gathering credible evidence:
This step includes many of the items mentioned in Chapter 5 of this text. At this
step, evaluators need to decide on the measurement indicators, sources of evidence,
quality and quantity of evidence, and logistics for collecting the evidence.
Step 5- Justifying Conclusions
This step includes the comparison of the evidence against the standards of
acceptability; interpreting those comparisons; judging the worth, merit, or significance of
the program; and creating recommendations for actions based upon the results of the
evaluations.
Figure 14.2.
Step 1
The first step is to orient oneself to the situation. The evaluator must identify
resources (time, personnel), constraints, and hidden agendas (unspoken goals). During
this step, the evaluator must determine what is to be expected from the program and what
can be observed.
Step 2
The second step involves defining the problem- determining what is to be
evaluated. During this step, definitions are needed for independent variables (what the
sponsors think makes the difference), dependent variables (what will show the
difference), and confounding variables (what the evaluator thinks could explain
additional differences).
Step 3
The third step involves making a decision about the design- that is, whether to use
qualitative or quantitative methods of data collection or both.
Quantitative Method
The quantitative method is destructive in nature (applying a generally accepted
principle to an individual case), so that the evaluation produces numeric (hard) data, such
as counts, ratings, scores, or classifications. Examples of quantitative data would be the
number of participants in a stress-management program, the ratings on a participant
satisfaction survey, and the pretest scores on a nutrition knowledge test. This approach is
suited to programs that are well defined and compares outcomes of programs with those
of other groups or the general population. It is the method most often used in evaluation
designs.
Qualitative Method
The qualitative method is an inductive method (individual cases are studied to
formulate a general principle) and produces narrative (soft) data, such as descriptions.
This is a good method to use for programs that emphasize individual outcomes or in
cases where other descriptive information from participants is needed.
Comparison Group
When participants cannot be randomly assigned to an experimental or control
group, a nonequivalent control group may be selected.
It is important to find a group that is as similar as possible to the experimental
group, such as two classrooms of students with similar characteristics or a group of
residents in two comparable cities. Factors to consider include:
– Participant's age
– Gender
– Education
– Location
– Socioeconomic status
– Experience
– As well as any other variable that might have an impact on program results.
Evaluation Designs
Measurements used in evaluation designs can be collected at three different times:
after the program; both before and after the program; and several times before, during,
and after the program.
Measurement is defined as the method or procedure of assigning numbers to
objects, events, and people.
Pretest- measurement before the program begins
Posttest- measurement after the completion of the program
Figure 14.5 Evaluation Designs
Figure 14.5 Evaluation Designs
Figure 14.5 Evaluation Designs
Experimental Design
Offers the greatest control over the various factors than may influence the results
Random assignment to experimental and control groups with measurement of
both groups.
Quasi-experimental design
Results in interpretable and supportive evidence of program effectiveness.
Usually cannot control for all factors that affect the validity of the results.
There is no random assignment tot eh groups, and comparisons are made on
experimental and comparison groups.
Non-experimental design
Without the use of a comparison or control group, has little control over the
factors that affect the validity of the results.
The most powerful design is the experimental design, in which participants are
randomly assigned to the experimental and control groups. The difference between I.1.
And I.2. in figure 14.5 is the use of a pretest to measure the participants before the
program begins. Use of a pretest would help assure that the groups are similar. Random
assignment should equally distribute any of the variables (such as age, gender, and race)
between the different groups. Potential disadvantages of the experimental design are that
it requires a relatively large group of participants and that the intervention may be
delayed for those in the control group.
A design more commonly found in evaluations of health promotion programs is
the quasi-experimental pretest-posttest design using a comparison group (II.1 in figure
14.5). This design is often used when a control group cannot be formed by random
assignment. In such a case, a comparison group (a nonequivalent control group) is
identified, and both groups are measured before and after the program. For example, a
program on fire safety for two fifth-grade classrooms could be evaluated by using pre-
and post knowledge test. Two other fifth-grade classrooms not receiving the program
could serve as the comparison group. Similar pretest scored between the comparison and
experimental groups would indicate that the groups were equal at the beginning of the
program. However, without random assignment, it would be impossible to be sure that
other variables (a unit on fire safety in a 4-H group, distribution of smoke detectors,
information from parents) did not influence the results.