ke-unit-2-notes
ke-unit-2-notes
KE Unit 2 notes
lOMoARcPSD|35902250
KNOWLEDGE ENGINEERING
UNIT - 2 NOTES
83
lOMoARcPSD|35902250
Table 3.2 Problem to Be Solved with a Knowledge-based Agent (from Buchanan et al.,
p. 132).
The director of the Oak Ridge National Lab (ORNL) faces a problem. Environmental Protection
Agency (EPA) regulations forbid the discharge of quantities of oil or hazardous chemicals into or
upon waters of the United States when this discharge violates specified quality standards. ORNL
has approximately two thousand buildings on a two-hundred-square-mile government
reservation, with ninety-three discharge sites entering White Oak Creek. Oil and hazardous
chemicals are stored and used extensively at ORNL. The problem is to detect, monitor,
and contain spills of these materials, and this problem may be solved with a
knowledge-based agent.
Table 3.3 Specification of the Actual Problem to Be Solved (from Buchanan et al., p.133).
When an accidental inland spill of an oil or chemical occurs, an emergency situation may exist,
depending on the properties and the quantity of the substance released, the location of the
substance, and whether or not the substance enters a body of water.
The observer of a spill should:
1. Characterize the spill and the probable hazards.
2. Contain the spill material.
3. Locate the source of the spill and stop any further release.
4. Notify the Department of Environmental Management.
What issues may concern the subject matter expert? First of all, the expert may be
concerned that once his or her expertise is represented into the agent, the organization
may no longer need him or her because the job can be performed by the agent. Replacing
human experts was a bad and generally inaccurate way of promoting expert systems.
Usually, the knowledge-based agents and even the expert systems are used by experts in
order to better and more efficiently solve problems from their areas of expertise. They are
also used by people who need the expertise but do not have access to a human expert, or
the expert would be too expensive.
What are some examples of knowledge-based agents? Think, for instance, of any tax-
preparation software. Is it a knowledge-based agent? What about the software systems that
help us with various legal problems, such as creating a will? They all are based on large
amounts of subject matter expertise that are represented in their knowledge bases.
Once the subject matter expert is identified and agrees to work on this project, the
knowledge engineer and the expert have a series of meetings to better define the actual
problem to be solved, which is shown in Table 3.3.
The knowledge engineer has many meetings with the subject matter expert to elicit his
or her knowledge on how to solve the specified problem. There are several knowledge
elicitation methods that can be employed, as discussed later in Section 6.3. Table 3.4
illustrates the unstructured interview, where the questions of the knowledge engineer and
the responses of the expert are open-ended.
lOMoARcPSD|35902250
KE: Suppose you were told that a spill had been detected in White Oak Creek one mile
before it enters White Oak Lake. What would you do to contain the spill?
SME: That depends on a number of factors. I would need to find the source in order to
prevent the possibility of further contamination, probably by checking drains and manholes for
signs of the spill material. And it helps to know what the spilled material is.
KE: How can you tell what it is?
SME: Sometimes you can tell what the substance is by its smell. Sometimes you can
tell by its color, but that's not always reliable since dyes are used a lot nowadays. Oil,
however, floats on the surface and forms a silvery film, while acids dissolve completely in
the water. Once you discover the type of material spilled, you can eliminate any
building that either doesn’t store the material at all or doesn’t store enough of it to account
for the spill.
Table 3.5 Identification of the Basic Concepts and Features Employed by the
Subject Matter Expert
KE: Suppose you were told that a spill had been detected in White Oak Creek one mile
before it enters White Oak Lake. What would you do to contain the spill?
SME: That depends on a number of factors. I would need to find the source in order
to prevent the possibility of further contamination, probably by checking drains
and manholes for signs of the spill material. And it helps to know what the spilled
material is.
KE: How can you tell what it is?
SME: Sometimes you can tell what the substance is by its smell. Sometimes you can
tell by its color, but that's not always reliable since dyes are used a lot nowadays. Oil, however,
floats on the surface and forms a silvery film, while acids dissolve completely in the water. Once
you discover the type of material spilled, you can eliminate any building that either doesn’t store
the material at all or doesn’t store enough of it to account for the spill.
object
subconcept of
building
odor
appearance
instance of
subconcept of
building 3024 source spill
no film
building 3023 instance of
instance of
pungent odor
location
substance
has as type
instance of instance of
subconcept of subconcept of
feature
has as amount
The features are also represented hierarchically, as shown in Figure 3.2. Ontologies are
discussed in detail in Chapter 5.
SME: Sometimes you can tell what the substance is by its smell. Sometimes you can tell by its color,
but that’s not always reliable since dyes are used a lot nowadays. Oil, however, floats on the
surface and forms a silvery film, while acid dissolves completely in the water.
IF the spill . . .
THEN the substance of the spill is oil
IF the spill . . .
THEN the substance of the spill is acid
Table 3.7 Iterative Process of Rules Development and Refinement (Based on Buchanan et al.,
p.138)
KE: Here are some rules I think capture your explanation about determining the substance of the
spill. What do you think?
IF the spill does not dissolve in water
and the spill forms a silvery film
THEN the substance of the spill is oil
IF the spill dissolves in water
and the spill does not form a film
THEN the substance of the spill is acid
SME: Uh-huh (long pause). Yes, that begins to capture it. Of course, if the substance is silver nitrate,
it will dissolve only partially in water.
KE: I see. Rather than talking about a substance dissolving or not dissolving in water, we should talk
about its solubility, which we may consider as being high, moderate, or low. Let’s add that
information to the knowledge base and see what it looks like.
IF the solubility of the spill is low
and the spill forms a silvery film
THEN the substance of the spill is oil
IF the solubility of the spill is moderate
THEN the substance of the spill is silver-nitrate
Testing of the agent involves three types of activity: verification, validation, and certifi-
cation (O’Keefe et al., 1987; Awad, 1996).
In essence, verification attempts to answer the question: Are we building the agent right?
Its goal is to test the consistency and the completeness of the agent with respect to its initial
specification. For example, in the case of a rule-based agent, one would check the rules to
identify various types of errors, such as the existence of rules that are redundant, conflicting,
subsumed, circular, dead-end, missing, unreachable, or with unnecessary IF conditions.
Validation, on the other hand, attempts to answer the question: Are we building the
right agent? In essence, this activity checks whether the agent meets the user’s needs and
requirements.
Finally, certification is a written guarantee that the agent complies with its specified
requirements and is acceptable for operational use.
Various types of tools can be used to develop a knowledge-based agent. We will briefly
discuss three different types: expert system shells, learning agent shells, and learning agent
shells for evidence-based reasoning. We will also discuss the reuse of knowledge in the
development of a knowledge-based agent.
OPS (Cooper and Wogrin, 1988), which has a general rule engine
CLIPS (Giarratano and Riley, 1994), which also has a general rule engine
CYC (Lenat, 1995; CYC, 2008; 2016), a very large knowledge base with ontologies
covering many domains, and with several rule engines
EXPECT (Gil and Paris, 1995; EXPECT 2015), a shell that enables the acquisition of
problem-solving knowledge both from knowledge engineers and from end-users
lOMoARcPSD|35902250
JESS (Friedman-Hill, 2003; JESS, 2016), which is a version of CLIPS with a Java-based
rule engine
CommonKADS (Schreiber et al., 2000), which is a general methodology with support-
ing tools for the development of knowledge-based systems
Jena, which is a toolkit for developing applications for the Semantic Web (Jena, 2012)
Pellet, an ontology (OWL2) reasoner that can be used to develop knowledge-based
applications for the Semantic Web (Pellet, 2012)
Protégé (Musen, 1989; Protégé, 2015), an ontology editor and knowledge base frame-
work, also used to develop knowledge-based applications for the Semantic Web
TopBraid Composer (Allemang and Hendler, 2011; TopBraid Composer, 2012) Ontology
Development Tool for Semantic Web applications
At the power end of the spectrum are shells that employ much more specific problem-
solving methods, such the propose-and-revise design method used in SALT (Marcus, 1988)
to design elevators. The knowledge for such a system can be elicited by simply filling in
forms, which are then automatically converted into rules. Thus the shell provides signifi-
cant assistance in building the system, but the type of systems for which it can be used is
much more limited.
In between these two types of shells are the shells applicable to a certain type of
problems (such as planning, or diagnosis, or design). A representative example is EMYCIN
(van Melle et al., 1981), a general rule-based shell for medical diagnosis.
base structured into an ontology and a set of rules (see Figure 3.3). Building a knowledge-
based agent for a specific application consists of customizing the shell for that application
and developing the knowledge base. The learning engine facilitates the building of the
knowledge base by subject matter experts and knowledge engineers.
Examples of learning agent shells are Disciple-LAS (Tecuci et al., 1999), Disciple-COA
(Tecuci et al., 2001), and Disciple-COG/RKF (Tecuci et al., 2005b), the last two being
presented in Section 12.4.
Problem
Interface
Solving
Ontology
+ Rules
Learning
Problem EBR KB
Solving
Interface
Domain KB Domain KB Domain KB
Scenario KB Scenario KB
Learning
KB = Ontology + Rules
Figure 3.4. The overall architecture of a learning agent shell for evidence-based reasoning.
1. Shell
Customization
Disciple Developer and Knowledge Engineer
Disciple
Learning Agent Multi-
Disciple Ontology
Shell Strategy
Development
Learning
Modules
Modules
EBR KB
Mixed-
Tutoring
Initiative
Domain KB Domain KB Domain KB
Modules
Interaction
Scenario KB Scenario KB
Knowledge Base
En
Management
d
-U
se
r
Problem- Evidence-
Solving Specific
Specialized
Modules Modules
Agent
End-User
4. Field Use
where, based on the specification of the type of problems to be solved and the agent to be
built, the developer and the knowledge engineer may decide that some extensions of the
Disciple shell may be necessary or useful. It is through such successive extensions during
the development of Disciple agents for various applications that the current version of the
Disciple shell for evidence-based reasoning problems (which includes the EBR knowledge
base) has emerged.
The next stage is agent teaching by the subject matter expert and the knowledge
engineer, supported by the agent itself, which simplifies and speeds up the knowledge
base development process (Tecuci et al., 2001; 2002b; 2005b). Once an operational agent is
developed, it is used for the education and training of the end-users, possibly in a
classroom environment.
The fourth stage is field use, where copies of the developed agent support users in their
operational environments. During this stage, an agent assists its user both in solving
lOMoARcPSD|35902250
problems and in collaborating with other users and their cognitive assistants. At the same
time, it continuously learns from this problem-solving experience by employing a form of
nondisruptive learning. In essence, it learns new rules from examples. However, because
there is no learning assistance from the user, the learned rules will not include a formal
applicability condition. It is during the next stage of after action review, when the user and
the agent analyze past problem-solving episodes, that the formal applicability conditions
are learned based on the accumulated examples.
In time, each cognitive assistant extends its knowledge with expertise acquired from its
user. This results in different agents and creates the opportunity to develop a more
competent agent by integrating the knowledge of all these agents. This can be accom-
plished by a knowledge engineer, with assistance from a subject matter expert, during the
next stage of knowledge integration. The result is an improved agent that may be used in a
new iteration of a spiral process of development and use.
Table 3.8 Some Relevant Questions to Consider When Assessing a Potential PhD Advisor
(1) What is the reputation of the PhD advisor within the professional community at large?
(2) Does the advisor have many publications?
(3) Is his or her work cited?
(4) What is the opinion of the peers of this PhD advisor?
(5) What do the students think about this PhD advisor?
(6) Is the PhD advisor likely to remain on the faculty for the duration of your degree program?
(7) What is the placement record of the students of this PhD advisor? Where do they get jobs?
(8) Is the PhD advisor expert in your areas of interest?
(9) Does the PhD advisor publish with students?
(10) Does the PhD advisor have a research group or merely a string of individual students?
(11) Is the PhD advisor’s research work funded?
An analysis of the questions in Table 3.8 shows that some of them point to necessary
conditions that need to be satisfied by the PhD advisor, while others refer to various
desirable qualities. Which questions from Table 3.8 point to necessary conditions? The
answers to questions (6) and (8) need to be “yes” in order to further consider a potential
PhD advisor.
Now let us consider the desirable qualities of a PhD advisor revealed by the other
questions in Table 3.8. Some of these qualities seem to be more closely related than
others. It would be useful to organize them in classes of quality criteria. Could you identify
a class of related criteria? Questions (2), (3), (4), and (11) all characterize aspects of the
professional reputation of the advisor.
What might be other classes of related criteria suggested by the questions in Table 3.8?
Questions (7) and (9) characterize the results of the students of the PhD advisor, while
questions (5) and (10) characterize their learning experience.
lOMoARcPSD|35902250
criterion
instance of
Q2, 3, 4, 11 Q2
research publications criterion
professional
reputation criterion Q3
citations criterion
Q4
has as criterion
PhD advisor
Q7, 9
Q9 publications with advisor criterion
quality criterion
quality of student
results criterion Q7
employers of graduates criterion
has as criterion
Q5, 10
Q10
research group status criterion
student learning
experience criterion
Q5
student opinion criterion
They express the hypothesis in natural language and select the phrases that may be
different for other similar hypotheses, such as the names of the advisor and student.
The selected phrases will appear in blue, guiding the agent to learn a general hypothesis
pattern:
This top-level hypothesis will be successively reduced to simpler and simpler hypotheses,
guided by questions and answers, as shown in Figure 3.8 and discussed in this section.
Bob Sharp should be interested in an area of expertise of John Doe, who should
stay on the faculty of George Mason University for the duration of the PhD
dissertation of Bob Sharp and should have the qualities of a good PhD advisor.
Bob Sharp is John Doe will stay on the faculty John Doe would be a
Is Bob Sharp interested in an Which are the necessary quality criteria for a good PhD advisor?
It is certain that Bob Sharp John Doe would be a good John Doe would be a John Doe would be a good
is interested in an area of PhD advisor with respect good PhD advisor with PhD advisor with respect
expertise of John Doe. to the professional respect to the quality of to the student learning
The reductions of the subhypotheses continues in the same way, until solutions are
obtained for them:
John Doe would be a good PhD advisor with respect to the PhD advisor quality criterion.
Which are the necessary quality criteria for a good PhD advisor?
professional reputation criterion, quality of student results criterion, and student learning
experience criterion.
John Doe would be a good PhD advisor with respect to the professional reputation
criterion.
John Doe would be a good PhD advisor with respect to the quality of student results
criterion.
John Doe would be a good PhD advisor with respect to the student learning experience
criterion.
Each of these subhypotheses can now be reduced to simpler hypotheses, each corres-
ponding to one of the elementary criteria from the right side of Figure 3.7 (e.g.,
research funding criterion). Since each of these reductions reduces a criterion to a
subcriterion, the agent could be asked to learn a general reduction pattern, as shown
in Figure 3.9.
Why is pattern learning useful? One reason is that the pattern can be applied
to reduce a criterion to its subcriteria, as shown in Figure 3.10. Additionally, as will
be illustrated later, the pattern will evolve into a rule that will automatically generate all
the reductions of criteria to their sub-criteria. If, instead of learning a pattern and
Pattern
learning
Which are the necessary quality criteria for a good PhD advisor?
John Doe would be a good John Doe would be a good John Doe would be a good
?O1 would be a good PhD advisor with respect to PhD advisor with respect PhD advisor with respect to
PhD advisor with the professional reputation to the quality of student the student learning
Pattern instantiation
Which is a ?O2? Which is a quality of student results criterion? Which is a quality of student results criterion?
?O1 would be a good John Doe would be a good PhD John Doe would be a good PhD
PhD advisor with advisor with respect to the advisor with respect to the
respect to the ?O3. publications with advisor criterion. employers of graduates criterion.
applying it, the user would manually define these reductions, then any syntactic differ-
ences between these reductions would lead to the learning of different rules. These
rules would only be superficially different, leading to an inefficient and difficult to
maintain agent.
After the top-level criterion (i.e., PhD advisor quality criterion) is reduced to a set of
elementary criteria, specific knowledge and evidence about the advisor need to be used
to evaluate John Doe with respect to each such elementary criterion. For example, the
following hypothesis will be evaluated based on favoring and disfavoring evidence from
John Doe’s peers:
John Doe would be a good PhD advisor with respect to the peer opinion criterion.
A learning agent shell for evidence-based reasoning already knows how to assess such
hypotheses based on evidence.
Through this process, the initial hypothesis is reduced to elementary hypotheses for
which assessments are made. Then these assessments are successively combined, from
bottom-up, until the assessment of the initial hypothesis is obtained, as illustrated in
Figure 3.11.
Notice at the bottom-right side of Figure 3.11 the assessments corresponding to the
subcriteria of the quality of student results criterion:
It is likely that John Doe would be a good PhD advisor with respect to the publications
with advisor criterion.
It is very likely that John Doe would be a good PhD advisor with respect to the employers
of graduates criterion.
It is very likely that John Doe would be a good PhD advisor with respect to the quality of
student results criterion.
lOMoARcPSD|35902250
very likely
min
Bob Sharp should be interested in an area of expertise of John Doe, who should stay on
the faculty of George Mason University for the duration of the PhD dissertation of Bob
Bob Sharp is John Doe will stay on the faculty John Doe would be a
min
It is certain that Bob Sharp John Doe would be a good John Doe would be a John Doe would be a good
is interested in an area of PhD advisor with respect to good PhD advisor with PhD advisor with respect to
expertise of John Doe. the professional reputation respect to the quality of the student learning
max
Which is a quality of student results criterion? Which is a quality of student results criterion?
John Doe would be a good PhD John Doe would be a good PhD
Figure 3.11. Reduction and synthesis tree for assessing a specific hypothesis.
Then this assessment is combined with the assessments corresponding to the other major
criteria (very likely for the professional reputation criterion, and almost certain for the student
learning experience criterion), through a minimum function (because they are necessary
conditions), to obtain the assessment very likely for the PhD advisor quality criterion.
Finally, consider the assessments of the three subhypotheses of the top-level hypothesis:
These assessments are combined by taking their minimum, leading to the following
assessment of the initial hypothesis:
It is very likely that John Doe would be a good PhD advisor for Bob Sharp.
Could you justify the preceding solution synthesis function? We used minimum because
each of the three subhypotheses of the initial hypothesis corresponds to a necessary
condition. If any of them has a low probability, we would like this to be reflected in the
overall evaluation.
lOMoARcPSD|35902250
Notice that, at this point, the knowledge engineer and the subject matter expert have
completely modeled the assessment of the specific hypothesis considered. This is the most
creative and the most challenging part of developing the agent. Once such a model for
assessing hypotheses (or solving problems, in general) is clarified, the agent can be rapidly
prototyped by modeling a set of typical hypotheses. The rest of the agent development
process consists of developing its knowledge base so that the agent can automatically
assess other hypotheses. The knowledge base will consist of an ontology of domain
concepts and relationships and of problem/hypothesis reduction and solution synthesis
rules, as was discussed in Section 1.6.3.1 and illustrated in Figure 1.15 (p. 38). As will be
discussed in the following, the way the preceding assessments were modeled will greatly
facilitate this process.
instance of
is interested in
Is Bob Sharp interested in an
is expert in
is interested in an area of
The goal of this phase is to develop an ontology that is as complete as possible. This will
enable the agent to learn reasoning rules based on the concepts and the features from the
ontology, as will be briefly illustrated in the following.
area of expertise
subconcept of
instance of
Bob Sharp
is interested in
is expert in
John Doe
Figure 3.13. Expanded ontology based on the specification from Figure 3.12.
Bob Sharp should be interested in an area of expertise of John Doe, who should
stay on the faculty of George Mason University for the duration of the PhD
dissertation of Bob Sharp and should have the qualities of a good PhD advisor.
Is Bob Sharp interested in an Which are the necessary quality criteria for a good PhD advisor?
It is certain that Bob Sharp is John Doe would be a good John Doe would be a good John Doe would be a good
interested in an area of PhD advisor with respect to PhD advisor with respect PhD advisor with respect to
expertise of John Doe. the professional reputation to the quality of student the student learning
Which is a professional
reputation criterion?
funding criterion.
Bob Sharp is
interested an area of
is interested in an area of
indicates the conditions that need to be satisfied by these variables so that the IF
hypothesis can be assessed as indicated in the example. For example, ?O1 should be a
PhD student or possibly a person (the agent does not yet know precisely what concept to
apply because the rule is only partially learned), ?O1 should be interested in ?O3, and ?O3
should be Artificial Intelligence or possibly any area of expertise.
The way the question and its answer from the reduction step are formulated is very
important for learning. What could the agent learn if the answer were simply “yes”? The
agent would only be able to learn the fact that “Bob Sharp is interested in an area of
expertise of John Doe.” By providing an explanation of why this fact is true (“Yes, Artificial
Intelligence” meaning: “Yes, because Bob Sharp is interested in Artificial Intelligence which is
an area of expertise of John Doe”), we help the agent to learn a general rule where it will
check that the student ?O1 is interested in some area ?O3, which is an area of expertise of
the advisor ?O2. This is precisely the condition of the rule that can be easily verified
because this type of knowledge was represented in the ontology, as discussed previously
and shown in Figure 3.12.
What is the difference between the pattern learning illustrated in Figure 3.9 and the rule
learning illustrated in Figure 3.15? The difference is in the formal applicability condition of
the rule, which restricts the possible values of the rule variables and allows the automatic
application of the rule in situations where the condition is satisfied. A learned pattern,
such as that from Figure 3.9, cannot be automatically applied because the agent does not
know how to instantiate its variables correctly. Therefore, its application, during the
modeling phase, is controlled by the user, who selects the instances of the variables.
A remarkable capability of the agent is that it learns a general rule, like the one in
Figure 3.15, from a single example rather than requiring the rule to be manually developed
by the knowledge engineer and the subject matter expert. Rule learning will be discussed
in detail in Chapter 9.
lOMoARcPSD|35902250
As indicated in the preceding, the rule in Figure 3.15 is only partially learned, because
instead of an exact applicability condition, it contains an upper and a lower bound for this
condition. The upper bound condition (represented as the larger ellipse from Figure 3.15)
corresponds to the most general generalization of the example (represented as the point
from the center of the two ellipses) in the context of the agent’s ontology, which is used as
a generalization hierarchy for learning. The lower bound condition (represented as the
smaller ellipse) corresponds to the least general generalization of the example.
The next phase is to refine the learned rules and, at the same time, test the agent with
new hypotheses. Therefore, the subject matter expert will formulate new hypotheses,
for example:
Using the learned rules, the agent will automatically generate the reasoning tree from
Figure 3.16. Notice that, in this case, the area of common interest/expertise of Dan Smith
and Bob Sharp is Information Security. The expert will have to check each reasoning step.
Those that are correct represent new positive examples that are used to generalize the
lower bound conditions of the corresponding rules. Those that are incorrect are used as
negative examples. The expert will interact with the agent to explain to it why a reasoning
step is incorrect, and the agent will correspondingly specialize the upper bound condition
of the rule, or both conditions. During the rule refinement process, the rule’s conditions
will converge toward one another and toward the exact applicability condition. During this
process, the ontology may also be extended, for instance to include the new concepts used
to explain the agent’s error. Rule refinement will be discussed in detail in Chapter 10.
Bob Sharp should be interested in an area of expertise of Dan Smith, who should
stay on the faculty of George Mason University for the duration of the PhD
dissertation of Bob Sharp and should have the qualities of a good PhD advisor.
Bob Sharp is Dan Smith will stay on the faculty Dan Smith would be a
area of expertise the duration of the PhD respect to the PhD advisor
Is Bob Sharp interested in an Which are the necessary quality criteria for a good PhD advisor?
It is certain that Bob Sharp Dan Smith would be a good Dan Smith would be a good Dan Smith would be a good
is interested in an area of PhD advisor with respect to PhD advisor with respect to PhD advisor with respect to
expertise of Dan Smith. the professional reputation the quality of student the student learning
of John Doe.
in an area of expertise of
Domain-
John Doe?
dependent
generic i
Domain KB Domain KB
Yes, Artificial Intelligence.
concepts,
instance Domain knowledge … Domain knowledge
generic
adv
(PhD advisor) (Insider threat)
It is certain that Bob instances,
of John Doe.
specific
Domain-
instance i
Scenario KB Scenario KB
dependent
Scenario knowledge … Scenario knowledge
specific
(George Mason Univ.) (Univ. of Virginia)
instances
specification assessment
P1 S1
P1 S1
QuestionP1 S1
Question
Answer
Question
Answer
Question
Answer
Question
Answer
Question
Answer
Rapid Answer IF the problem to solve is P1g
P1 S1
1 1 P 1 S1
n n
IF the problem to solve is P1g
IF the problem to solve is P1g
IF the problem to solve is P1g
P11 S11 Pn1 S1n THEN solve its sub-problems
THEN solve its sub-problems
prototyping
P11 S11 Pn1 S1n 1 … P
P1g 1
THEN
Png
1 … P solve
1
THEN
its sub-problems
solve its sub-problems
Question 1g Png
1 … P 1
Question
Answer Examples: E1, … , Ek 1g Png
1 … P1
ng
Question
Answer Examples: E1, … , Ek 1g
Question
Answer Examples: E1, … , Ek
Question Examples: E1, … , Ek
Answer
Question
Answer
P21 S2 Answer
2 2
1 Pm Sm
P21 S21 2
Pm 2
Sm
P21 S21 2
Pm 2
Sm
actor
Ontology
person
organization
graduate educational
faculty member staff member company
student organization
actor
ontology refinement
… Except-When Condition company
Except-When Condition organization
… faculty member staff member
Except-When Condition
THEN solve its sub-problems graduate
Except-When Condition
THEN solve its sub-problems
P
1 1
… THEN
P Except-When Condition student
1g 1ng 1
solve its sub-problems
P … THEN
P solve its sub-problems
1ng 1
1g
P … THEN
P solve its sub-problems
1g 1ng 1
P … THEN
P solve its sub-problems instructor professor PhD advisor
1g 1ng 1
P … P
1g 1 ng 1
P … P college university
1g ng
P1 S1
P1 S1
QuestionP1 S1
Question
Answer IF the problem to solve is P1g
Question
Answer IF the problem to solve is P1g
Question
Answer
Question IF the problem to solve is P1g
Answer THEN solve its sub-problemsIF the problem to solve is P1g
Question
Agent use and
Answer THEN
1 … P
P1g
solve
1 its sub-problems
Answer THEN solve its sub-problems
P1 S1
1 P
1
1 S1
n n
Png
1 … P
1g
1
THEN
Png
1 … P solve
1 its sub-problems
Examples: E1, … , Ek 1g Png
1 … P1
P11 S11 Pn1 S1n Examples: E1, … , Ek 1g
ng
person
IF the problem to solve is P1g organization
… Except-When
Except-When Condition Condition graduate undergraduate
… Except-When Condition
optimization
THEN Except-When Condition
1 … P1
P1g
solve
THEN
its sub-problems
solve its sub-problemsExcept-When …
Condition
Png its sub-problemsExcept-When Condition
1 … P 1
THEN solve instructor professor PhD advisor
Figure 3.18. Main phases of agent development when using learning technology.
lOMoARcPSD|35902250
Successively reducing it, from the top down, to simpler and simpler hypotheses
Assessing the simplest hypotheses
Successively combining, from the bottom up, the assessments of the subhypotheses,
until the assessment of the top-level hypothesis is obtained
Periodically, the agent can undergo an optimization phase, which is the last phase in
Figure 3.18. During this phase, the knowledge engineer and the subject matter expert will
review the patterns learned from the end-user, will learn corresponding rules from them,
and will correspondingly refine the ontology. The current version of the Disciple-EBR shell
reapplies its rule and ontology learning methods to do this. However, improved methods
can be used when the pattern has more than one set of instances, because each represents
a different example of the rule to be learned. This is part of future research.
Figure 3.19 compares the conventional knowledge engineering process of developing a
knowledge-based agent (which was discussed in Section 3.1) with the process discussed in
this section, which is based on the learning agent technology.
The top part of Figure 3.19 shows the complex knowledge engineering activities that are
required to build the knowledge base. The knowledge engineer and the subject matter
expert have to develop a model of the application domain that will make explicit the way
the subject matter expert assesses hypotheses. Then the knowledge engineer has to
develop the ontology. He or she also needs to define general hypotheses decomposition
rules and to debug them, with the help of the subject matter expert.
As shown at the bottom of Figure 3.19, each such activity is replaced with an equivalent
activity that is performed by the subject matter expert and the agent, with limited assistance
from the knowledge engineer. The knowledge engineer still needs to help the subject matter
expert to define a formal model of how to assess hypotheses and to develop the ontology.
After that, the subject matter expert will teach the agent how to assess hypotheses, through
examples and explanations, and the agent will learn and refine the rules by itself.
The next chapters discuss each phase of this process in much more detail.
The knowledge bases developed with Disciple-EBR are located in the repository folder,
which is inside the installation folder. As shown in Figure 3.17, the knowledge bases are
Subject
ct Dialogue
Conventional
Matter
er Knowledge
Knowledge Engineer
Engineering
Expert
rtt
Programming
Model Define Verify and
Understand Develop
problem reasoning update rules
domain ontology
solving rules and ontology
Develop
D Define concepts Provide and Analyze
Explain
reasoning with hierarchical explain agent’s
errors
model organization examples solutions
Knowledge
Assist in Extend ontology Learn Learn
Engineer Refine
(support role)) solution based on modeled reasoning ontology
rules
development solutions rules elements
organized hierarchically, with the knowledge base for evidence-based reasoning at the top
of the hierarchy. The user cannot change this knowledge base, whose knowledge elements
are inherited in the domain and scenario knowledge bases. From the user’s point of view,
each knowledge base consists of a top-level domain part (which contains knowledge
common to several applications or scenarios in a domain) and one scenario part (con-
taining knowledge specific to a particular application or scenario). As illustrated in
Figure 3.17, there can be more than one scenario under a domain. In such a case, the
domain and each of the scenarios correspond to a different knowledge base. Loading,
saving, or closing a scenario will automatically load, save, or close both the scenario part
and the corresponding domain part of the knowledge base.
Loading and selecting a knowledge base are described in Operation 3.1 and illustrated
in Figure 3.20.
5. Click on Select
4. Click on Scen
Once a knowledge base is selected, you can invoke different modules of Disciple-EBR to
use it. Each module is accessible in a specific workspace. As illustrated in Figure 3.21,
there are three workspaces:
The user can switch between the workspaces to use the corresponding modules. For
example, you must switch to the Evidence workspace to work with evidence items. Then,
to save the knowledge base, you must switch to the Scenario workspace.
The user should work with only one set of the three workspaces at a time (correspond-
ing to the same KB). Therefore, you should close the workspaces corresponding to a
knowledge base (by clicking on the icon with the minus sign [–]) before opening the
workspaces corresponding to another knowledge base.
The steps to save all the knowledge bases loaded in memory are described in Operation
3.2 and illustrated in Figure 3.22.
Scenario workspace
Domain workspace
Evidence workspace
2. Scenario workspace
3. Select System
Save All
The following are knowledge engineering guidelines for knowledge base development.
Guideline 3.1. Work with only one knowledge base loaded in memory
To maintain the performance of the Disciple-EBR modules, work with only one knowledge
base loaded in memory. Therefore, close all the knowledge bases before loading a new one.
If several knowledge bases are loaded, work only with the set of three workspaces
corresponding to the knowledge base you are currently using. Close all the other
workspaces.
“WA” has errors, and the most recently saved version is “WA-4o.”
Delete “WA” in Windows Explorer, copy “WA-4o,” and rename this copy as “WA.”
Continue with the development of “WA.”
Finalize the project team, specify the type of hypotheses to be analyzed by the agent to
be developed, and study the application domain. Prepare a short presentation of the
following:
The application domain of your agent and why your agent is important.
A bibliography containing the expertise domain and your current familiarity with the
domain, keeping in mind that you should choose a domain where you already are or
could become an expert without investing a significant amount of time.
Three examples of hypotheses and a probability for each.
lOMoARcPSD|35902250
3.1. Which are the main phases in the development of a knowledge-based agent?
3.5. Consider the fact that a knowledge-based agent may need to have hundreds or
thousands of rules. What can be said about the difficulty of defining and refining
these rules through the conventional process discussed in Section 3.1.4?
3.6. Use the scenario from Section 3.1 to illustrate the different difficulties of building a
knowledge-based agent discussed in Section 1.6.3.1.
3.11. What is the organization of the knowledge repository of a learning agent shell for
evidence-based reasoning?
3.12. What is the difference between a specific instance and a generic instance? Provide
an example of each.
3.13. Are there any mistakes in the reasoning step from Figure 3.24 with respect to the
goal of teaching the agent? If the answer is yes, explain and indicate corrections.
3.14. Which are the main stages of developing a knowledge-based agent using learning
agent technology?
Yes
Analysis and synthesis, introduced in Section 1.6.2, form the basis of a general divide-and-
conquer problem-solving strategy that can be applied to a wide variety of problems. The
general idea, illustrated in Figure 4.1, is to decompose or reduce a complex problem P1 to
n simpler problems P11, P12, . . . , P1n, which represent its components. If we can then find
the solutions S11, S12, . . . , S1n of these subproblems, then these solutions can be combined
into the solution S1 of the problem P1.
If any of the subproblems is still too complex, it can be approached in a similar way, by
successively decomposing or reducing it to simpler problems, until one obtains problems
whose solutions are known, as illustrated in Figure 4.2.
Figures 4.3 and 4.4 illustrate the application of this divide-and-conquer approach to
solve a symbolic integration problem. Notice that each reduction and synthesis operation
is justified by a specific symbolic integration operator.
Cognitive assistants require the use of problem-solving paradigms that are both natural
enough for their human users and formal enough to be automatically executed by the
agents. Inquiry-driven analysis and synthesis comprise such a problem-solving paradigm
where the reduction and synthesis operations are guided by corresponding questions and
answers. The typical questions are those from Rudyard Kipling’s well-known poem “I Keep
Six Honest . . .”
P1 S 1
113
lOMoARcPSD|35902250
A complex problem P is P1 S1
1
solved by:
• Successively reducing
it to simpler and
simpler problems;
P11 S11 P12 S12 … P1n S1n
• Finding the solutions
of the simplest
problems;
• Successively
problem is obtained
Problem 1
Question Q
Question Q
Answer B
Answer A
We have already illustrated this paradigm in Sections 2.2 and 3.3.2. To better understand
it, let us consider the simple abstract example from Figure 4.5. To solve Problem 1, one
asks Question Q related to some aspect of Problem 1. Let us assume that there are two
answers to Q: Answer A and Answer B. For example, the question, “Which is a sub-
criterion of the quality of student results criterion?” has two answers, “publications with
advisor criterion” and “employers of graduates criterion.”
Let us further assume that Answer A leads to the reduction of Problem 1 to two simpler
problems, Problem 2 and Problem 3. Similarly, Answer B leads to the reduction of
Problem 1 to the other simpler problems, Problem 4 and Problem 5.
Let us now assume that we have obtained the solutions of these four subproblems. How
do we combine them to obtain the solution of Problem 1? As shown in Figure 4.6, first the
lOMoARcPSD|35902250
Problem 1
Solution
of Problem 1
Problem-
level
synthesis
Question Q
Question Q
Answer B
Answer A
Solution B
Solution A
of Problem 1
of Problem 1
Reduction-
level
synthesis
Problem 2 Problem 3 Problem 4 Problem 5
Figure 4.6. A more detailed view of the analysis and synthesis process.
Test whether President Roosevelt has the critical capability to maintain support.
Which are the critical requirements for President Roosevelt to maintain support?
President Roosevelt needs means to secure support from the government, means to
secure support from the military, and means to secure support from the people.
Test whether President Roosevelt Test whether President Roosevelt Test whether President Roosevelt
has means to secure support from has means to secure support from has means to secure support from
Test whether President Roosevelt has the critical capability to maintain support.
President Roosevelt has the critical capability to maintain support because President
Roosevelt has means to secure support from the government, has means to secure
support from the military, and has means to secure support from the people.
critical capability is
an emerging property
needs means to secure support from the government, has means to secure support from the government,
means to secure support from the military, and means has means to secure support from the military, and
Yes, because President Roosevelt has all the needed critical requirements.
Test whether President Roosevelt has Test whether President Roosevelt Test whether President Roosevelt
means to secure support from the has means to secure support from has means to secure support from
President Roosevelt has means to President Roosevelt has means to President Roosevelt has means to
secure support from the government. secure support from the military. secure support from the people.
Figure 4.8. Illustration of reduction and synthesis operations in the COG domain.
Moreover, it is assumed that the questions guiding the synthesis operations may have
only one answer, which typically indicates how to combine the solutions. Allowing more
questions and more answers in the synthesis tree would lead to a combinatorial explo-
sion of solutions.
Another interesting aspect is that the three leaf solutions in Figure 4.8 are about the
means of President Roosevelt, while their composition is about a capability. Thus this
illustrates how synthesis operations may lead to emerging properties.
A third aspect to notice is how the reduction-level composition is actually performed. In
the example from Figure 4.8, the solutions to combine are:
President Roosevelt has the critical capability to maintain support because President
Roosevelt has means to secure support from the government, has means to secure
support from the military, and has means to secure support from the people.
Figure 4.9 shows another example of reduction and synthesis in the COG domain. In
this case, the solutions to combine are:
PM Mussolini does not have the critical capability to maintain support because PM
Mussolini does not have means to secure support from the people.
Additional examples of solution synthesis from the COG domain are presented in
Figures 12.27 (p. 372), 12.28 (p. 373), 12.29 (p. 373), and 12.30 (p. 374) from Section 12.4.2.
As suggested by the preceding examples, there are many ways in which solutions may
be combined.
One last important aspect related to problem solving through analysis and synthesis is
that the solutions of the elementary problems may be obtained by applying any other type of
reasoning strategy. This enables the solving of problems through a multistrategy approach.
Chapter 12 presents Disciple cognitive assistants for different types of tasks, illustrating
the use of the inquiry-driven analysis and synthesis in different domains. Section 12.2
discusses this problem-solving paradigm in the context of military engineering planning.
Section 12.3 discusses it in the context of course of action critiquing. Section 12.4 discusses
it in the context of center of gravity analysis, and Section 12.5 discusses it in the context of
collaborative emergency response planning.
PM Mussolini does not have the critical capability to maintain support because
PM Mussolini does not have means to secure support from the people.
Which are the critical requirements for PM Mussolini to PM Mussolini does not have the critical
maintain support? PM Mussolini needs means to secure capability to maintain support because PM
support from the government, means to secure support from Mussolini does not have means to secure
the military, and means to secure support from the people. support from the people.
No, because PM Mussolini does not have all the needed critical requirements.
Test whether PM Mussolini has Test whether PM Mussolini has Test whether PM Mussolini has
means to secure support from the means to secure support from means to secure support from the
PM Mussolini has means to secure PM Mussolini has means to PM Mussolini does not have means
support from the government. secure support from the military. to secure support from the people.
Figure 4.9. Another illustration of reduction and synthesis operations in the COG domain.
lOMoARcPSD|35902250
Successively reducing it, from the top down, to simpler and simpler hypotheses
(guided by introspective questions and answers).
Assessing the simplest hypotheses based on evidence.
Successively combining, from the bottom up, the assessments of the simpler hypoth-
eses, until the assessment of the top-level hypothesis is obtained.
Figure 4.10 shows a possible analysis of the hypothesis that Country X has nuclear
weapons.
likely
max
likely
min
max
likely no support
max max
likely no support
max
Which is an indicator?
likely
enriched uranium.
likely
Figure 4.10. An example of different types of reductions and corresponding synthesis functions.
lOMoARcPSD|35902250
Hypothesis 1
Assessment of
Hypothesis 1
max
Scenario 1 Scenario 2
min min
Figure 4.11. Reductions and syntheses corresponding to two sufficient conditions (scenarios).
4.3.4 Indicators
Many times when we are assessing a hypothesis, we have only indicators. For example, as
shown at the bottom part of Figure 4.10, having the capability to produce enriched uranium
is an indicator that a country can build nuclear weapons. An indicator is, however, weaker
than a sufficient condition. If we determine that a sufficient condition is satisfied (e.g., a
scenario has actually happened), we may conclude that the hypothesis is true. But we
cannot draw such a conclusion just because we have discovered an indicator. However, we
may be more or less inclined to conclude that the hypothesis is true, based on the relevance
(strength) of the indicator. Therefore, given the symbolic probabilities from Table 2.5, we
distinguish between three types of indicators of different relevance (strength): “likely indica-
tor,” “very likely indicator,” and “almost certain indicator.”
A “likely indicator” is one that, if discovered to be true, would lead to the conclusion
that the considered hypothesis is likely. Similarly, a “very likely indicator” would lead to the
conclusion that the hypothesis is very likely, and an “almost certain indicator” would lead to
the conclusion that the hypothesis is almost certain.
In the example from the bottom part of Figure 4.10 it is likely that Country X can
produce enriched uranium, and this is a very likely indicator that Country X can build
nuclear weapons. Therefore, we can conclude that the probability of the hypothesis
that Country X can build nuclear weapons is likely, the minimum between likely (the
probability of the indicator) and very likely (the strength of the indicator).
In general, the probability of a hypothesis H based on an indicator I is the minimum
between the probability of the indicator and the relevance (strength) of the indicator (which
could be likely, very likely, or almost certain).
lOMoARcPSD|35902250
It makes no sense to consider the type “certain indicator,” because this would be a
sufficient condition. Similarly, it makes no sense to consider the type “no support indica-
tor,” because this would not be an indicator.
As an abstract example, Figure 4.12 shows a hypothesis that has two likely indicators,
A and B, if only one of them is observed. However, if both of them are observed, they
synergize to become an almost certain indicator.
As a concrete example, consider Person , who has been under surveillance in
connection with terrorist activities. We suspect that will attempt to leave the country
in a short while. Three days ago, we received information that sold his car. Today, we
received information that he closed his account at his bank. Each of these is only a likely
indicator of the hypothesis that plans to leave the country. He could be planning to buy
a new car, or he could be dissatisfied with his bank. But, taken together, these two
indicators suggest that it is almost certain that is planning to leave the country.
Coming back to the abstract example in Figure 4.12, let us assume that indicator A is
almost certain and indicator B is very likely. In such a case, the assessment of Hypothesis
1, based only on indicator A, is minimum(almost certain, likely) = likely. Similarly, the
assessment of Hypothesis 1, based only on indicator B, is minimum(very likely, likely) =
likely. But the assessment of Hypothesis 1, based on both indicators A and B, is
minimum(minimum(almost certain, very likely), almost certain) = very likely. Also, the
assessment of Hypothesis 1 based on all the indicators is the maximum of all the
individual assessments (i.e., very likely), because these are three alternative solutions
for Hypothesis 1.
Now we discuss the assessment of the leaf hypotheses of the argumentation structure,
based on the identified relevant evidence. Let us consider an abstract example where the
leaf hypothesis to be directly assessed based on evidence is Q (see Figure 4.13).
We begin by discussing how to assess the probability of hypothesis Q based only on one
item of favoring evidence Ek* (see the bottom of Figure 4.13). First notice that we call
this likeliness of Q, and not likelihood, because in classic probability theory, likelihood is
Hypothesis 1
very likely
max
Q very likely
on balance
c
Q based only on almost Q based only on
likely
favoring evidence certain disfavoring evidence
max
min
How likely is Q, based only on Ek* How likely is it that Ek* is true?
P(Ek*|Q), while here we are interested in P(Q|Ek*), the posterior probability of Q given Ek*.
To assess Q based only on Ek*, there are three judgments to be made by answering
three questions:
The relevance question is: How likely is Q, based only on Ek* and assuming that Ek* is
true? If Ek* tends to favor Q, then our answer should be one of the values from likely
to certain. If Ek* is not relevant to Q, then our answer should be no support, because
Ek* provides no support for the truthfulness of Q. Finally, if Ek* tends to disfavor Q,
then it tends to favor the complement of Q, that is, Qc. Therefore, it should be used
as favoring evidence for Qc, as discussed later in this section.
The believability question is: How likely is it that Ek* is true? Here the answer should be
one of the values from no support to certain. The maximal value, certain, means
that we are sure that the event Ek reported in Ek* did indeed happen. The minimal
value, no support, means that Ek* provides us no reason to believe that the event
Ek reported in Ek* did happen. For example, we believe that the source of Ek* has
lied to us.
The inferential force or weight question is: How likely is Q based only on Ek*? The agent
automatically computes this answer as the minimum of the relevance and believ-
ability answers. What is the justification for this? Because to believe that Q is true
based only on Ek*, Ek* should be both relevant to Q and believable.
When we assess a hypothesis Q, we may have several items of evidence, some favoring Q
and some disfavoring Q. The agent uses the favoring evidence to assess the probability of
Q and the disfavoring evidence to assess the probability of Qc. As mentioned previously,
because the disfavoring evidence for Q is favoring evidence for Qc, the assessment process
for Qc is similar to the assessment for Q.
When we have several items of favoring evidence, we evaluate Q based on each of them
(as was explained previously), and then we compose the obtained results. This is illus-
trated in Figure 4.13, where the assessment of Q based only on Ei* (almost certain) is
lOMoARcPSD|35902250
composed with the assessment of Q based only on Ek* (likely), through the maximum
function, to obtain the assessment of Q based only on favoring evidence (almost certain). In
this case, the use of the maximum function is justified because it is enough to have one
item of evidence that is both very relevant and very believable to persuade us that the
hypothesis Q is true.
Let us assume that Qc based only on disfavoring evidence is likely. How should we
combine this with the assessment of Q based only on favoring evidence? As illustrated at
the top of Figure 4.13, the agent uses an on-balance judgment: Because Q is almost certain
and Qc is likely, it concludes that, based on all available evidence, Q is very likely.
In general, as indicated in the right and upper side of Table 4.1, if the assessment of Qc
(based on disfavoring evidence for Q) is higher than or equal to the assessment of Q
(based on favoring evidence), then we conclude that, based on all the available evidence,
there is no support for Q. If, on the other hand, the assessment of Q is strictly greater than
the assessment of Qc, then the assessment of Q is decreased, depending on the actual
assessment of Qc (see the left and lower side of Table 4.1).
One important aspect to notice is that the direct assessment of hypotheses based on
favoring and disfavoring evidence is done automatically by the agent, once the user
assesses the relevance and the believability of evidence.
Another important aspect to notice is that the evaluation of upper-level hypotheses (such
as those from Figure 4.10) requires the user to indicate what function to use when
composing the assessments of their direct subhypotheses. This was discussed in Section 4.3.
no no no no no no
no no no no
likely likely
support support support support
very very no no no
likely
likely likely support support support
almost very no
certain certain likely
certain likely support
lOMoARcPSD|35902250
established that the cesium-137 canister is missing (see Figure 2.8 on p. 62). The next step
is to consider the competing hypotheses:
We have to put each of these hypotheses to work, to guide the collection of relevant
evidence. In Section 2.2.4, we have already discussed, at a conceptual level, the collec-
tion of evidence for hypothesis H2. Table 4.2 shows the result of our information
collection efforts.
The collected information from Table 4.2 suggests that the cesium-137 canister was
stolen with the panel truck having Maryland license MDC-578. This has led to the
development of the analysis tree in Figure 2.9 (p. 63). In this case study, you are going
to actually perform this analysis. You have to identify the “dots” in the information from
Table 4.2, which are fragments representing relevant items of evidence for the leaf
hypotheses in Figure 2.9. These dots are presented in Table 4.3.
This case study has several objectives:
When you associate an item of evidence with a hypothesis, the agent automatically
generates a decomposition tree like the one in Figure 4.14. The bottom part of Figure 4.14
shows the abstraction of the tree that is automatically generated by the agent when you
indicate that the item of evidence E005-Ralph favors the leaf hypothesis “The XYZ hazardous
material locker was forced.”
The agent also automatically generates the reduction from the top of Figure 4.14, where
the leaf hypothesis, “The XYZ hazardous material locker was forced,” is reduced to the
elementary hypothesis with the name, “The XYZ hazardous material locker was forced,” to be
directly assessed based on evidence. Although these two hypotheses are composed of the
same words, internally they are different, the latter being an instance introduced in the
agent’s ontology. This elementary hypothesis corresponds to the hypothesis Q in
Figure 4.13. The agent decomposes this hypothesis as shown in the bottom part of
Figure 4.14, which corresponds to the tree in Figure 4.13 except that there is only one
item of favoring evidence, namely E005-Ralph. After that, you have to assess the relevance
of this item of evidence to the considered hypothesis (e.g., likely), as well as its believability
(e.g., very likely), and the agent automatically composes them, from the bottom up, to
obtain the assessment of the leaf hypothesis. When you add additional items of evidence
as either favoring or disfavoring evidence, the agent extends the reasoning tree from
Figure 4.14 as indicated in Figure 4.13.
Figure 4.15 illustrates the selection of a synthesis function indicating how to evaluate
the probability of a node based on the probability of its children. You have to right-click on
the node (but not on any word in blue), select New Solution with. . ., and then select the
function from the displayed list.
Now you can perform the case study. Start Disciple-EBR, select the case study know-
ledge base “02-Evidence-based-Analysis/Scen,” and proceed as indicated in the instruc-
tions from the bottom of the opened window.
This case study illustrates several basic hypothesis analysis operations described in the
following.
label. Clicking on [REMOVE] will restore the leaf hypothesis under the Irrelevant
to label.
To associate another evidence item to a hypothesis, click on it in the left panel and
repeat the preceding operations.
To return to the Reasoner module, click on [REASONING] following the hypothesis.
lOMoARcPSD|35902250
The objective of this case study is to learn how to use Disciple-EBR to analyze hypotheses based
on evidence retrieved from the Internet, by associating search criteria with elementary hypoth-
eses, invoking various search engines (such as Google, Yahoo!, or Bing), identifying relevant
information, extracting evidence from it, and using the evidence to evaluate the hypotheses.
This case study concerns the hypothesis that the United States will be a global leader in
wind power within the next decade.
To search for evidence that is relevant to a leaf hypothesis, the agent guides you to
associate search criteria with it and to invoke various search engines on the Internet.
Figure 4.19 shows the corresponding interface of the Evidence module. Because the
[COLLECTION GUIDANCE] mode is selected in the left panel, it shows all the leaf hypotheses
and their current evidential support. If you click on one of these hypotheses, such as
“United States imports huge quantities of oil,” it displays this hypothesis in the right panel,
enabling you to define search criteria for it. You just need to click on the [NEW] button
following the Search criterion label, and the agent will open an editor in which you can
enter the search criterion.
Figure 4.20 shows two defined search criteria: “oil import by United States” and “top oil
importing countries.” You can now invoke Bing, Google, or Yahoo! with any one of these
criteria to search for relevant evidence on the Internet. This will open a new window with
the results of the search, as shown in Figure 4.21.
lOMoARcPSD|35902250
You have to browse the retrieved documents shown in Figure 4.21 and determine
whether any of them contains information that is relevant to the hypothesis that the
United States imports huge quantities of oil. Such a document is the second one, whose
content is shown in Figure 4.22.
You can now define one or several items of evidence with information copied from the
retrieved document, as illustrated in Figure 4.23. In the left panel of the Evidence module,
you switch the selection mode to [AVAILABLE EVIDENCE] and then click on [NEW]. As a
result, the right panel displays a partial name for the evidence E001- to be completed by
you. You then have to click on the [EDIT] button, which opens an editor where you can
copy the description of this item of evidence from the retrieved document. The result is
shown in the right panel of Figure 4.23.
You can define additional characteristics of this item of evidence, such as its type (as
will be discussed in Section 4.7), and you should indicate whether this item of evidence
favors or disfavors the hypothesis that the United States imports huge quantities of oil, as
explained previously.
lOMoARcPSD|35902250
In this case study, you will first select the hypothesis, ”United States will be a global
leader in wind power within the next decade,” and then you will browse its analysis tree to
see how it is reduced to simpler hypotheses that you have to assess by searching evidence
on the Internet. You will associate specific search criteria with the leaf hypotheses, invoke
specific search engines with those criteria, identify relevant Web information, define
evidence from this information, associate evidence with the corresponding hypotheses,
and evaluate its relevance and believability, with the goal of assessing the probability of the
top-level hypothesis.
Start Disciple-EBR, select the case study knowledge base “03-Evidence-Search/Scen,”
and proceed as indicated in the instructions from the bottom of the opened window.
This case study illustrates the following hypothesis analysis operation:
In the previous sections, we have discussed and illustrated how you may directly assess
the believability of an item of evidence. However, the Disciple-EBR agent has a significant
amount of knowledge about the various types of evidence and its believability credentials,
enabling you to perform a much deeper believability analysis, as will be discussed in this
section. You may wish to perform such a detailed believability analysis for those items of
evidence that are critical to the final result of the analysis. We will start with presenting a
classification or ontology of evidence.
Attempts to categorize evidence in terms of its substance or content would be a fruitless
task, the essential reason being that the substance or content of evidence is virtually
unlimited. What we have termed a substance-blind classification of evidence refers to a
classification of recurrent forms and combinations of evidence, based not on substance or
content, but on the inferential properties of evidence (Schum, 1994 [2001a], pp. 114–130;
Schum, 2011). In what follows, we identify specific attributes of the believability of various
recurrent types of evidence without regard to their substance or content.
Here is an important question you are asked to answer regarding the individual
kinds of evidence you have: How do you stand in relation to this item of evidence? Can
you examine it for yourself to see what events it might reveal? If you can, we say that the
evidence is tangible in nature. But suppose instead you must rely upon other persons
to tell you about events of interest. Their reports to you about these events are
examples of testimonial evidence. Figure 4.24 shows a substance-blind classification
of evidence based on its believability credentials. This classification is discussed in
the following sections.
evidence
There are two different kinds of tangible evidence: real tangible evidence and demon-
strative tangible evidence (Lempert et al., 2000, pp. 1146–1148). Real tangible evidence is
an actual thing and has only one major believability attribute: authenticity. Is this object
what it is represented as being or is claimed to be? There are as many ways of generating
deceptive and inauthentic evidence as there are persons wishing to generate it. Docu-
ments or written communications may be faked, captured weapons may have been
tampered with, and photographs may have been altered in various ways. One problem
is that it usually requires considerable expertise to detect inauthentic evidence.
Demonstrative tangible evidence does not concern things themselves but only repre-
sentations or illustrations of these things. Examples include diagrams, maps, scale models,
statistical or other tabled measurements, and sensor images or records of various sorts
such as IMINT, SIGINT, and COMINT. Demonstrative tangible evidence has three believ-
ability attributes. The first concerns its authenticity. For example, suppose we obtain a
hand-drawn map from a captured insurgent showing the locations of various groups in his
insurgency organization. Has this map been deliberately contrived to mislead our military
forces, or is it a genuine representation of the location of these insurgency groups?
The second believability attribute is accuracy of the representation provided by the
demonstrative tangible item. The accuracy question concerns the extent to which the
device that produced the representation of the real tangible item had a degree of sensitiv-
ity (resolving power or accuracy) that allows us to tell what events were observed. We
would be as concerned about the accuracy of the hand-drawn map allegedly showing
insurgent groups locations as we would about the accuracy of a sensor in detecting traces
of some physical occurrence. Different sensors have different resolving power that also
depends on various settings of their physical parameters (e.g., the settings of a camera).
The third major attribute, reliability, is especially relevant to various forms of sensors
that provide us with many forms of demonstrative tangible evidence. A system, sensor,
or test of any kind is reliable to the extent that the results it provides are repeatable or
lOMoARcPSD|35902250
consistent. You say that a sensing device is reliable if it provides the same image or report
on successive occasions on which this device is used.
The left side of Figure 4.25 shows how the agent assesses the believability of an item of
demonstrative tangible evidence Ei* as the minimum of its authenticity, accuracy, and
reliability.
Here are additional examples involving evidence that is tangible and that you can
examine personally to see what events it reveals.
Have a look at evidence item E009-MDDOTRecord in Table 4.3 (p. 126). The Maryland
DOT record, in the form of a tangible document, could be given to the analyst to verify
that the vehicle carrying MD license plate number MDC-578 is registered in the name of
the TRUXINC Company in Silver Spring, Maryland.
Now consider evidence item E008-GuardReport in Table 4.3. Here we have a document
in the form of a log showing that the truck bearing license plate number MDC-578 exited
the XYZ parking lot at 8:30 pm on the day in question. This tangible item could also be
made available to analysts investigating this matter.
min min
i
Source’s Source’s
competence credibility
An objective observer is one who bases a belief on the sensory evidence instead of desires
or expectations. Finally, if the source did base a belief on sensory evidence, how good was
this evidence? This involves information about the source's relevant sensory capabilities
and the conditions under which a relevant observation was made.
As indicated in Figure 4.24, there are several types of testimonial evidence. If the source
does not hedge or equivocate about what the source observed (i.e., the source reports that
he or she is certain that the event did occur), then we have unequivocal testimonial
evidence. If, however, the source hedges or equivocate in any way (e.g., "I'm fairly sure
that E occurred"), then we have equivocal testimonial evidence. The first question we
would ask a source of unequivocal testimonial evidence is: How did you obtain information
about what you have just reported? It seems that this source has three possible answers to
this question. The first answer is, "I made a direct observation myself.” In this case, we have
unequivocal testimonial evidence based upon direct observation. The second possible
answer is, "I did not observe this event myself but heard about its occurrence (or
nonoccurrence) from another person." Here we have a case of second hand or hearsay
evidence, called unequivocal testimonial evidence obtained at second hand. A third answer
is possible: "I did not observe event E myself nor did I hear about it from another source.
But I did observe events C and D and inferred from them that event E definitely occurred."
This is called testimonial evidence based on opinion, and it requires some very difficult
questions. The first concerns the source's credibility as far as his or her observation of
events C and D; the second involves our examination of whether we ourselves would infer
E based on events C and D. This matter involves our assessment of the source's reasoning
ability. It might well be the case that we do not question this source's credibility in
observing events C and D, but we question the conclusion that the source has drawn
from his or her observations that event E occurred. We would also question the certainty
with which the source has reported the opinion that E occurred. Despite the source’s
conclusion that “event E definitely occurred," and because of many sources of uncertainty,
we should consider that testimonial evidence based on opinion is a type of equivocal
testimonial evidence.
There are two other types of equivocal testimonial evidence. The first we call completely
equivocal testimonial evidence. Asked whether event E occurred or did not, our source
says, "I don't know," or, "I can't remember."
But there is another way a source of HUMINT can equivocate: The source can provide
probabilistically equivocal testimonial evidence in various ways: "I'm 60 percent sure that
event E happened”; or "I'm fairly sure that E occurred”; or, "It is very likely that
E occurred." We could look upon this particular probabilistic equivocation as an assess-
ment by the source of the source’s own observational sensitivity.
The right side of Figure 4.25 shows how a Disciple-EBR agent assesses the believability
of an item of testimonial evidence based upon direct observation Ek* by a source, as the
minimum of the source’s competence and credibility. The source’s competence is
assessed as the minimum of the source’s access and understandability, while the source’s
credibility is assessed as the minimum of the source’s veracity, objectivity, and observa-
tional sensitivity.
Here are some examples involving testimonial evidence from human sources that is not
hedged or qualified in any away.
Evidence item E014-Grace in Table 4.3 (p. 126) is Grace’s testimony that no one at the
XYZ Company had checked out the canister for work on any project. Grace states this
unequivocally. You should also note that she has given negative evidence saying the
lOMoARcPSD|35902250
cesium-137 was not being used by the XYZ Company. This negative evidence is very
important, because it strengthens our inference that the cesium-137 canister was stolen.
E006-Clyde in Table 4.3 is unequivocal testimonial evidence. It represents positive evidence.
Here are some examples involving testimonial evidence given by human sources who
equivocate or hedge in what they tell us.
Consider the evidence item E005-Ralph in Table 2.4 (p. 60). Here Ralph hedges a bit by
saying that the lock on the hazardous materials storage area appears to have been forced.
He cannot say for sure that the lock had been forced, so he hedges in what he tells us.
In new evidence regarding the dirty bomb example, suppose we have a source code-
named “Yasmin.” She tells us that she knew a man in Saudi Arabia named Omar al-
Massari. Yasmin says she is “quite sure” that Omar spent two years “somewhere” in
Afghanistan “sometime” in the years 1998 to 2000.
believability of Ei*
min
Source’s
believability
authenticity of Ei* accuracy of Ei*
min
Source’s Source’s
competence credibility
min
min
Source’s Source’s
access understandability
Source’s
observational sensitivity
important concept from the field of law, where a chain of custody refers to the persons or
devices having access to the original evidence, the time at which they had such access, and
what they did to the original evidence when they had access to it. These chains of custody
add three major sources of uncertainty for intelligence analysts to consider, all of which
are associated with the persons in the chains of custody, whose competence and credibil-
ity need to be considered. The first and most important question involves authenticity:
Is the evidence received by the analyst exactly what the initial evidence said, and is it
complete? The other questions involve assessing the reliability and accuracy of the
processes used to produce the evidence if it is tangible in nature or also used to take
various actions on the evidence in a chain of custody, whether the evidence is tangible or
testimonial. As an illustration, consider the situation from Figure 4.27. We have an item of
testimonial HUMINT coming from a foreign national whose code name is “Wallflower,”
who does not speak English. Wallflower gives his report to the case officer Bob. This
report is recorded by Bob and then translated by Husam. Then Wallflower’s translated
report is transmitted to the report’s officer Marsha, who edits it and transmits it to the
analyst Clyde, who evaluates it.
Figure 4.28 shows how a Disciple-EBR agent may determine the believability of the
evidence received by the analyst. A more detailed discussion is provided in Schum
et al. (2009).
The case officer might have intentionally overlooked details in his recording of Wall-
flower’s report. Thus, as shown at the bottom of Figure 4.28, the believability of the
recorded testimony of Wallflower is the minimum between the believability of Wallflower
and the believability of the recording. Then Husam, the translator, may have intentionally
altered or deleted parts of this report. Thus, the believability of the translated recording is
the minimum between the believability of the recorded testimony and the believability of
the translation by Husam. Then Marsha, the report’s officer, might have altered or deleted
parts of the translated report of Wallflower’s testimony in her editing of it, and so on.
Wallflower’s
received Wallflower’s
reported Wallflower’s
testimony about
Wallflower’s
testimony about translated
Emir Z. in English.
recorded
Emir Z. in English. testimony about
testimony about Wallflower’s
Emir Z. in English.
Emir Z. in Farsi. testimony about
recording
translation
Clyde editing
transmission
Wallflower
• Competence
• Credibility
Bob
Husam A.
Marsha • Competence ce
e
Marsha
• Competence • Competence
• Competence • Veracity Sony Recorder
ce
e
• Credibility • Credibility
• Veracity SN 247
• Fidelity
• Fidelity • Reliability
• Reliability
• Security
min
Believability of Believability of
min
Believability of Believability of
min
Believability of Believability of
min
Believability of Believability of
The result of these actions is that the analyst receiving this evidence almost certainly
did not receive an authentic and complete account of it, nor did he receive a good account
of its reliability and accuracy. What Clyde received was the transmitted, edited, translated,
and recorded testimony of Wallflower. Although the information to make such an analysis
may not be available, the analyst should adjust the confidence in his conclusion in
recognition of these uncertainties.
This case study, which continues the analysis from Section 4.5 with the analysis of the
hypothesis, “The cesium-137 canister is used in a project without being checked-out from
the XYZ warehouse,” has two main objectives:
In Section 4.6, we have presented how you can define an item of evidence, and Figure 4.23
(p. 132) shows the definition of E001-US-top-oil-importer with type evidence. You can
specify the type by clicking on the [CHANGE] button. Figure 4.29, for instance, shows the
definition of E014-Grace. After you click on the [CHANGE] button, the agent displays the
various evidence types from the right panel. You just need to click on the [SELECT] button
following the correct type, which in this case is unequivocal testimonial evidence based upon
direct observation.
Once you have selected the type of E014-Grace, the agent displays it after the label Type
and asks for its source, which is Grace (see Figure 4.30).
As shown in Figure 4.30, we have also indicated that this item of evidence disfavors the
hypothesis “The missing cesium-137 canister is used in a project at the XYZ company.” As a
result, the agent introduced it into the analysis tree and generated a more detailed analysis
of its believability, which is shown in Figure 4.31.
You can now perform a more detailed believability analysis, as illustrated in Figure 4.32,
where we have assessed the competence, veracity, objectivity, and observational sensitiv-
ity of Grace, and the agent has automatically determined her believability.
In this case study, you will practice the preceding operations. You will first select the
hypothesis, “The cesium-137 canister is used in a project without being checked out from
lOMoARcPSD|35902250
Figure 4.31. Decomposition of the believability assessment for an item of testimonial evidence.
the XYZ warehouse.” Then you will browse its analysis to see how it is reduced to simpler
hypotheses that need to be assessed based on the evidence. After that, you will represent a
new item of evidence, will associate it with the hypothesis to which it is relevant, assess its
relevance, evaluate its believability by assessing its credentials, and browse the resulting
analysis tree.
lOMoARcPSD|35902250
An important feature of the Disciple-EBR agent is that it allows you to perform analyses at
different levels of detail. What this means is that a hypothesis may be reduced to many levels
of subhypotheses or just a few levels that are then assessed based on relevant evidence. The
same applies to assessing the believability of evidence. You may directly assess it, as was
illustrated in Figure 4.14 (p. 127), where the believability of E005-Ralph was assessed as very
likely. But if an item of evidence has an important influence on the analysis, then you may
wish to perform a deeper believability analysis, as was illustrated in Figure 4.32, where the
user assessed lower-level believability credentials. The user could have drilled even deeper
to assess the source’s access and understandability instead of his or her competence.
It may also happen that you do not have the time or the evidence to assess a
subhypothesis, in which case you may make various assumptions with respect to its
probability. Consider, for example, the analysis from the case study in Section 4.5, partially
shown in Figure 4.15 (p. 128) and the four subhypotheses of the top-level hypothesis. The
first three of these subhypotheses have been analyzed as discussed in the previous
sections. However, for the last subhypothesis, you have made the following assumption:
It is certain that the MDC-578 truck left with the cesium-137 canister.
Assumptions are distinguished from system-computed assessments by the fact that the
assumed probabilities have a yellow background.
lOMoARcPSD|35902250
You may provide justifications for the assumptions made. You may also experiment
with various what-if scenarios, where you make different assumptions to determine their
influence on the final result of the analysis.
Thus the agent gives you the flexibility of performing the analysis that makes the best
use of your time constraints and available evidence.
The Disciple-EBR shell includes a customized modeling assistant to model the hypoth-
esis analysis process. The following two case studies demonstrate its use.
The objective of this case study is to learn how to use Disciple-EBR to model the analysis of
a hypothesis. More specifically, you will learn how to:
This case study will guide you through the process of defining and analyzing a hypothesis
by using, as an example, the following hypothesis: “CS580 is a potential course for Mike
Rice.” You will first define the reduction tree shown in Figure 4.33. Then you will formalize
it and specify the synthesis functions.
Start Disciple-EBR, select the case study knowledge base “05-Modeling-Learning/Scen,”
and proceed as indicated in the instructions from the bottom of the opened window.