Tim Menzies
Dept. of Software Development, Monash University,
Melbourne, Australia
timm@cse.unsw.edu.au
-
Simon Goss
Air Operations Division, Aeronautical & Maritime
Research Laboratory, Melbourne, Australia.
simon.goss@dsto.defence.gov.au
Tue Oct 1 16:38:50 EST 1996
This file is also available in postscript form.
Informal vague causal diagrams such as Figure 1 are a common technique for illustrating and sharing intuitions about a domain. Normally, such diagrams are viewed as precursors to other modeling techniques. That is, the standard approach is to continue the analysis process till the vagueness goes away.
Figure 1: An economic vague causal diagram; from [5].
Here, we take another approach and explore what we can do with these vague diagrams without requiring precise analysis. Our approach is useful for both the acquisition and maintenance of knowledge. Standard software engineering methodologies are typically prescriptions on how to develop some initial system. Here we formalise the process of handling an existing, possibly poorly understood, system. Our goal is the structuring and the optimisation of the resources available during the maintenance phase. In the usual case, software maintenance is the process of understanding someone else's idiosyncratic code which may or may not be documented. An extreme software maintenance problem occurs when handling legacy systems where the maintainer may not even have access to the source code.
Vague models must be processed differently to standard software systems. The standard software design activity cycle shown in Figure 2 [11]. Logical design comprises a brainstorming elicitation stage followed by a representation stage that systematises the concepts found during elicitation. To make the representations exhibit behaviour, they must be transformed into a lower-level representation which can be operationalised. During this transformation stage, implementation-dependent details will contort the representation. Once operationalised, these concepts can be executed and tested. Testing can be divided into verification (i.e. was the system built right?) and validation (i.e. was the right system built?). Feedback from the testing process can improve the physical and logical designs.
Figure 2: The design activity cycle;
adapted from [11]
In modeling of vague domains (see §2.1), systems development spends more time in testing and correction than initial logical design. Hence, development should move through the logical design stage as quickly as possible since most of the specification will be found in the testing stage. Vague modeling cannot produce correct models. Rather, it can only produce models that contain portions that are useful for known purposes.
Our preferred test engine is the HT4 abductive validation engine (summarised in §2.2). Previously, we have discussed HT4 and its applications to KBS inference and verification [18,17]. This article maps these low-level discussions into the high-level processes referred to in Figure 2. At its core, HT4 maps a theory down into a directed graph expressing the connections and invariants between entities in the theory. This low-level data structure, we will argue, is a unifying framework for operationalising most of the vague modeling process.
Figure 3 shows our proposed design activity cycle for vague domains. Elicitation (§3.1) lets us articulate and represent (§3.2) our theory; a library of desired behaviour (called the observations); and a success criteria for judging between competing inferences. In the transformation stage (§3.3), we:
Figure 3: Vague Modeling: the Details
At execution time, each pair is processed.
Portions of the dependency graph are extracted that connect members of
back to members of
. Recall that we are executing
vague models. Vague models are necessarily incomplete;
i.e. they are under-specified. When executing such under-specified
models, guesses have to be made and mutually exclusive guesses must be
managed separately. The extracted portions of the dependency graph
must be divided into internally consistent worlds.
The generation and assessment of the these worlds is a framework for
many expert system inference tasks, including validation. Good
theories can generate worlds that contain most of
; i.e. they
can explain a majority of the known behaviour.
Vague models contain inaccuracies. Therefore, vague modeling is only possible when the theory can be tested. That is, the computational properties of the validation engine constrain the modeling process; i.e. if we can't test it then we shouldn't model it. We will note below where the limits of the test device constrain the design process.
In order to explicate our argument, we begin with some background material about example vague domains (§2.1) and abduction (§2.2).
Elsewhere [15], we have characterised vague domains as: (i) poorly measured, i.e. known data from that domain is insufficient to confirm or deny that some inferred state is valid; (ii) hypothetical, i.e. the domain lacks an authoritative oracle (e.g. a person or a large library of known behaviour) that can declare knowledge to be ``right'' or ``wrong''; (iii) Indeterminate, i.e. inferencing over a knowledge base could generate numerous, mutually exclusive, outcomes.
The approached proposed here was developed by the first
author for a neuroendocrinological domain (§2.1.1) which we
would describe as the analysis of perturbations of an equilibrium
system.
Our current goal is to test the generality of that previous
work approach in the domain of black-box comprehension (§2.1.2).
We would describe the air traffic control simulators found in this
second, black-box comprehension domain as spatio-temporal reasoning
over non-equilibrium systems. Note that our work in this second
black-box comprehension domain is still in its preliminary stages.
Neuroendocrinology is the study of the interaction of glands and nerves. Obtaining values for certain chemicals within the body is not as simple as, say, attaching a volt meter to an electric circuit:
Measurement in this domain can therefore be an expensive process and
not all entities are fully measured. For example, in
studies [17,8] of a sketch of a model of glucose
regulation [23], data was collected from six journal
articles. This sketch was represented as a compartmental
model . In all, none of the flow constants
were known and only 39% of the compartments were measured and not all
at the same time interval.
Operations Research makes extensive use of large complex
simulation models. These can be the result of many person-years of
development and incorporate modules obtained from external sources. It
is common that such a system may be obtained from a third party.
Typically, some group called derives some model
representing their initial understanding of a problem (e.g.
modeling the performance of a missile system). This model is
operationalised in some third generation language to become
.
Perhaps an attempt is made to document
in a manual
which may be incomplete.
and
are then shipped to another
site where a second team (
) tries to understand them.
Conceptually,
builds
, a model representing the
local understanding of
and the incomplete
.
Figure 4: Commissioning Remote Models
The effort required to use with confidence can be non-trivial,
especially when object code is supplied
without source code or access to the author.
Current
practice is for
to be documented in an incomplete manner
(e.g. some procedural manual advising parametric sensitivity and
constants relating to the local physical and operational environment).
can be regarded as a black-box model that
must convert
(with some support from
) into a gray-box model
. Once validated,
would be used for planning, prediction, and
optimisation studies. However, first it is essential that we can
validate
the vague model
since
local conditions may invalidate
.
For example, the Australian Defence Forces use aircraft in
configurations that are different to how they are used overseas.
Certain parameters representing important domain knowledge
are stored in compiled numerical matrices and are
inaccessible to
. For example,
these parameters may (i) be based on experimental
data from tests in other climates or (ii) contain
certain tacit assumptions about
aircraft operation.
Prior to relying on
,
would like to validate this model under local
conditions.
Elsewhere [18], we have given a precise overview of abduction.
Here, we offer an approximate characterisation of abduction as the
search for consistent subsets of some background theory that are
relevant for achieving some goal. If multiple subsets can be
extracted, then a preference operator (which we call and
Bylander et. al. [2] call the plausibility operator pl)
must be applied to select the preferred subset(s). Abduction is
different to deduction. In deductive inference, all consequences of
known literals must be found by the inference procedure. In
abduction, the inference engine can ignore parts of the background
theory if they are not relevant for the current goal.
Abduction is a very general process. Abductive frameworks for various KBS tasks have been defined by numerous researchers; e.g. frame-based reasoning [20], Bayesian reasoning [2], diagnosis [6], explanation [13], and default reasoning [21]. We have also argued that abduction is also useful for planning, monitoring, verification, prediction, classification, qualitative reasoning [18], and validation [17]. The reason for this generality is that the processing of subset extraction is a synonym of known general frameworks for expert systems; i.e. Breuker's [1] components of solutions and Clancey's [4] situation-specific models [18]. For example, the formal difference between heuristic classification and heuristic construction [3] is the extraction of one subset or multiple conflicting subsets, respectively.
Figure: A portion of Figure 1.
In order to present a detailed example of abduction, we will run an example through our HT4 abductive inference engine [18]. A portion of Figure 1 is shown in Figure 5. The edge symbols of Figure 5 are explained in Figure 6.
Figure: Edge symbols of Figure 1 and
Figure 5.
In the case where ={foriegnSalesUp, domesticSalesDown}
and
= {investorConfidenceUp, inflationDown, wagesDown},
we can generate the following proofs
for
:
Figure: Worlds from Figure 5.
Which is the world? We have argued elsewhere [18]
that, depending on the choice of
operators, a variety of KBS
tasks can be emulated.
operators can be defined for one of
three levels: vertex, proof, and world. When extending a proof from a
particular node, a vertex-level
could elect to use
edges with the highest numerical rating. This could be used to emulate
a MYCIN-style certainty factor traversal. A proof-level
could execute when some proofs or partial proofs are known;
e.g. beam search during planning. A worlds-level
could execute when the worlds are known. Minimal-fault diagnosis is a
world-level
assessment operator that returns world(s) which
maximise the number of
puts obtained while minimising the
number of
puts to those world.
returns
the worlds(s) with the greatest overlap with known user profiles;
i.e. return the worlds a particular user is liable to understand. Our
preferred validation approach is a worlds-level
that returns
the worlds that offer explanations for the most number of
puts. This
would therefore favour
(which explains 3 members of
) over
(which only explains 2 members of
).
The computational limits of the test engine constrains what can be
represented in vague modeling.
Elsewhere, we have explored the practicality of HT4 for medium
to large theories [17]. In our mutation study, 257
theories were artificially generated by adding random vertices and
edges to the and-or graph from a theory called Smythe '89. These
were run using 5970 experimental comparisons. In the changing N
mutation study, the fanout was kept constant and the size of the model
was increased. Model size N was measured by the number of vertices
in the ground graph. For example, foriegnSales
contributes three vertices to the ground graph: {foriegnSalesUp,
foriegnSalesSteady, foriegnSalesDown}. For that study, a ``give up''
time of 840 seconds was built into HT4. HT4 did not
terminate for
in under that ``give up''time. We
conclude that the ``knee'' in the exponential runtime curve kicks-in
at around 800 literals. These result came from a Smalltalk V
implementation on a Macintosh Powerbook 170. A port to ``C'' on a
Sparc Station is underway.
The changing fanout mutation study examined the practicality of
HT4 for models of varying fanout (fanout is the number of edges
that touch each vertex
; i.e.
).
In that study, the Smythe
'89 theory size was kept constant, but edges were added at random to
produce new graphs of larger fanouts. Six models were used of sizes
.
At low fanouts, many
behaviours were inexplicable. However, after a fanout of 4.4, most
behaviours were explicable. Further, after a fanout of 6.8, nearly all
the behaviours were explicable [17].
Figure 8: Parameters from fielded expert systems.
From [19].
In practice, how restrictive is a limit of 850 vertices and
? The only reliable data we could
find on fielded expert systems is shown in
Figure 8 [19].
Figure 8 suggests that a practical inference
engine must work at least for the range
and
. We therefore conclude
that vague modeling is practical for the size of expert systems seen
in current practice and for nearly the range of fanouts seen in
fielded expert systems. However, after a certain level of
inter-connectivity, a theory is able to reproduce any input/output
pairs. An inference procedure that condones any behaviour at all from
a theory is not a useful inference procedure. After the point where
%
covered approaches 100%, vague modeling
becomes impractical.
Above, we have defined vague domains and the abductive framework that is the backbone of the process. We have explored the computational limits of validation in that framework. We can show where the testing process constrains the various stages of the design process; i.e. elicitation (§3.1), representation (§3.2), transformation (§3.3), verification (§3.4), and validation (§3.5).
The elicitation stage is an exploration of the domain. The goal of this stage is to make a first pass at discovering (i) list of terms for ``things'' in this domain; (ii) a rough idea of how these ``things'' influence each other; (iii) what ``things'' are incompatible (the invariants); (iv) a set of scenarios recording what behaviour your models should be able to reproduce.
In the elicitation, domain experts are encouraged to sketch out vague causal diagrams like Figure 1. Note that these models will be under-specified and can generate conflicts. At this stage, experts should not try and resolve these conflicts. Rather, their goal is to generate the set of all possible explanations that are acceptable to them. The process of extracting relevant and consistent subsets will be handled later.
Once drawn, the next stage in vague modeling is to walk through the
diagrams with scenarios; i.e. examples of how things impact the
system. These scenarios will later become our puts and
our
puts.
Tacit in the vague diagrams is invariant knowledge. For example, having said that inflation can go up or down is the statement that we can't believe that inflation can go up and down simultaneously. More generally, these diagrams can include invariants between any pairs of states. Invariants can be absolute or gradual:
When working with sensor data, it may be useful to introduce
summary nodes that generalise certain data. For example,
prolongedExposure to sunlight and sunLamps might both connect
to tanned; e.g.
prolongedExposure tanned and
sunLamps
tanned.
Vague causal diagrams can be generated from many pre-existing sources such as:
Domains typically have pre-existing special constructs
which may be useful in elicitation of the vague causal diagrams.
For example,
neuroendocrinology using
compartmental models.
In compartmental models,
compartments are connected by ``pipes'' and each pipe has a flow
control
box. These flow rate through these flow controllers
can be changed by other compartments; e.g. glucose (a compartment)
can increase the glucagonProduction (a flow controller on a
pipe going into glucagon);
i.e.
glucose
glucagonProduction and
glucagonProduction
glucagon.
Details of the test engine constrain the elicitation stage. In particular, we must restrict the size of vague models. In practice, this means avoiding low-level detail and removing time.
Avoiding low-level detail: Vague modeling is not a technique for recording of low-level detail since the HT4 inference engine has problems handling very big models (recall §2.2.2). Hence, if models get too big, we ne continuous variables with up, down, Steady; i.e. one of three states representing the sign of the change in those variables. Note that this introduces some indeterminacy into the model. Recall that in Figure 1, if foriegnSales and domesticSales go up and down respectively, then we have contradictory possible values for companyProfits ( up OR down). In vague modeling, we defer the resolution of these conflicts till the testing phase. In the meantime, recall that if HT4 generates multiple worlds for companyProfitsUp or companyProfitsDown, then these contradictory assumptions will be stored in multiple worlds.
Removing time:
In practice, for the current work, we plan to build our models without time. In the neuroendocrinological domain, time is removed by only using scenarios without time series data. This restriction still provides us with many data points. Much of the experimentation in neuroendocrinology is performed on separate groups of rats. Conclusions are made by comparing measurements in some control group to measurements in some treated group. In such scenarios, there is no sense of time flowing from the control group to the treated group.
In the black-box comprehension domain, time is harder to remove. Many of the systems in the target domain are large FORTRAN or C++ simulations which explicitly model time-series events. Our current approach in this domain is to develop an elicitation framework aimed at declarative representation. Time and position are often inter-changeable indexes in the navigation strategies of mariners and pilots. Traditional AI planning (e.g. STRIPS) avoids explicit representation of time by creating ordered lists representing priorities in schedules. We conjecture that pilot knowledge expressed in the this prioritised list framework, need not mention time. Rather, our rules will be reactions to some immediate situation rather than a reflection of some between-situations plan.
The representation stage fills in the details sketched in the elicitation stage. The goal of this stage is creation of (i) a theory (§3.2.1); (ii) some observations containing the scenarios that the theory should be able to reproduce (§3.2.2); and the success criteria (§3.2.3).
Smythe '87 [22] is a theory proposing connections between serum adrenocorticotropin ( acth), serum corticosterone ( cortico), and neuro-noradrenergic activity ( nna). Nna was measured as the ratio of noradrenaline to its post-cursor, 3,4-dihydroxphenyl-ethethyleneglycol. This theory is shown as the vague causal diagram of Figure 10. The temp vertex was introduced to model the user's statement that dex has the same effect as cortico.
Figure 10: The Smythe '87 theory
The Smythe '87 theory is shown in Figure 10 in HT4 syntax. This theory starts with a name = 'a string' statement and is followed by two object statements. The object statements describe the events and the measures of the diagram. Events (denoted objects(e)) are like booleans while measures (denoted objects(m)) are for continuous variables. The body, which follows the head, contains the influence statements. The influence statements are pairs of measures connected by one of the topoi symbols (see Figure 6).
The Smythe '87 paper contains values comparing these values in various
experiments: (i) control i.e. no treatments; (ii) dex i.e.
an injection of dexamethasone at 100 ; (iii)
coldSwim i.e. a two minute swim in a bath of ice cold water; and
(iv) coldSwim, dex i.e. both a coldSwim and an injection
of dex. A sample of experimental results from Smythe '87 is
shown as the observations of Figure 11 (in HT4 syntax).
The experiments
are shown in the columns and the measures taken in the different
experiments are shown in the rows.
Figure: The observations for the Smythe '87 model
of Figure 10.
Our scenarios are comparisons between two experiments.
When we compare (e.g.) [coldSwim] to
[dex, coldSwim], we say that the input () to scenario was
change(coldSwim)=arrived, change(dex)=arrived
and the output (
) of the scenario
was that acth=down,
cortico=down, nna=up.
The scenario
that our theory has to try and reproduce is
change(coldSwim)=arrived and change(dex)=arrived
leads to nna=up;
that is
cold swims and injections
of steroids increases stress levels in a rat.
The success criteria will be mapped into the HT4
operator. We use the success criteria to select preferred inferences.
Success criteria extract summaries of the inferences, and
then cull some inferences based on those summaries.
The exact
operator is domain dependent. For
neuroendocrinology, we have used a
operator that returns the
world(s) that cover the most
puts. Note that this answers the
following question: ``can a theory of X reproduce known
behaviour of X?''. Elsewhere (§3.5) we describe
experiments were this was applied as a validation criteria which
successful found faults in published models of neuroendocrinology.
For other domains, (e.g. machine learning) we may chose to (i) return
the smaller worlds; (ii) the world(s) that make the fewest
assumptions; that (iii) use the least controversial assumptions; (iv)
the ``cheapest'' worlds (assuming that each edge is associated with
some notion of the cost of its inclusion); (v) etc.
The details of what can be represented are restricted by what can be handled at the transformation stage.
The physical design is the bridge between the logical design and an executable version which we can test. Specifically, the goals of the transformation stage is to be able to:
Two styles of model-compilers are planned for the black-box comprehension domain. The first style is a generic partial-evaluator of rules with variables into the network of connections between literals in the rule. This partial-evaluator will need knowledge of the range of each field of each fact that can be asserted. The second style will be an attempt to ``shoe-horn'' the rule representation into the representation used in the neuroendocrinological domain. A hand inspection of the rules will be used to draw vague causal diagrams which will be mapped into the neuroendocrinological theory.
With the neuroendocrinological model compiler, if a vertex can be in one of N states, then one vertex is added for each such state. If an edge condones connections between two states of adjacent vertices, then an edge is added between those states. For example, power ++ lights is expanded into Figure 12.
Figure 12: Light is directly proportional to power.
All the vertices of Figure 12 are or vertices; i.e. belief in these vertices requires a belief in only one of its parents. The model compiler can add the disabling effects of rats via and vertices; i.e. vertices which we can only believe if we believe all their parents. For example, if we know that the rats in the basement disable the connection between the power and the lights (when they bump the cable), then we say if not rats then power ++ lights which expands into Figure 13.
Figure 13: The lights/power relationship works when rats are absent.
At this point, the neuroendocrinological model compiler gets intricate. A conjunction of competing upstream influences can combine to explain a steady vertex. Also, changes to an object's value downstream of an abler link can be explained in terms of changes to the abler. Returning to our rats, in the case of power not rising (but on) and the rats being present, the lights are dark. Now consider the same situation, but the rats suddenly disappearing. The lights going up can now be explained in terms of a change in the rat population. If we take these into account, we get Figure 14. In Figure 14 we include for the first time the incompatible vertices (see the double-headed arrows marked with a cross).
Figure 14: The final rats model.
For more details on the neuroendocrinology model compiler, see [16].
The data compiler converts tables like Figure 11 into
and
sets of vertices in the dependency graph. For example, in
the scenario [] to [dex] in Figure 11,
= dex=arrived, coldSwim=absent and
=
nna=up, cortico=down, acth=down.
There are five other scenarios:
<,
> pairs: [] to [coldSwim],
[] to [dex,coldSwim], [dex] to [coldSwim],
[dex] to [dex,coldSwim], and [coldSwim] to
[dex,coldSwim]. Therefore, we can
run six scenarios through the Smythe '87 model
of Figure 10.
Smythe '87 is a symmetrical model.
In symmetrical models, a comparison
from to
is the same as
to
; i.e. after studying [dex] to [coldSwim],
we do not need to study [coldSwim] to
[dex]. Non-symmetrical domains are those that include asymmetric
influences. Direct and inverse are symmetric influences
while ablers
are asymmetric. In asymmetric
domains, therefore, twice as many scenarios must be studied.
Verification is an exploration of the internal syntactic structure of a program. Preece, Shinghal and Batarekh [19] (hereafter, PSB) define a taxonomy of structural ``anomalies'' in rule-based expert systems (see Figure 15) and argue that a variety of verification tools target different subsets of these anomalies (perhaps using different terminology).
Some of these anomalies require meta-knowledge about knowledge base literals:
A rule is PSB-redundant if the same final hypotheses are reachable if that rule was removed. An unusable redundant rule has some impossible premise. A knowledge base is PSB-deficient if a consistent subset of askables leads to zero final hypotheses. A PSB-duplicate redundant rule has a premise that is a subset of another rule premise. PSB define duplicate rules for the propositional case and subsumed redundant rules for first- order case (where instantiations have to be made to rule premise variables prior to testing for subsets). PSB define ambivalence as the case where, given a consistent subset of askables, a rule-base cannot infer some final hypotheses.
PSB stress that the entries in their taxonomy of KBS anomalies may not be true errors. For example, the dependency network from a rule-base may show a circularity anomaly between literals. However, this may not be a true error. Such circularities occur in (e.g.) user input routines that only terminate after the user has supplied valid input. For this reason, the "errors" detected by internal testing are called anomalies, not faults. Internal test anomalies are used as pointers into the system which direct the developer to areas that require a second glance.
In vague domains, the anomalies shown in italics of Figure 15 are to be expected. For example:
However, a graph-theoretic analysis of the dependency
graph can detect many of the other PSB anomalies:
Other useful verification studies are:
Validation is an assessment of the program based on external semantic criteria. For each model generated from the logical design, we find:
In the case of competing models, we reject the models that explain
significantly less behaviour than their competitors. Returning to the
Smythe '87 theory (Figure 10), in the scenario
[coldSwim] to [dex, coldSwim], we can infer from
Figure 11 that ={acthDown,corticoDown, nnaUp}
and
= { change(dex)=arrived, coldSwim=present}. In
this comparison nnaUp can't be explained since their exists no
proof from nnaUp to
which does not violate the proof
invariants. Another error can be found in the comparison [dex]
to [dex,coldSwim]. In this comparison
=
{change(coldSwim)=arrived, dex=present} and
=
{acthSteady, corticoUp, nnaUp} and only nnaUp can be
explained. What is interesting here is that the Smythe '87 theory had
been published in an international refereed journal [22]
and its validation errors were found by a detailed examination of the
data published to support it! Further, when these errors were shown to
the author of the theory, he found them novel and exciting; i.e. they
were significant errors.
Smythe '87 is a small theory. The Smythe '89 study explores how well abductive validation scales up to medium-sized models. Smythe '89 [23] is a theory of human glucose regulation. It contains 27 possible inputs and 53 measurable entities which partially evaluated into an and-or graph with 554 vertices and 1257 edges. Smythe '89 is a review paper that summaries a range of papers from the field of neuroendocrinology. Those papers offer 870 experimental comparisons with between 1 to 4 inputs and 1 to 10 outputs.
Smythe '89 was originally studied by QMOD, the technical precursor to HT4. That study found that 32% of the observations were inexplicable. QMOD could not explain studies or handle multiple causes. These restrictions implied that it could only handle of 24 possible comparisons. Even with these restrictions, QMOD found several errors in Smythe '89 that were novel and exciting to Smythe himself [8]. Like the Smythe '87 study, these errors had not been detected previously by international peer review.
When HT4 ran over the full 870 comparisons, it found more errors than
QMOD. Only 150 of the comparisons could explain 100% of their
puts. On average, 45% of the
s in those comparisons were
inexplicable. The level of critique offered by QMOD and HT4 is
surprisingly high. This is both a disturbing and exciting finding.
It is disturbing in the sense that if the very first large-scale
medical theory analysed by abductive validation contains significant
numbers of errors, then it raises doubts as to the accuracy of theories in
general. This result is exciting in the sense that the level of
critique is so high. Abductive validation promises to be a powerful
technique for theory review.
Vague modeling makes little use of two of the dominant themes in KA:
ontological engineering [10] and reusable problem solving
strategies (e.g. KADS [24]). While an
appropriate ontology may be useful for ``kick-starting'' the
elicitation stage, they may not be productivity tools for the rest of
the design activity life cycle in vague domains. Further, it is not
clear what (e.g.) KADS offers for the testing phase of a KBS. Our
framework easily supports validation and verification. While KADS has
a ``verification'' inference model, in its current form it cannot be applied
reflectively on the KADS models themselves. Also, this KADS
``verification model'' makes no mention to assumption space
management; a technical issue that is central to our architecture and
validation process.
Vague modeling is the process of making what sense we can of half-formed ideas in a poorly measured domain. In this paper we have explored the design activity cycle of vague modeling based around vague causal diagrams and abductive inference. We have contrasted vague modeling with conventional modeling. In conventional modeling, most of the design is generated in the analysis stage. In vague modeling, most of the model is found during testing. Further, the limits of the computational device used in the testing phase restricts the modeling process back at the logical design phase. Hence, the test engine should be built first since, in vague domains, what can't be tested shouldn't be modeled. HT4 imposes two restrictions: the need to avoid low-level detail and the need for declarative representations that avoid time. Models that satisfy these restrictions can be verified and validated.
Originally, the approach described here was developed as a general approach, using the neuroendocrinology domain as an example. Our current task is to test the generality of this approach in the black-box comprehension domain. While the work done to date is preliminary, it has served to highlighted certain restrictions with the software developed for the neuroendocrinology domain. For example, we need better tools for building of model and data compilers (see §3.3.1 and §3.3.2). While ultimately these must be customised to a particular domain, model compilers for rules and frames should be added to the current system.
Some of the Menzies papers can be found at
http:// www.sd.monash.edu.au/ timm/pub/
docs/papersonly.html.
Vague Models and Their Implications for the KBS Design Cycle
This document was generated using the LaTeX2HTML translator Version 95.1 (Fri Jan 20 1995) Copyright © 1993, 1994, Nikos Drakos, Computer Based Learning Unit, University of Leeds.
The command line arguments were:
latex2html -split 0 words.tex.
The translation was initiated by Timothy Menzies on Tue Oct 1 16:37:53 EST 1996