Retrieval, Reuse, Revision, and Retention in Case-Based Reasoning
Retrieval, Reuse, Revision, and Retention in Case-Based Reasoning
Artificial Intelligence Research Institute, CSIC, Campus UAB, 08193 Bellaterra, Spain
e-mail: mantaras@[Link]
2
School of Computing and Information Engineering, University of Ulster, Coleraine BT52 1SA, Northern
Ireland, UK
e-mail: [Link]@[Link]
3
Department of Computer Science, University College Cork, Ireland
e-mail: [Link]@[Link]
4
Computer Science Department, Indiana University, Lindley Hall 215, 150 S. Woodlawn Avenue,
Bloomington, IN 47405, USA
e-mail: leake@[Link]
5
School of Computer Science and Informatics, University College Dublin, Belfield, Dublin 4, Ireland
email: [Link]@[Link]
6
School of Computing, The Robert Gordon University, Aberdeen AB25 1HG, Scotland, UK
e-mail: [Link]@[Link]
7
AI-Lab, Swiss Federal Institute of Technology, CH-1015 Lausanne, Switzerland
e-mail : [Link]@[Link]
8
School of Information Technologies, University of Sydney, Australia
e-mail : marym@[Link]
9
BBN Technologies, Cambridge, MA 02138, USA
e-mail: mcox@[Link]
10
EECS Department, Northwestern University, Evanston, IL 60208, USA
e-mail : forbus@[Link]
11
School of Computer Science and Informatics, University College Dublin, Belfield, Dublin 4, Ireland.
e-mail : [Link]@ucd..ie
12
Department of Computer and Information Science, Norwegian University of Science and Technology,
Trondheim, Norway
[Link]@[Link]
13
Department of Computer Science, University of Auckland, Auckland, New Zealand
e-mail: ian@[Link]
Abstract
Case-based reasoning (CBR) is an approach to problem solving that emphasizes the role of prior
experience during future problem solving (i.e., new problems are solved by reusing and if
necessary adapting the solutions to similar problems that were solved in the past). It has enjoyed
considerable success in a wide variety of problem solving tasks and domains. Following a brief
overview of the traditional problem-solving cycle in CBR, we examine the cognitive science
foundations of CBR and its relationship to analogical reasoning. We then review a representative
selection of CBR research in the past few decades on aspects of retrieval, reuse, revision, and
retention.
1 Introduction
While much of the inspiration for the study of case-based reasoning (CBR) came from cognitive
science research on human memory (e.g., Schank, 1982), the resulting methodology has been
shown to be useful in a wide range of applications (e.g., Watson, 1997; Aha, 1998; Bergmann,
2002). Unlike most problem solving methodologies in artificial intelligence (AI), CBR is
memory based, thus reflecting human use of remembered problems and solutions as a starting
point for new problem solving. An observation on which problem solving is based in CBR,
namely that similar problems have similar solutions (e.g., Leake & Wilson, 1999), has been
shown to hold in expectation for simple scenarios (Faltings, 1997b), and is empirically validated
in many real-world domains.
Solving a problem by CBR involves obtaining a problem description, measuring the similarity
of the current problem to previous problems stored in a case base (or memory) with their known
solutions, retrieving one or more similar cases, and attempting to reuse the solution of one of the
retrieved cases, possibly after adapting it to account for differences in problem descriptions. The
solution proposed by the system is then evaluated (e.g., by being applied to the initial problem or
assessed by a domain expert). Following revision of the proposed solution if required in light of
its evaluation, the problem description and its solution can then be retained as a new case, and
the system has learned to solve a new problem.
Figure 1 shows Aamodt & Plaza's (1994) classic model of the problem solving cycle in CBR.
The individual tasks in the CBR cycle (i.e., retrieve, reuse, revise, and retain) have come to be
known as the "4 REs". Because of the pivotal role of retrieval in the CBR cycle, a considerable
amount of research has focused on retrieval and similarity assessment. As illustrated in Figure 2,
Leake (1996b) expresses the role of similarity through the concepts of retrieval and adaptation
distances. Also captured in Leakes diagram is the relationship between problem and solution
spaces in CBR.
Problem
RETRIEVE
RETAIN
PRIOR
CASES
Similar
Cases
REUSE
CASE BASE
Updated
Solution
REVISE
Proposed
Solution
Figure 1 The CBR cycle. Adapted from (Aamodt & Plaza, 1994).
In Figure 2, the retrieval distance R increases as the similarity between the input problem
description and a stored problem description decreases (i.e., lower similarity means greater
distance). A common assumption in CBR is that the retrieval distance R is commensurate with A,
the adaptation distance (or effort). However, several authors have questioned this assumption and
its implication that the most similar case is the easiest to adapt (e.g., Smyth & Keane, 1998). This
is an issue that we will return to in our discussion of retrieval methods in Section 3.
.Figure 2 Relationship between problem and solution spaces in CBR. Adapted from (Leake, 1996b).
As will be clear from the literature review we present in this article, aspects of reuse and
retention, and to a lesser extent revision, have also attracted significant research interest. Several
books examine fundamental aspects of CBR and present case studies of research and applications
(e.g., Riesbeck & Schank, 1989; Kolodner, 1993; Leake, 1996a; Watson, 1997; Lenz et al., 1998;
Bergmann, 2002). However, our aim in this article is to present a concise summary of research
that focuses on the problem-solving cycle in CBR, including some of the most recent advances.
In Section 2, we briefly examine the cognitive science foundations of CBR and the relationship
between CBR and analogical reasoning. In Sections 3, 4, and 5, we review a representative
selection of CBR research on aspects of retrieval, reuse and revision, and retention. Our
conclusions are presented in Section 6.
The study of CBR has been strongly influenced by research in cognitive science. A major current
underlying early CBR research was the study of human story understanding (Schank & Abelson,
including reasoning and learning. Psychological studies have shed light on some basic properties
of human analogical reasoning. For example, there is ample evidence that, for people, retrieval is
heavily influenced by surface properties more than by deep similarities, unlike most CBR
systems. Yet when people are given the analogues, they find the comparisons easy (e.g., Gick &
Holyoak, 1980; Keane, 1988; Gentner et al., 1993). The centrality of relational information in
human similarity judgments is now reasonably well-established (e.g., Gentner et al., 2001;
Markman & Gentner, 2001; Kokinov & French, 2003). This reinforces the practice of using
structured representations that was common in early CBR, but it also suggests that CBR and
machine learning systems that use feature vectors are unlikely to be good models of human
cognition.
Psychological theories of analogical processes have led to computational models that can be
used for building CBR systems. For example, the Structure-Mapping Engine (SME)
(Falkenhainer et al., 1986; Forbus et al., 1994a) is based on Gentners (1983) structure-mapping
theory. SME has been used as a cognitive modeling tool to account for some existing
psychological findings, and several predictions made based on SME have been subsequently
confirmed in psychological experiments (Forbus, 2001). SME has also been used in a variety of
performance-oriented systems, ranging from a case-based tutor for engineering thermodynamics
(Forbus, 2001) to sketch understanding systems (Forbus et al., 2003). The Incremental Analogy
Machine (Keane & Brayshaw, 1988; Keane et al., 1994) was first to explore incremental
mapping of analogies, crucial for extended problem solving. Some computational models have
focused on how such computations can be implemented in neural architectures, such as Hummel
& Holyoaks (1997) LISA model and Larkey & Loves (2003) CAB model. Other computational
models explore how statistical association models can be combined with structural models (e.g.,
Ramscar & Yarlett, 2003).
One of the major differences in approach between CBR and analogy research is their focus on
generality. In analogy research, processes like matching and retrieval are typically assumed to be
broadly general cognitive processes, operating universally (or nearly so) over peoples mental
representations. In contrast, CBR often focuses on creating a system to perform a specific task
well on existing computing hardware, and generality is often traded off for efficiency or other
performance measures, with an emphasis on content theories that reflect the knowledge required
for particular task domains. Domain-specific matchers, retrieval systems, and even similarity
metrics are fair game. This can lead to controversies between the two communities. This
dissociation between similarity-based reminding and analogical inference was greeted with great
skepticism and shock in the CBR community because it contradicted a common assumption that
human memory retrieval relied on extensive indexing, using abstract principles. Forbus et al.s
(1994b) MAC/FAC model captures this dissociation by postulating a first stage of retrieval that is
non-structural, a cheap filter that generally lets highly similar items through, followed by a much
more constrained structural match stage using SME. If the MAC/FAC model is correct, then
CBR index-based retrieval schemes would best be viewed as good engineering tools, rather than
as cognitive models. This is still an open question.
CBR also differs from analogy research in its treatment of adaptation. Case adaptation is a
major issue for CBR, and must often be addressed in CBR applications. Consequently, it is
possible that the experience of CBR in this area will suggest fruitful questions to consider in the
context of analogy. Studies focusing on integrating adaptability concerns with other CBR
processes include Smyth & Keane's (1998) adaptation-guided retrieval and Leake et al.s (1997)
use of adaptation cases to reuse prior adaptations and predict adaptability.
2.2 Discussion
How closely should CBR systems mirror what people do? It can be argued that organizing CBR
systems to operate differently from people in some ways might make them more useful. Just as
eyeglasses and cars help us see and move farther than we can unaided, carefully designed CBR
systems could help us retrieve more relevant memories more often, and help us to work through
problems that we could not solve unaided due to working memory limitations. On the other hand,
no CBR system approaches the amount of knowledge and experience that people accumulate, nor
has any CBR system ever operated over the breadth of problems that people can solve. The
processes of human cognition may well hold the key to creating such capabilities.
Retrieval in CBR
An important step in the CBR cycle (Figure 1) is the retrieval of previous cases that can be used
to solve the target problem. As will be clear from our discussions in Sections 3.1 and 3.2,
improving retrieval performance through more effective approaches to similarity assessment has
been the focus of a considerable amount of research. However, several authors have questioned
the basic assumption on which similarity-based retrieval is based, namely that the most similar
cases are most useful for solving the target problem. In Section 3.3, we examine alternative
approaches to retrieval that have been motivated by a growing awareness of the limitations of
similarity-based retrieval, particularly in light of new application requirements.
Domeshek, 1992; Leake, 1992) as a way to avoid the need for computationally expensive
structure mapping and case matching procedures.
3.1.1 Assessment of surface similarity
In approaches to retrieval based on surface features, the similarity of each case to the target
problem, typically represented as a real number in [0, 1], is computed according to a given
similarity measure. Usually the retrieved cases are the k most similar to the target problem, an
approach often referred to as k nearest neighbor retrieval or simply k-NN (e.g., Cover & Hart,
1967). Alternatively, the retrieved cases may be those whose similarity to the target problem
exceeds a predefined threshold.
There are many ways of measuring similarity and different approaches are appropriate for
different case representations. For example, it is common practice for each case to be represented
as a simple feature vector (or set of attribute-value pairs). With this representation, a local
similarity measure is usually defined for each attribute and a global similarity measure is
computed as a weighted average of the local similarities. The weights assigned to case attributes
allow them to have varying degrees of importance and may be selected by a domain expert or
user, or as we shall see in Section 3.2, determined by an adaptive learning process.
A CBR system can guarantee that it retrieves the k cases that are maximally similar to the
target problem by computing the similarity of the target problem to every case in memory.
However, sequentially processing all cases in memory has complexity O(n), where n is the
number of cases. This may not be an acceptable overhead if n is very large. One approach to
reducing retrieval time, as in the pioneering work of Stanfill & Waltz (1986), involves the use of
massively parallel computers. While the requirement for expensive hardware is an obvious
drawback, the approach still guarantees finding the maximally similar cases by performing an
exhaustive memory search. Stanfill & Waltz describe the implementation of a memory-based
reasoning algorithm on a fine-grained SIMD parallel machine. Their Connection Machine
performs a highly parallel search for similar cases and was applied to the problem of
pronouncing English words using a case memory containing thousands of examples of correctlypronounced words.
Another approach to reducing retrieval time relies on the organization of cases in memory. For
example, Wess et al. (1993) propose an approach to retrieval in which organization of the case
memory is based on similarities between cases. A binary tree called a k-d tree is used to split the
case memory into groups of cases in such a way that each group contains cases that are similar to
each other according to a given similarity measure. To ensure that the most similar cases are
retrieved, the retrieval algorithm computes similarity bounds to determine which groups of cases
should be considered first.
Smyth & McKenna (1999a; 2001b) propose an alternative model of case retrieval that is
informed by the availability of an explicit model of case-base competence (Smyth & McKenna,
1998; 2001a). The so-called footprint-based retrieval algorithm is a two-stage retrieval approach
that searches two distinct populations of cases. First, it involves the search of a small subset of
so-called footprint cases which have been identified as providing a covering set for the entire
case base (i.e., can solve the same set of problems). They are drawn from the key competence
groups that exist within the case base as made available by the competence model. The first stage
of retrieval identifies the footprint case that is most similar to the target problem as the reference
case, and the second stage of retrieval searches another small subset of cases that are related to
this reference case. This related set is chosen because its cases either cover (solve) the reference
case or because they can be covered by a reference case. The final case chosen for retrieval is the
related set case that is most similar to the target problem. The approach has produced significant
retrieval efficiency benefits by searching only a small fraction of the cases in the case base, while
at the same time guaranteeing the selection of near optimal cases. Footprint-based retrieval is
related to Shaafs (1996) fish and shrink strategy in which cases are linked according to specific
aspect similarities. The latter approach assumes that if a case does not fit a query then this will
reduce the likely usefulness of its neighbors. This allows for the efficient elimination of many
cases during retrieval.
Simoudis & Miller (1990) argue that retrieval based only on surface similarity may not be
sufficiently discriminating when applied to large case memories, and needs to be combined with
other techniques to reduce the number of cases considered for adaptation. They present an
approach called validated retrieval that can dramatically reduce the number of potentially
relevant cases. Retrieval based on surface similarity is combined in the approach with validation
of the retrieved cases to determine if they are applicable to the target problem. Associated with
each case in memory is a validation procedure consisting of a set of domain-specific tests and
their results for that case. To validate a retrieved case, its associated tests are applied to the target
problem. The retrieved case is deemed relevant to the target problem only if all the tests give the
same results for the target problem as they do for the retrieved case.
The validation phase that follows the initial retrieval of cases in validated retrieval resembles
the justification phase in CASEY, a CBR system for medical diagnosis (Koton, 1988). The goal
of CASEYs justification component is to determine whether the causal explanation of a
retrieved case applies to the target problem. Often this enables CASEY to avoid invoking its
causal model when creating an explanation for the target problem. Other systems that combine
retrieval based on surface similarity with an additional filter to improve retrieval performance
include CHEF (Hammond, 1986), SWALE (Kass et al., 1986), KRITIK (Goel &
Chandrasekaran, 1989), and PROTOS (Porter et al., 1990).
3.1.2 Assessment of structural similarity
Though computationally expensive because it relies on extensive use of domain knowledge,
retrieval based on structural similarity has the advantage that more relevant cases may be
retrieved. As in Forbus et al.s (1994b) MAC/FAC model (Section 2.1), one way of mitigating
the extra cost is to combine surface and structural similarity in a two-stage retrieval process.
Experiments based on human assessment of similarities and analogies have confirmed that both
surface and structural similarity assessment are necessary for sound retrieval (Forbus et al.,
1994b). Inspired by previous work by Gentner & Forbus (1991), Brner (1993) proposes an
approach to retrieval in which fast retrieval of candidate cases based on their surface similarity to
the target problem is followed by a more expensive assessment of their structural similarity. She
defines structural similarity as the most specific graph structure that the target problem has in
common with a stored case, and a set of transformation rules, given as background knowledge,
needed to determine this common structure.
10
the practical efficiency of the approach, they introduce a subgraph matching algorithm that
works on a compact version of the case memory in which subgraphs that are common to multiple
cases are stored only once. In a similar vein, Champin & Solnon (2003) propose a similarity
measure, based on graph editing operations within a modification of Tverskys (1977) contrast
model, to compare cases represented by labeled graphs where vertices and edges can have more
than one label. To account for the intractability of this representation, they propose a heuristic
greedy algorithm.
Arcos & Lpez de Mntaras (1997) describe a retrieval mechanism called perspectives for
structured case representations. Cases and degrees of similarity are represented as feature terms,
which are equivalent to first-order terms and can also be viewed as directed acyclic graphs
labeled by features and values (Plaza, 1995). Their knowledge-intensive approach to retrieval
uses a subsumption mechanism between the feature terms to obtain an order relation between
case descriptions on the basis of a set of user-defined relevant aspects of the target problem. The
system is implemented in an object-oriented language (Arcos, 1997) based on feature terms and
has been applied to the problem of synthesizing expressive music (Arcos & Lpez de Mntaras,
2001; Lpez de Mntaras & Arcos, 2002).
Emde & Wettschereck (1996) propose an alternative way of measuring the similarity of firstorder terms. They also present a generalization of a propositional instance-based learner
(distance-weighted k-NN) to first order representations. Issues addressed in the approach, which
the authors refer to as relational instance-based learning (RIBL), include the generation of cases
from the knowledge base, assessment of similarity between arbitrarily complex cases, and
estimation of the relevance of predicates and attributes. Their empirical results suggest that RIBL
can achieve high levels of classification accuracy in a variety of domains.
3.1.3 Similarity frameworks
With so many ways of measuring similarity, it is not surprising that some researchers view
similarity in a general way, independent of any specific algorithm. For example, Richter (1992)
discusses the notion of similarity in the context of a formal mathematical framework. He
describes approaches to modeling similarities with increasing complexity and informativeness.
These range from simple predicates (least informative) to relations and functions (most
informative). He also discusses general forms of distance functions and similarity measures,
including a generalization of Tverskys (1977) contrast model, which uses a set-theoretic
approach to express the similarity between objects as a linear combination of their numbers of
matching and mismatching features. One limitation of Tverskys model is that all features are
assumed to be equally important, whereas Richters generalization allows different weights to be
assigned to features. However, Richter emphasizes that to allow for changes in the problemsolving environment, the parameters of a similarity measure should be the result of an adaptive
learning process, an idea we explore further in Section 3.2.
Osborne & Bridge (1996) present another general framework that distinguishes, in particular,
between ordinal and cardinal similarity measures. Ordinal measures use a description of the
target problem to induce a partial ordering over the cases in the case memory. No information
about the degree of similarity is given; the cases are merely ordered, with the implication that
cases higher in the ordering should be retrieved prior to any that are lower in the ordering. On the
11
other hand, cardinal measures are functions that score the cases, returning real numbers to denote
degrees of similarity. Osborne & Bridge present a set of operators that allows the flexible and
systematic construction of both ordinal and cardinal similarity measures. In later work (e.g.,
Osborne & Bridge, 1997), the framework is generalized to accommodate similarity measures in
which the degree of similarity is denoted by any value drawn from an ordered set. With this
extension, the framework accounts for similarity measures in which the degree of similarity is
denoted by common subgraphs (Brner, 1993) or feature terms (e.g., Jantke, 1994; Plaza, 1995).
12
classification accuracy, Wilke & Bergmann (1996) argue that decision costs should also be
considered in many applications. Experimental results support the hypothesis that classification
based on weights learned using cost optimization leads to lower decision costs than classification
based on weights learned by accuracy optimization.
Improving the adaptability of retrieved cases can also be the subject of an adaptive learning
process. In case-based planning, for example, Muoz-Avila & Hllen (1996) extend the footprinted similarity metric used in Prodigy/Analogy (Veloso, 1994; Veloso & Carbonell, 1994) by
incorporating feature weights in a new metric which counts the weights of relevant features that
match features in the target problem. A feature is considered relevant to a planning goal with
respect to a solution if it contributes to achieving the goal in the solution. The authors also
present an algorithm for analyzing the performance of retrieved cases to identify features whose
weights need to be recomputed. The algorithm provides a bridge between the new similarity
metric and a feature weighting model based on incremental optimizers. Experimental results
show that integrating the proposed similarity metric and analysis algorithm in the feature
weighting model improves the adaptability of the retrieved cases by converging to best weights
over a period of multiple problem-solving episodes.
Many CBR applications rely on domain knowledge encoded in the similarity measures used
by the system to guide the retrieval of relevant cases. Such a knowledge-intensive approach to
similarity assessment typically relies on knowledge acquired from a domain expert. Stahl &
Gabel (2003) and Gabel & Stahl (2004) investigate the use of machine learning techniques to
reduce the knowledge-acquisition overheads associated with the construction and maintenance of
domain-specific similarity measures. Feedback about the quality of retrieval results provided by a
domain expert is used in the approach to guide the automatic refinement of similarity measures.
13
case may be impossible to adapt, for example if adaptation knowledge is incomplete, as is often
the case in weak-theory domains that are commonly targeted by CBR. To address this issue, they
introduce the notion of adaptation-guided retrieval in which the adaptation requirements of cases
are explicitly assessed during retrieval by means of domain-specific adaptation knowledge. In
contrast to traditional approaches that relied on heuristics to predict the ease with which a given
case could be adapted, adaptation-guided retrieval combines local and global measures of
adaptability to ensure that the most adaptable case is always selected. In this way, it bridges the
gap between retrieval (and similarity knowledge) and reuse (and adaptation knowledge).
Empirical results show that the approach can significantly reduce adaptation failures and
adaptation costs by performing preliminary adaptation work during retrieval. Leake et al. (1997)
propose a case-based approach to this problem, predicting adaptation effort based on prior
adaptation experiences.
3.3.2 Diversity-conscious retrieval
In CBR recommender systems, descriptions of available products are stored in a product case
base and retrieved in response to a query describing the user's requirements. An important
advantage of similarity-based retrieval in this context is that if there is no case that exactly
matches the user's requirements, she can be shown the cases that are most similar to her query.
However, one problem is that the most similar cases are often very similar to each other, with the
result that the user is offered a limited choice (Smyth & McClave, 2001). In other words, the
recommended cases may be lacking in diversity. To address this issue, retrieval algorithms have
recently been introduced that combine measures of similarity and diversity in the retrieval
process to achieve a better balance between these often conflicting characteristics (e.g., Smyth &
McClave, 2001; McSherry, 2002; McGinty & Smyth, 2003). For example, Smyth & McClave
(2001) propose an approach to retrieval that incrementally selects a diverse set of cases from a
larger set of similarity-ordered cases. Experimental results have shown that major gains in
diversity can often be achieved at the expense of relatively small reductions in similarity.
3.3.3 Compromise-driven retrieval
McSherry (2003b; 2004a) proposes a compromise-driven approach to retrieval in recommender
systems inspired by the observation that the cases that are most similar to the users query are
often not sufficiently representative of compromises (i.e., unsatisfied requirements) that the user
may be prepared to accept. A basic assumption in similarity-based retrieval is that a given case
(or product) is more acceptable than another if it is more similar to the users query.
Compromise-driven retrieval is based on the weaker assumption that a given case is more
acceptable than another if it is more similar to the users query and it involves a subset of the
compromises that the other case involves. As well as being less likely to be contradicted by user
behavior, this weaker assumption provides the basis of a more principled approach to deciding
which cases are included in the retrieval set than arbitrarily limiting the number of retrieved cases
as in k-NN. For example, no case is included in the retrieval set if there is a more similar case
that involves a subset of the compromises it involves.
Though not relying on an explicit measure of diversity in the retrieval process, compromisedriven retrieval shares with other approaches to enhancing diversity (e.g., Smyth & McClave,
14
2001; McSherry, 2002) the goal of offering users a better choice of alternatives. Also, it
guarantees that the retrieval set provides full coverage of the available cases; that is, for any case
that is not included in the retrieval set, one of the recommended cases is at least as good in terms
of its similarity to the users query and the compromises it involves.
3.3.4 Order-based retrieval
Order-based retrieval is another approach with particular application to recommender systems
(Bridge & Ferguson, 2002a). Rather than scoring the cases, it offers an expressive query
language for defining and combining ordering relations, and the result of query evaluation is to
partially order the cases in the case base. The query language supports queries that naturally
combine preferred values with other preference information such as maximum values, minimum
values, and values that the user would prefer not to consider. As in compromise-driven retrieval
(McSherry, 2003b), there is no need for an explicit measure of recommendation diversity
because the set of retrieved cases is inherently diverse (Bridge & Ferguson, 2002b).
3.3.5 Explanation-oriented retrieval
It is often important for CBR systems to explain their reasoning and to justify their suggestions
or solutions (e.g., Rissland, et al. 1984; Ashley & Aleven, 1992; Leake, 1996b; McSherry,
2001b; Cunningham et al., 2003; Doyle et al., 2004; McSherry, 2004b; Leake & McSherry,
2005). Explanations serve many different goals, such as teaching the user about the domain or
explaining the relevance of a question the user is asked (Leake, 1991; 1992; Srmo & Cassens,
2004; Srmo et al., 2005). For example, McSherry (2003a; 2005) proposes a conversational CBR
approach to product recommendation in which the system can explain why a question is being
asked in terms of its ability to discriminate between competing cases. Explaining retrieval
failures is another potential role of explanation in CBR recommender systems (e.g., McSherry,
2004c).
More commonly, the goal is to explain how the system reached its conclusions. In
applications such as classification and diagnosis, an attractive feature of CBR is the ability to
explain the predicted outcome by showing the user one or more of the target problems nearest
neighbors. As noted by Leake (1996b), ... the results of CBR systems are based on actual prior
cases that can be presented to the user to provide compelling support for the system's
conclusions. Such explanations are known as precedent-based explanations and have long been
a feature of case-based models of legal argumentation (e.g., Ashley, 1991; Branting, 1991;
Rissland & Skalak, 1991). An empirical study by Cunningham et al. (2003) has shown that they
are often more compelling than alternative forms of explanation. However, several authors have
questioned the effectiveness of precedent-based explanations in which the user is simply shown
the case that is most similar to the target problem.
For example, McSherry (2004b) argues that such explanations are often less informative than
might be expected, and should ideally be supported by an analysis of the pros and cons of the
proposed solution. Case-based legal argumentation (Ashley, 1989, 1991; Ashley & Aleven,
1992; 1997; Aleven, 2003) broadens the notion of case similarity to include other considerations
(e.g., noteworthy distinctions among cases, the existence of counterexamples) with a view to
explaining and distinguishing the strengths and weaknesses of relevant cases, providing examples
15
The reuse process in the CBR cycle is responsible for proposing a solution for a new problem
from the solutions in the retrieved cases. In the 4 REs of Aamodt & Plazas (1994) classic CBR
cycle (Figure 1), reuse appears second, after retrieve, and is followed by revise and retain.
Reusing a retrieved case can be as easy as returning the retrieved solution, unchanged, as the
proposed solution for the new problem. This is often appropriate for classification tasks, where
each solution (or class) is likely to be represented frequently in the case base, and therefore the
most similar retrieved case, if sufficiently similar, is likely to contain an appropriate solution. But
reuse becomes more difficult if there are significant differences between the new problem and the
retrieved cases problem. In these circumstances the retrieved solution may need to be adapted to
account for these important differences. Medical decision making is one domain in which
adaptation is commonly required.
Adaptation becomes particularly relevant when CBR is used for constructive problem-solving
tasks such as design, configuration, and planning. For such tasks it is unlikely that each solution
(design, configuration, or plan) will be represented in the case base. Thus the retrieved solution is
simply an initial solution and any differences between the new problem and the retrieved cases
problem are likely to alter the retrieved solution.
Adaptation methods differ in complexity with respect to two dimensions: what is changed in
the retrieved solution, and how the change is achieved. Substitution adaptation simply
reinstantiates some part(s) of the retrieved solution, whereas transformation adaptation alters the
structure of the solution (Kolodner, 1993). Adaptation is commonly achieved by altering the
retrieved solution directly, but the more complex generative adaptation replays the method of
deriving the retrieved solution on the new problem. These three adaptation methods will be used
to structure this section. The contributions we discuss are different approaches to adaptation for
reuse (i.e., adaptation during solution formulation). Adaptation can also be used when feedback
about a proposed solution indicates that a repair is needed; this is part of the revise stage in the
CBR cycle.
Hammond (1990) describes the reuse of recipes in CHEF, a menu-planning system.
Substitution adaptation is used to substitute ingredients in the retrieved recipe to match the menu
requirements (e.g., when a recipe containing beef and broccoli is retrieved for a meal requiring
16
chicken and snow peas, the meat component is replaced by chicken and the vegetable component
is substituted by snow peas). Transformation adaptation may also be needed to amend the
proposed recipe further by adding or removing steps in the recipe that result from any ingredient
substitutions (e.g., for chicken, rather than beef, a new skinning step should be added). Further
transformations may occur at the revise stage where critics analyze the failure of a recipe and
repair strategies are applied to the proposed recipe to add or remove steps in the failed recipe.
CHEFs learning of critics introduced the topic of case-based planning and many of its themes
(e.g., indexing, use of cases in memory, failure-driven learning).
SWALE (Schank & Leake, 1989) is a case-based explanation system for story understanding
that reuses old explanations by applying substitution adaptation to amend the actor, their role or
the action in the retrieved explanation (Kass, 1989). Transformation adaptation may again be
needed to add or remove components in the current explanation resulting from these
substitutions.
Dj Vu (Smyth & Keane, 1995a; 1998) is a CBR system for the automated design of plantcontrol software. It builds on some of the ideas proposed in CHEF (Hammond, 1990) by utilizing
transformation adaptation knowledge in the form of general adaptation strategies and more
specialized adaptation specialists. An important and novel contribution of Dj Vu is its
representation of complex plant-control software designs as hierarchies of related cases and its
adoption of a hierarchical model of case retrieval and reuse. For example, complex target
problems lead to the retrieval and adaptation of the solutions to abstract cases, the elements of
which in turn lead to the retrieval and adaptation of more detailed sub-cases. In this way, the
solution transformation is performed by a combination of problem decomposition, sub-case
adaptation, and solution re-integration. A unique feature of Dj Vu, as discussed in Section 3.3,
concerns its ability to leverage existing adaptation knowledge during retrieval to evaluate the
adaptability of cases. This alleviates some adaptation problems by guaranteeing the retrieval of
cases that can be adapted easily.
Model-based adaptation is a popular approach to transformation adaptation in which causal
reasoning is integrated with CBR. Kotons (1988) CASEY is an early example of model-based
adaptation, in a medical diagnosis CBR system that utilizes domain independent repair strategies
to adapt the retrieved explanation to account for differences between the symptoms of new and
retrieved patients. KRITIK (Goel & Chandrasekaran, 1989) relies on model-based transformation
adaptation to reuse designs for physical devices. Model-based reasoning creates a causal
explanation for the new design by transforming the explanation of the retrieved one. Faltings
(1997a) CADRE also applies model-based reasoning in case-based design. Reuse involves the
combination of retrieved design cases and the transformation adaptation of the retrieved design.
This approach was evaluated in two design prototypes: CADRE for architectural design and
FAMING (Faltings & Sun, 1996) for mechanism design.
Evolutionary methods have also been explored for adaptation, in the context of architectural
design (Gmez de Silva Garza & Maher, 2000). The retrieved designs become the initial
population for a genetic algorithm and mutation and crossover operators are used to generate new
designs for the population. Mutation is a substitution adaptation that randomly alters parts of one
design to produce a new design. Crossover is a transformation adaptation that can alter the
structure of the design. It generates two new designs from two parent designs by interchanging
17
parts of the design in each parent. The genetic algorithms fitness function evaluates the designs
by calculating how well they match the design requirements. The design with the highest match
to the requirements is the new design.
Purvis & Pu (1995) present adaptation as a constraint satisfaction problem. The design cases
are represented as constraint satisfaction problems, where the design requirements are the
constraints and the design is the solution. The retrieved designs are adapted by applying a
minimum conflicts heuristic to guide the repair of the design to match the new design
requirements.
The work discussed so far in this section has been devoted to substitution and transformation
adaptation. Based on studies of several CBR systems that use adaptation, Fuchs & Mille (1999)
propose a knowledge-level task model for substitution and transformation adaptation processes.
Their reuse task is composed of copy and adapt subtasks. The adapt subtask comprises selecting
a problem difference, modifying the solution, and finally verifying the solution. The modify task
can remove or substitute/add part of the solution, and finally the substitute/add task searches for a
suitable replacement by using additional cases, applying a heuristic, or accessing domain
knowledge such as explanations, abstractions or specializations.
Generative adaptation differs from substitution and transformation adaptation in that it does
not adapt the retrieved solution directly, but instead derives the new solution by replaying the
method used to derive the retrieved solution. Generative adaptation may result in a reinstantiation
of parts of the retrieved solution, like substitution, or in a transformation that alters the structure
of the solution. Prodigy/Analogy (Veloso, 1994; Veloso & Carbonell, 1994), a general purpose
planning system, applies derivational replay to recompute a replacement for a faulty element of
the retrieved solution by recalling how the element was computed and replaying the computation
for the new problem. Derivational replay is a variant of derivational analogy (Carbonell, 1986) in
which the complete solution is recomputed. In this work an analogy between the new and
retrieved problems is used to adapt the method of deriving the solution.
Although CBR systems avoid reasoning from first principles by remembering and reusing past
solutions, substitution and transformation adaptation of retrieved solutions is often achieved by
reasoning about how the problem differences should be reflected in the adaptation to the
proposed solution. Therefore the acquisition of adaptation knowledge can require a substantial
knowledge engineering effort. The difficulty of acquiring adaptation knowledge was identified in
early CBR research but, until recently, relatively little effort has been devoted to automating the
acquisition of adaptation knowledge. Leake et al.s (1995) DIAL system for disaster response
planning builds up its adaptation knowledge as it applies case-based planning. The adaptation
knowledge it learns is a set of adaptation cases that capture the steps in successful manual plan
adaptations. DIAL applies a mixed-initiative adaptation process: if an adaptation case matches
the current adaptation need it is reused by a CBR process, otherwise DIAL attempts to apply a
general rule to revise the plan (e.g., add or remove a step), but as a final option it resorts to
manual adaptation. This last option offers the opportunity to acquire a new adaptation case when
the manual adaptation generates a successful plan.
Several systems exploit the knowledge already captured in the cases as a source of adaptation
knowledge. McSherry (1999) reuses pairs of cases from the case base that contain the same
differences as those found between the new problem and the retrieved case. The solution
18
difference from the pair of cases is replayed on the retrieved case. Rather than reusing differences
directly from cases, Hanney & Keane (1997) use the case base as a source of case pairs that are
used as training data to learn rule-based adaptation knowledge that generalizes the adaptations
represented in the case pairs. Wilke et al.s (1997) knowledge-light learning provides a
framework for such approaches, in which knowledge contained elsewhere in the CBR system,
like the case base, is used to learn or improve the adaptation knowledge. Jarmulak et al. (2001)
use a set of adaptation cases created from the original case-base as the knowledge source of a
case-based adaptation system. Further work has applied different learning methods to assemble
an ensemble of rule-based adaptation experts learned from these adaptation cases (Wiratunga et
al., 2002).
Retention in CBR
In the classic review paper by Aamodt & Plaza (1994), retention is presented as the final step in
the CBR cycle, in which the product of the most recent problem-solving episode is incorporated
into the systems knowledge. To a great extent this has traditionally translated into a variety of
approaches for recording the product of problem solving as a new case that can be added to the
case base. Of course, there are various issues concerning how best to learn a new case and
different systems record different types of information in their cases. Most, for example, simply
record the target problem specification and the final solution, with the implicit assumption that
the outcome was successful. For example, when CBR is integrated with a generative problemsolving system for speed-up learning, the success of the systems solutions may be guaranteed
(e.g., Veloso, 1994; Veloso & Carbonell, 1994). When outcomes are less reliable or when the
criteria for success are more complex, case representations must include additional information
on the outcome of the solution, which may also include fine-grained information on how well the
solution addressed many dimensions of the systems goals (e.g., Goel et al., 1991). Another
question is what to store concerning the solution itself. Many systems store only the solution, but
others record a much deeper representation of the problem solving process that brought about the
particular solution. In Veloso & Carbonells work, for example, derivational traces are stored in
cases. These rich knowledge structures describe precisely how a given solution was derived,
providing a trace of the decision-making processes that led to a particular solution.
In general, the modern view of retention accommodates a much broader perspective of what it
means for a CBR system to learn from its problem solving experience, a view that is largely a
response to certain critical issues that have arisen during the practical application of CBR
systems in complex problem solving scenarios. In this section we will review this body of work,
highlighting many of the critical issues associated with open-ended case learning policies and
how these issues have been resolved by novel approaches to case-base optimization. Moreover,
we will argue that these issues have served as an important catalyst for research in the area of
case-base maintenance and the maintenance of other aspects of a CBR system, which
accommodates a broader perspective on case learning and retention.
19
20
examples from a set of instances. For example, Hart's condensed nearest neighbor (CNN)
approach (Hart, 1968) is an early attempt to eliminate redundancy from a collection of instances,
which in turn led to a family of related redundancy detection methods (e.g., Gates, 1972; Tomek,
1976b). Wilson (1972) adopts a complementary stance, focusing on the elimination of noise
rather than redundancy from training data. For example, Wilson's edited nearest-neighbor (ENN),
and Tomek's (1976a) repeated edited nearest neighbor (RENN) techniques represent the genesis
of a variety of noise reduction techniques that remain as valuable coping strategies even today.
In the 1990's, with the revival of lazy learning techniques and a renewed interest in nearest
neighbor methods, researchers from the machine learning community once again focused their
attention on ways to eliminate harmful data from training sets. Algorithms proposed by Aha et al.
(1991) and Wilson & Martinez (1997) are good examples of a new breed of machine learning
inspired coping strategies that ultimately came to have a significant influence on the CBR
community. Markovitch & Scott (1993) propose a unifying framework for the systematic
discussion of all of the various strategies for coping with harmful knowledge in general, and the
utility problem in particular. Their framework is based on different types of filters for eliminating
harmful knowledge at various stages in the problem solving cycle. One approach that is
especially relevant in CBR is to simply delete harmful cases from the case base so that they
cannot actively contribute to ongoing problem solving costs deletion policies in CBR
correspond to selective retention filters in the Markovitch & Scott framework. Surprisingly
enough, in many speed-up learners even the apparently naive random deletion of knowledge
items (to maintain the knowledge base to some predefined size) works quite well for optimizing
efficiency. Even though random deletion removes both useful and redundant items it can equal
the success of more sophisticated methods (Markovitch & Scott, 1993). More sophisticated
deletion policies have been developed and are guided by some assessment of the utility of
individual knowledge items. For example, Minton (1990) uses a utility metric that takes into
account the cost of including the item in the set of candidates to consider (match cost) and the
expected savings offered by the item (average savings multiplied by its application frequency) to
deliver even greater protection against the damaging effects of the utility problem.
Unfortunately it soon became clear that the same type of coping strategies would not translate
directly over to case-based reasoners. The problem stems from the fact that many case-based
reasoners are not simply using case knowledge as a form of speed-up knowledge. Instead, cases
are often a primary source of problem solving knowledge. Without cases, certain problems
cannot be solved and thus the act of deleting cases may irrevocably reduce the competence of the
system to solve new problems; CBR systems may not be able to reconstruct deleted cases from
an internal domain model. To address this problem, Smyth & Keane (1995b) proposed the use of
a competence model to evaluate the contributions of individual cases to problem solving
competence. In particular, they developed methods for categorizing cases according to their
competence characteristics with a view to guiding the selection of cases for deletion. These
categories facilitate the preservation of key cases (called pivotal cases) that might otherwise be
deleted, in favor of deleting less critical cases whose loss is expected to least harm system
competence. Competence-guided case deletion provides a safe way to eliminate cases from a
growing case base to stave off the harmful effects of the utility problem while at the same time
protecting against reductions in competence.
21
Later work brought the introduction of a more fine-grained model of case competence (Smyth
& McKenna, 1998) as a pre-cursor to a variety of related retention models and other forms of
case-base editing (e.g., McKenna & Smyth, 2000a; 2000b; Smyth & McKenna, 2000). For
example, as an alternative to case deletion, Smyth & McKenna (1999b) use their competence
model to develop a competence-guided case addition algorithm. In related work, Zhu & Yang
(1999) describe a case addition algorithm that has the added advantage of providing a guaranteed
lower bound on resulting competence. Leake & Wilson (2000) highlight the importance of
considering both competence and performance during case-base optimization. They argue the
need for more fine-grained performance metrics with which to guide the maintenance of a case
base and show how one such metric can help to guide case-base editing in a way that gives due
consideration to competing factors such as case-base size, coverage, and adaptation performance.
Over the past few years there has been a broad range of research addressing these key issues
of case deletion, addition, and case-based editing in general. Further discussion is beyond the
scope of this article but the interested reader is referred to work by Surma & Tyburcy (1998), Lei
et al. (1999), Portinale & Torasso (2001), Yang & Zhu (2001), Salam & Golobardes (2002),
Wiratunga et al. (2003), and Woon et al. (2003).
22
23
therefore inextricably tied to many related issues for managing the multiple forms of knowledge
within CBR systems and adapting CBR systems to the needs of the environments in which they
function.
Figure 3 An extension of the classical 4-stage CBR model to emphasize the importance of maintenance in
overall system performance, illustrating the setup, initialization, application and maintenance phases of the
SIAM methodology for maintaining CBR systems. Adapted from (Iglezakis et al., 2004).
Maintenance strategies can also be used to assist the case author during the early stages of
case acquisition. For example, Ferrario & Smyth (2001) describe a distributed approach to case
authoring in which a community of authors contribute to the validation of new case knowledge.
McSherry (2000; 2001a) also focuses on the case acquisition task, and presents CaseMaker, a
system that performs background reasoning on behalf of the case author while new cases are
being added, in order to help the user determine the best cases to add in light of their competence
contributions. The system uses its evaluations of the contributions of potential cases to suggest
cases to add to the case library. McKenna & Smyth (2001) propose an approach to providing
authoring support that attempts to identify competence holes within an evolving case base. They
demonstrate how their model of competence (Smyth & McKenna, 1998; 2001a) can be used to
prioritise gaps in case knowledge and, like McSherry (2000; 2001a), propose a technique for
automatically suggesting the type of cases that an author might want to consider to fill these gaps
with a view to maximizing the potential coverage and contributions that are available. To provide
a systematic framework for organizations needing to capture and maintain case-based
24
knowledge, Nick et al. (2001) developed systematic practical strategies for guiding the
maintenance of corporate experience repositories.
In this section we summarize d research in the area of retention and maintenance in CBR. Due
to space limitations, we could only scratch the surface of this dynamic and rich area of research.
Retention and case-base editing and, more generally, case-base maintenance, continues to be a
rich source of research ideas, and even recent developments could not be discussed here in the
detail they deserve. The interested reader is referred to Wilson & Leake (2001) for a thorough
examination of the dimensions of maintenance strategies and survey of additional maintenance
research in terms of those dimensions. In addition, a recent collection of maintenance articles
addressing numerous facets of maintenance is available in Leake et al. (2001).
Conclusions
Our aim in this paper has been to provide a concise overview of the cognitive science
foundations of CBR and of the four main tasks involved in the CBR cycle, namely retrieval,
reuse, revision, and retention. Rather than presenting a comprehensive survey, we have focused
on a representative selection of work from the CBR literature in the past few decades. We have
tried to strike a balance between research that can be seen as laying the foundations of CBR and
more recent contributions. The fact that many of the cited papers were published in the last few
years is also evidence of a significant amount of ongoing research activity. It should be clear
from our discussion that much of the recent research has been motivated by an increased
awareness of the limitations of traditional approaches to retrieval, reuse, and retention. This is a
trend that seems likely to continue with the emergence of new and more demanding applications
of CBR, and we look forward to the challenges and opportunities that lie ahead.
References
Aamodt, A. 1994 Explanation-driven case-based reasoning. In Wess, S., Althoff, K.-D. & Richter, M.
(eds.), Topics in Case-Based Reasoning. Berlin: Springer, pp. 274-288.
Aamodt, A. 2004 Knowledge-intensive case-based reasoning in Creek. In Proceedings of the Seventh
European Conference on Case-Based Reasoning. Berlin: Springer, pp. 1-15.
Aamodt, A. & Plaza. E. 1994 Case-based reasoning: Foundational issues, methodological variations, and
system approaches. AI Communications 7(1), 3959.
Aha, D. W. 1998 The omnipresence of case-based reasoning in science and application. Knowledge-Based
Systems 11(5-6), 261-273.
Aha, D. W., Breslow, L. A. & Muoz-Avila, H. 2001 Conversational case-based reasoning. Applied
Intelligence 14(1), 9-32.
Aha, D. W., Kibler, D. & Albert, M. K. 1991 Instance-based learning algorithms. Machine Learning 6,
3766.
Aleven, V., 2003 Using background knowledge in case-based legal reasoning: A computational model and
an intelligent learning environment. Artificial Intelligence 150, 183-237.
Arcos, J. L. 1997 The NOOS Representation Language. PhD Thesis, Artificial Intelligence Research
Institute and Technical University of Catalonia. IIIA-CSIC Monographies Series.
Arcos, J. L. & Lpez de Mntaras, R. 1997 Perspectives: A declarative bias mechanism for case retrieval. In
Proceedings of the Second International Conference on Case-Based Reasoning. Berlin: Springer, pp. 279290.
Arcos, J. L. & Lpez de Mntaras, R. 2001 An interactive case-based reasoning approach for generating
25
26
Domeshek, E. 1992 Do the Right Things: A Component Theory for Indexing Stories as Social Advice. PhD
Thesis, Institute for the Learning Sciences, Northwestern University, Evanston, IL.
Doyle, D., Cunningham, P., Bridge, D. & Rahman, Y. 2004 Explanation oriented retrieval. In Proceedings
of the Seventh European Conference on Case-Based Reasoning. Berlin: Springer, pp. 157-168.
Emde, W. & Wettschereck, D. 1996 Relational instance-based learning. In Proceedings of the Thirteenth
International Conference on Machine Learning. San Mateo, CA: Morgan Kaufmann, pp. 122-130.
Falkenhainer, B., Forbus, K. & Gentner, D. 1986 The structure-mapping engine. In Proceedings of the Fifth
National Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press, pp. 272-277.
Faltings, B. 1997a Case reuse by model-based interpretation. In Maher, M. L. & Pu, P. (eds.) Issues and
Applications of Case-Based Reasoning in Design. Mahwah, NJ: Lawrence Erlbaum, pp. 3960.
Faltings, B. 1997b Probabilistic indexing for case-based prediction, In Proceedings of the Second
International Conference on Case-Based Reasoning, Berlin: Springer, pp. 611-622.
Faltings, B. & Sun, K. 1996 FAMING: Supporting innovative mechanism shape design. Computer-Aided
Design 28(3), 207-216.
Ferrario, M-A. & Smyth, B. 2001 Distributing case-base maintenance: The collaborative maintenance
approach. Computational Intelligence 17(2), 315-330.
Forbus, K. 2001 Exploring analogy in the large. In Gentner, D., Holyoak, K. & Kokinov, B. (eds.) The
Analogical Mind: Perspectives from Cognitive Science. Cambridge, MA: MIT Press, pp. 23-58.
Forbus, K., Ferguson, R. & Gentner, D. 1994a Incremental structure-mapping. In Proceedings of the
Sixteenth Annual Conference of the Cognitive Science Society, Hillsdale, NJ: Lawrence Erlbaum, pp. 313318.
Forbus, K., Gentner, D. & Law, K. 1994b MAC/FAC: A model of similarity-based retrieval. Cognitive
Science 19(2), 141-205.
Forbus, K., Usher, J. & Chapman, V. 2003 Qualitative spatial reasoning about sketch maps. In Proceedings
of the Fifteenth Annual Conference on Innovative Applications of Artificial Intelligence. Menlo Park, CA:
AAAI Press, pp. 85-92.
Fox, S. & Leake, D. 1995 Using introspective reasoning to refine indexing. Proceedings of the Fourteenth
International Joint Conference on Artificial Intelligence. San Mateo, CA: Morgan Kaufmann, pp. 391397.
Fox, S. & Leake, D. 2001 Introspective reasoning for index refinement in case-based reasoning. Journal of
Experimental and Theoretical Artificial Intelligence 13(1), 63-88.
Francis, A. G. & Ram, A. 1995 A Comparative utility analysis of case-based reasoning and control-rule
learning systems. In Proceedings of the Eighth European Conference on Machine Learning. Berlin:
Springer, pp. 138-150.
Fuchs, B. & Mille, A. 1999 A knowledge-level task model of adaptation in case-based reasoning. In
Proceedings of the Third International Conference on Case Based Reasoning. Berlin: Springer, pp.
118131.
Gabel, T. & Stahl, A. 2004 Exploiting background knowledge when learning similarity measures. In
Proceedings of the Seventh European Conference on Case-Based Reasoning. Berlin: Springer, pp. 169183.
Gates, G. W. 1972 The reduced nearest neighbor rule. IEEE Transactions on Information Theory IT-18(3),
431433.
Gentner, D. 1983 Structure-mapping: A theoretical framework for analogy. Cognitive Science 7, 155-170.
Gentner, D. & Forbus, K. 1991 MAC/FAC: A model of similarity-based retrieval. In Proceedings of the
Thirteenth Annual Conference of the Cognitive Science Society. Hillsdale, NJ: Lawrence Erlbaum, pp.
504-509.
Gentner, D., Holyoak, K. & Kokinov, B. (eds.) 2001 The Analogical Mind: Perspectives from Cognitive
Science. Cambridge, MA: MIT Press.
Gentner, D., Rattermann, M. J. & Forbus, K. 1993 The roles of similarity in transfer: Separating retrieval
from inferential soundness. Cognitive Psychology 25, 524-575.
Gick, M. L. & Holyoak, K.J. 1980 Analogical problem solving. Cognitive Psychology 12, 306-355.
Goel, A. & Chandrasekaran, B. 1989 Use of device models in adaptation of design cases. In Proceedings of
the DARPA Workshop on Case-Based Reasoning. San Mateo, CA: Morgan Kaufmann, pp. 100-109.
Goel, A., Kolodner, J., Pearce, M., Billington, R. & Zimring, C. 1991. Towards a case-based tool for aiding
27
conceptual design problem solving. In Bareiss, R. (ed.) Proceedings of the DARPA Workshop on CaseBased Reasoning. San Mateo, CA: Morgan Kaufmann, pp. 109-120.
Gmez de Silva Garza, A. & Maher, M. L. 2000 A process model for evolutionary design case adaptation.
In Proceedings of the Artificial Intelligence in Design Conference. Dordrecht: Kluwer Academic
Publishers, pp. 393412.
Hammond, K. J. 1986 CHEF: A model of case-based planning. In Proceedings of the Fifth National
Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press, pp. 267-271.
Hammond, K. J. 1990 Explaining and repairing plans that fail. Artificial Intelligence 45, 173-228.
Hanney, K. & Keane, M. T. 1997 The adaptation knowledge bottleneck: How to ease it by learning from
cases. In Proceedings of the Second International Conference on Case Based Reasoning. Berlin: Springer,
pp. 359370.
Hart, P. E. 1968 The condensed nearest neighbor rule. IEEE Transactions on Information Theory IT-14,
515516.
Holt, A., Bichindaritz, I., Schmidt, R. & Perner, P. 2005 Medical applications in case-based reasoning.
Knowledge Engineering Review. This issue.
Hummel, J. E. & Holyoak, K. J. 1997 Distributed representations of structure: A theory of analogical access
and mapping. Psychological Review 104, 427-466.
Iglezakis, I. Reinartz, T. & Roth-Berghofer, T. 2004 Maintenance memories: beyond concepts and
techniques for case base maintenance. In Proceedings of the Seventh European Conference on CaseBased Reasoning. Berlin: Springer, pp. 227-241.
Jantke, K. P. 1994 Non-standard concepts in similarity in case-based reasoning. In Proceedings der
Jahrestagung der Gesellschaft fr Klassifikation. Berlin: Springer, pp. 29-44.
Jarmulak, J., Craw, S. & Rowe, R. 2001 Using case-base data to learn adaptation knowledge for design. In
Proceedings of the Seventeenth International Joint Conference on Artificial Intelligence. San Francisco,
CA: Morgan Kaufmann, pp. 10111016.
Kass, A. 1989 Strategies for adapting explanations. In Proceedings of the Second DARPA Workshop on
Case-Based Reasoning. San Mateo, CA: Morgan Kaufmann, pp. 119123.
Kass, A., Leake, D. & Owens, C. 1986 SWALE: A program that explains. In Schank, R. (ed.) Explanation
Patterns: Understanding Mechanically and Creatively. Hillsdale, NJ: Lawrence Erlbaum, pp. 232-254.
Keane, M. 1988 Analogical Problem Solving. Chichester: Ellis Horwood , New York: Wiley.
Keane, M. & Brayshaw, M. 1988 The incremental analogy machine: a computational model of analogy.
Proceedings of the Third European Working Session on Learning. London: Pitman, pp. 53-62.
Keane, M. T., Ledgeway, T., & Duff, S. 1994. Constraints on analogical mapping: A comparison of three
models. Cognitive Science 18, 287-334.
Kokinov, B. & French, R. 2003 Computational models of analogy-making. In Nadel, L. (ed.) Encyclopedia
of Cognitive Science. London: Nature Publishing Group, pp. 113-118.
Kolodner, J. L. 1993 Case-Based Reasoning. San Mateo, CA: Morgan Kaufmann.
Kolodner, J. L. 1994 From natural language understanding to case-based reasoning and beyond: A
perspective on the cognitive model that ties it all together. In Schank, R. C. & Langer, E. (eds.) Beliefs,
Reasoning, and Decision Making: Psycho-Logic in Honor of Bob Abelson. Hillsdale, NJ: Lawrence
Erlbaum.
Kolodner, J. L. 1996 Making the implicit explicit: Clarifying the principles of case-based reasoning. In
Leake, D. B. (ed.) Case-Based Reasoning: Experiences, Lessons & Future Directions. Cambridge, MA:
MIT Press, pp. 349-370.
Kolodner, J. L., Camp, P. J., Crismond, D., Fasse, B., Gray, J., Holbrook, J. & Puntembakar, S. 2003
Problem-based learning meets case-based reasoning in the middle-school science classroom: Putting
learning-by-design into practice. Journal of the Learning Sciences 12(4), 495-547.
Koton. P. 1988 Reasoning about evidence in causal explanations. In Proceedings of the Seventh National
Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press, pp. 256-261.
Larkey, L. B. & Love, B. C. 2003. CAB: Connectionist analogy builder. Cognitive Science 27, 781-794.
Leake, D. B. 1991 Goal-based explanation evaluation. Cognitive Science 15(4), 509-545.
Leake, D. B. 1992 Evaluating Explanations: A Content Theory. Hillsdale, NJ: Lawrence Erlbaum.
Leake, D. B. 1996a Case-Based Reasoning: Experiences, Lessons & Future Directions. Cambridge, MA:
MIT Press.
28
Leake, D. B. (ed.) 1996b CBR in context: The present and future. In Leake, D. B. (ed.) Case-Based
Reasoning: Experiences, Lessons & Future Directions. Cambridge, MA: MIT Press, pp. 3-30.
Leake, D. B. 1998 Cognition as case-based reasoning. In Bechtel, W. & Graham, G. (eds.) A Companion to
Cognitive Science. Oxford: Blackwell, pp. 465-476.
Leake, D. B., Kinley, A. & Wilson, D. 1995 Learning to improve case adaptation by introspective reasoning
and CBR. In Proceedings of the First International Conference on Case-Based Reasoning. Berlin:
Springer, pp. 229-240.
Leake, D. B., Kinley, A. & Wilson, D. 1997 Learning to integrate multiple knowledge sources for casebased reasoning. In Proceedings of the Fifteenth International Joint Conference on Artificial Intelligence.
San Francisco, CA: Morgan Kaufmann, pp. 674-679.
Leake, D. & McSherry, D. 2005 Introduction to the special issue on explanation in case-based reasoning.
Artificial Intelligence Review 24(2), 103-108.
Leake, D., Smyth, B., Wilson, D. & Yang, Q. 2001 Introduction to the special issue on maintaining casebased reasoning systems. Computational Intelligence 17(2), 193-195.
Leake, D. & Sooriamurthi, R. 2004 Case dispatching versus case-base merging: When MCBR matters.
International Journal of Artificial Intelligence Tools 13(1), 237-254.
Leake, D. & Wilson, D. 1999 When experience is wrong: Examining CBR for changing tasks and
environments. In Proceedings of the Third International Conference on Case-Based Reasoning. Berlin:
Springer, pp. 218-232.
Leake, D. & Wilson, D. 2000 Remembering why to remember: Performance-guided case-base maintenance.
In Proceedings of the Fifth European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 161172.
Lei, C., Babka, O. & Garanito, L. A. G. 1999 Case library reduction applied to pile foundations. In
Proceedings of the Third International Conference on Case-Based Reasoning. Berlin: Springer, pp. 233247.
Lenz, M. 1996 Applying case retrieval nets to diagnostic tasks in technical domains. In Proceedings of the
Third European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 219-233.
Lenz, M., Bartsch-Sprl, B., Burkhard, H.-D. & Wess, S. (eds.) 1998 Case-Based Reasoning Technology.
Berlin: Springer.
Lpez de Mntaras, R. & Arcos, J. L. 2002 AI and music: From composition to expressive performance. AI
Magazine 23(3), 43-57.
Markman, A. B. & Gentner, D. 2001 Thinking. Annual Review of Psychology 52, 223-247.
Markovitch, S. & Scott, P. D. 1993 Information filtering: selection mechanisms in learning systems.
Machine Learning 10, 113-151.
McGinty, L. & Smyth, B. 2003 On the role of diversity in conversational recommender systems.
Proceedings of the Fifth International Conference on Case-Based Reasoning. Berlin: Springer, pp. 276290.
McKenna, E. & Smyth, B. 2000a Competence-guided editing methods for lazy learning. In Proceedings of
the Fourteenth European Conference on Artificial Intelligence. Amsterdam: IOS Press, pp. 60-64.
McKenna, E. & Smyth, B. 2000b Competence-guided case-base editing techniques. In Proceedings of the
Fifth European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 186-197.
McKenna, E. & Smyth, B. 2001 Competence-guided case discovery. In Proceedings of the Twenty-First
BCS SGES International Conference on Knowledge Based Systems and Applied Artificial Intelligence.
London: Springer, pp. 97-108.
McSherry, D. 1998 An adaptation heuristic for case-based estimation. In Proceedings of the Fourth
European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 184195.
McSherry, D. 1999 Demand-driven discovery of adaptation knowledge. In Proceedings of the Sixteenth
International Joint Conference on Artificial Intelligence, San Francisco, CA: Morgan Kaufmann, pp. 222227.
McSherry, D. 2000 Automating case selection in the construction of a case library. Knowledge-Based
Systems 13(2-3), 65-76.
McSherry, D. 2001a Intelligent case-authoring support in CaseMaker-2. Computational Intelligence 17(2),
331-345.
McSherry, D. 2001b Interactive case-based reasoning in sequential diagnosis. Applied Intelligence 14(1),
29
65-76.
McSherry, D. 2001c Minimizing dialog length in interactive case-based reasoning. In Proceedings of the
Seventeenth International Joint Conference on Artificial Intelligence. San Francisco, CA: Morgan
Kaufmann, pp. 993-998.
McSherry, D. 2001d Precision and recall in interactive case-based reasoning. In Proceedings of the Fourth
International Conference on Case-Based Reasoning. Berlin: Springer, pp. 392-406.
McSherry, D. 2002 Diversity-conscious retrieval. In Proceedings of the Sixth European Conference on
Case-Based Reasoning. Berlin: Springer, pp. 219-233.
McSherry, D. 2003a Increasing dialogue efficiency in case-based reasoning without loss of solution quality.
In Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence. San Francisco,
CA: Morgan Kaufmann, pp. 121-126.
McSherry, D. 2003b Similarity and compromise. In Proceedings of the Fifth International Conference on
Case-Based Reasoning. Berlin: Springer, pp. 291-305.
McSherry, D. 2004a Balancing user satisfaction and cognitive load in coverage-optimised retrieval.
Knowledge-Based Systems 17, 113-119.
McSherry, D. 2004b Explaining the pros and cons of conclusions in CBR. In Proceedings of the Seventh
European Conference on Case-Based Reasoning. Berlin: Springer, pp. 317-330.
McSherry, D. 2004c Incremental relaxation of unsuccessful queries. In Proceedings of the Seventh
European Conference on Case-Based Reasoning. Berlin: Springer, pp. 331-345.
McSherry, D. 2005 Explanation in recommender systems. Artificial Intelligence Review 24(2) 179-197.
Minton, S. 1990 Qualitative results concerning the utility of explanation-based learning. Artificial
Intelligence 42, 363-391.
Mougouie, B. & Bergmann, R. 2002 Similarity assessment for generalized cases by optimization methods.
In Proceedings of the Sixth European Conference on Case-Based Reasoning. Berlin: Springer, pp. 249263.
Muoz-Avila, H. 2001 Case-base maintenance by integrating case index revision and case retention policies
in a derivational replay framework. Computational Intelligence 17(2), 280-294.
Muoz-Avila, H. & Hllen, J. 1996 Feature weighting by explaining case-based planning episodes. In
Proceedings of the Third European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 280-294.
Nick, M., Althoff, K.-D. & Tautz, C. 2001 Systematic maintenance of corporate experience repositories.
Computational Intelligence 17(2), 364-386.
Ontan, S. & Plaza, E. 2003 Collaborative case retention strategies for CBR agents. In Proceedings of the
Fifth International Conference on Case-Based Reasoning, Berlin: Springer, pp. 392-406.
Osborne, H. R. & Bridge, D. G. 1996 A case base similarity framework. In Proceedings of the Third
European Workshop on Case-Based Reasoning. Berlin: Springer, pp. 309-323.
Osborne, H. R. & Bridge, D. G. 1997 Similarity metrics: A formal unification of cardinal and non-cardinal
similarity measures. In Proceedings of the Second International Conference on Case-Based Reasoning.
Berlin: Springer, pp. 235-244.
Patel, V. L. & Groen, G. J. 1986. Knowledge-based solution strategies in medical reasoning. Cognitive
Science 10, 91-116.
Plaza, E. 1995 Cases as terms: A feature term approach to the structured representation of cases. In
Proceedings of the First International Conference on Case-Based Reasoning. Berlin: Springer, pp. 265276.
Porter, B., Bareiss, R. & Holte, R. 1990 Concept learning and heuristic classification in weak theory
domains. Artificial Intelligence 45(1-2), 229-263.
Portinale, L. & Torasso, P. 2001 Case-base maintenance in a multi-modal reasoning system. Computational
Intelligence 17(2), 263-279.
Purvis, L. & Pu, P. 1995 Adaptation using constraint satisfaction techniques. In Proceedings of the First
International Conference on Case Based Reasoning. Berlin: Springer, pp. 289300.
Racine, K. & Yang, Q. 1997 Maintaining unstructured case bases. In Proceedings of the Second
International Conference on Case-Based Reasoning, Berlin: Springer, pp. 553-564.
Ramscar, M. & Yarlett, D. 2003 Semantic grounding in models of analogy: An environmental approach.
Cognitive Science 27, 41-71.
Reinartz, T., Iglezakis, I. & Roth-Berghofer, T. 2001. Review and restore for case-based maintenance.
30
31
32