10 1 1 127 5463
10 1 1 127 5463
Spanish Council for Scientific Research (CSIC), Campus UAB, 08193 Bellaterra, Barcelona, Spain [email protected], http://www.iiia.csic.es/~richard
2Dept.
!"
#%$&$'$(*),+-./
!/021435176
[email protected], http://www.swi.psy.uva.nl/usr/richard/home.html Abstract This paper gives an overview about the development of the field of Knowledge Engineering over the last 15 years. We discuss the paradigm shift from a transfer view to a modeling view and describe two approaches which considerably shaped research in Knowledge Engineering: Role-limiting Methods and Generic Tasks. To illustrate various concepts and methods which evolved in the last years we describe three modeling frameworks: CommonKADS, MIKE, and PROTG-II. This description is supplemented by discussing some important methodological developments in more detail: specification languages for knowledge-based systems, problem-solving methods, and ontologies. We conclude with outlining the relationship of Knowledge Engineering to Software Engineering, Information Integration and Knowledge Management. Key Words Knowledge Engineering, Knowledge Acquisition, Problem-Solving Method, Ontology, Information Integration
Introduction
In earlier days research in Artificial Intelligence (AI) was focused on the development of 1
formalisms, inference mechanisms and tools to operationalize Knowledge-based Systems (KBS). Typically, the development efforts were restricted to the realization of small KBSs in order to study the feasibility of the different approaches. Though these studies offered rather promising results, the transfer of this technology into commercial use in order to build large KBSs failed in many cases. The situation was directly comparable to a similar situation in the construction of traditional software systems, called software crisis in the late sixties: the means to develop small academic prototypes did not scale up to the design and maintenance of large, long living commercial systems. In the same way as the software crisis resulted in the establishment of the discipline Software Engineering the unsatisfactory situation in constructing KBSs made clear the need for more methodological approaches. So the goal of the new discipline Knowledge Engineering (KE) is similar to that of Software Engineering: turning the process of constructing KBSs from an art into an engineering discipline. This requires the analysis of the building and maintenance process itself and the development of appropriate methods, languages, and tools specialized for developing KBSs. Subsequently, we will first give an overview of some important historical developments in KE: special emphasis will be put on the paradigm shift from the so-called transfer approach to the so-called modeling approach. This paradigm shift is sometimes also considered as the transfer from first generation expert systems to second generation expert systems [43]. Based on this discussion Section 2 will be concluded by describing two prominent developments in the late eighties: Role-limiting Methods [99] and Generic Tasks [36]. In Section 3 we will present some modeling frameworks which have been developed in recent years: CommonKADS [129], MIKE [6], and PROTG-II [123]. Section 4 gives a short overview of specification languages for KBSs. Problem-solving methods have been a major research topic in KE for the last decade. Basic characteristics of (libraries of) problem-solving methods are described in Section 5. Ontologies, which gained a lot of importance during the last years are discussed in Section 6. The paper concludes with a discussion of current developments in KE and their relationships to other disciplines. In KE much effort has also been put in developing methods and supporting tools for knowledge elicitation (compare [48]). E.g. in the VITAL approach [130] a collection of elicitation tools, like e.g. repertory grids (see [65], [83]), are offered for supporting the elicitation of domain knowledge (compare also [49]). However, a discussion of the various elicitation methods is beyond the scope of this paper.
Historical Roots
In the early eighties the development of a KBS has been seen as a transfer process of human 2
knowledge into an implemented knowledge base. This transfer was based on the assumption that the knowledge which is required by the KBS already exists and just has to be collected and implemented. Most often, the required knowledge was obtained by interviewing experts on how they solve specic tasks [108]. Typically, this knowledge was implemented in some kind of production rules which were executed by an associated rule interpreter. However, a careful analysis of the various rule knowledge bases showed that the rather simple representation formalism of production rules did not support an adequate representation of different types of knowledge [38]: e.g. in the MYCIN knowledge base [44] strategic knowledge about the order in which goals should be achieved (e.g. consider common causes of a disease first) is mixed up with domain specific knowledge about for example causes for a specific disease. This mixture of knowledge types, together with the lack of adequate justifications of the different rules, makes the maintenance of such knowledge bases very difficult and time consuming. Therefore, this transfer approach was only feasible for the development of small prototypical systems, but it failed to produce large, reliable and maintainable knowledge bases. Furthermore, it was recognized that the assumption of the transfer approach, that is that knowledge acquisition is the collection of already existing knowledge elements, was wrong due to the important role of tacit knowledge for an experts problem-solving capabilities. These deficiencies resulted in a paradigm shift from the transfer approach to the modeling approach. Knowledge Engineering as a Modeling Process Nowadays there exists an overall consensus that the process of building a KBS may be seen as a modeling activity. Building a KBS means building a computer model with the aim of realizing problem-solving capabilities comparable to a domain expert. It is not intended to create a cognitive adequate model, i.e. to simulate the cognitive processes of an expert in general, but to create a model which offers similar results in problem-solving for problems in the area of concern. While the expert may consciously articulate some parts of his or her knowledge, he or she will not be aware of a significant part of this knowledge since it is hidden in his or her skills. This knowledge is not directly accessible, but has to be built up and structured during the knowledge acquisition phase. Therefore this knowledge acquisition process is no longer seen as a transfer of knowledge into an appropriate computer representation, but as a model construction process ([41], [106]). This modeling view of the building process of a KBS has the following consequences: Like every model, such a model is only an approximation of the reality. In principle, the modeling process is innite, because it is an incessant activity with the aim of approximating the intended behaviour. The modeling process is a cyclic process. New observations may lead to a renement, modication, or completion of the already built-up model. On the other side, the model may guide the further acquisition of knowledge. The modeling process is dependent on the subjective interpretations of the knowledge engineer. Therefore this process is typically faulty and an evaluation of the model with respect to reality is indispensable for the creation of an adequate model. According to this feedback loop, the model must therefore be revisable in every stage of the modeling process.
Problem Solving Methods In [39] Clancey reported on the analysis of a set of first generation expert systems developed to solve different tasks. Though they were realized using different representation formalisms (e.g. production rules, frames, LISP), he discovered a common problem solving behaviour. Clancey was able to abstract this common behaviour to a generic inference pattern called Heuristic Classification, which describes the problem-solving behaviour of these systems on an abstract level, the so called Knowledge Level [113]. This knowledge level allows to describe reasoning in terms of goals to be achieved, actions necessary to achieve these goals and knowledge needed to perform these actions. A knowledge-level description of a problemsolving process abstracts from details concerned with the implementation of the reasoning process and results in the notion of a Problem-Solving Method (PSM). A PSM may be characterized as follows (compare [20]): A PSM species which inference actions have to be carried out for solving a given task. A PSM determines the sequence in which these actions have to be activated. In addition, so-called knowledge roles determine which role the domain knowledge plays in each inference action. These knowledge roles dene a domain independent generic terminology. When considering the PSM Heuristic Classification in some more detail (Figure 1) we can identify the three basic inference actions abstract, heuristic match, and refine. Furthermore, four knowledge roles are defined: observables, abstract observables, solution abstractions, and solutions. It is important to see that such a description of a PSM is given in a generic way. Thus the reuse of such a PSM in different domains is made possible. When considering a medical domain, an observable like 410 C may be abstracted to high temperature by the inference action abstract. This abstracted observable may be matched to a solution abstraction, e.g. infection, and finally the solution abstraction may be hierarchically refined to a solution, e.g. the disease influenca.
abstract observables
heuristic match
solution abstractions
abstract
rene
observables
solutions
role
In the meantime various PSMs have been identified, like e.g. Cover-and-Differentiate for solving diagnostic tasks [99] or Propose-and-Revise [100] for parametric design tasks. PSMs may be exploited in the knowledge engineering process in different ways: 4
PSMs contain inference actions which need specic knowledge in order to perform their task. For instance, Heuristic Classication needs a hierarchically structured model of observables and solutions for the inference actions abstract and rene, respectively. So a PSM may be used as a guideline to acquire static domain knowledge. A PSM allows to describe the main rationale of the reasoning process of a KBS which supports the validation of the KBS, because the expert is able to understand the problem solving process. In addition, this abstract description may be used during the problemsolving process itself for explanation facilities. Since PSMs may be reused for developing different KBSs, a library of PSMs can be exploited for constructing KBSs from reusable components. The concept of PSMs has strongly stimulated research in KE and thus has influenced many approaches in this area. A more detailed discussion of PSMs is given in Section 5.
design-extensions refer to knowledge for proposing a new value for a design parameter, constraints provide knowledge restricting the admissible values for parameters, and xes make potential remedies available for specic constraint violations. From this characterization of the PSM Propose-and-Revise, one can easily see that the PSM is described in generic, domain-independent terms. Thus the PSM may be used for solving design tasks in different domains by specifying the required domain knowledge for the different predefined generic knowledge roles. E.g. when SALT was used for building the VT-system [101], a KBS for configuring elevators, the domain expert used the form-oriented user interface of SALT for entering domain specific design extensions (see Figure 2). That is, the generic terminology of the knowledge roles, which is defined by object and relation types, is instantiated with VT specific instances. 1 2 3 4 5 Name: Precondition: Procedure: Formula: Justication: CAR-JAMB-RETURN DOOR-OPENING = CENTER CALCULATION [PLATFORM-WIDTH OPENING-WIDTH] / 2 CENTER-OPENING DOORS LOOK BEST WHEN CENTERED ON PLATFORM.
(the value of the design parameter CAR-JUMB-RETURN is calculated according to the formula - in case the precondition is fullled; the justication gives a description why this parameter value is preferred over other values (example taken from [100]))
Fig. 2 Design Extension Knowledge for VT
On the one hand, the predefined knowledge roles and thus the predefined structure of the knowledge base may be used as a guideline for the knowledge acquisition process: it is clearly specified what kind of knowledge has to be provided by the domain expert. On the other hand, in most real-life situations the problem arises of how to determine whether a specific task may be solved by a given RLM. Such task analysis is still a crucial problem, since up to now there does not exist a well-defined collection of features for characterizing a domain task in a way which would allow a straightforward mapping to appropriate RLMs. Moreover, RLMs have a fixed structure and do not provide a good basis when a particular task can only be solved by a combination of several PSMs. In order to overcome this inflexibility of RLMs, the concept of configurable RLMs has been proposed. Configurable Role-Limiting Methods (CRLMs) as discussed in [121] exploit the idea that a complex PSM may be decomposed into several subtasks where each of these subtasks may be solved by different methods (see Section 5). In [121], various PSMs for solving classification tasks, like Heuristic Classification or Set-covering Classification, have been analysed with respect to common subtasks. This analysis resulted in the identification of 6
shared subtasks like data abstraction or hypothesis generation and test. Within the CRLM framework a predefined set of different methods are offered for solving each of these subtasks. Thus a PSM may be configured by selecting a method for each of the identified subtasks. In that way the CRLM approach provides means for configuring the shell for different types of tasks. It should be noted that each method offered for solving a specific subtask, has to meet the knowledge role specifications that are predetermined for the CRLM shell, i.e. the CRLM shell comes with a fixed scheme of knowledge types. As a consequence, the introduction of a new method into the shell typically involves the modification and/or extension of the current scheme of knowledge types [121]. Having a fixed scheme of knowledge types and predefined communication paths between the various components is an important restriction distinguishing the CRLM framework from more flexible configuration approaches such as CommonKADS (see Section 3). It should be clear that the introduction of such flexibility into the RLM approach removes one of its disadvantages while still exploiting the advantage of having a fixed scheme of knowledge types, which build the basis for generating effective knowledge-acquisition tools. On the other hand, configuring a CRLM shell increases the burden for the system developer since he has to have the knowledge and the ability to configure the system in the right way. Generic Task and Task Structures In the early eighties the analysis and construction of various KBSs for diagnostic and design tasks evolved gradually into the notion of a Generic Task (GT) [36]. GTs like Hierarchical Classification or State Abstraction are building blocks which can be reused for the construction of different KBSs. The basic idea of GTs may be characterized as follows (see [36]): A GT is associated with a generic description of its input and output. A GT comes with a xed scheme of knowledge types specifying the structure of domain knowledge needed to solve a task. A GT includes a xed problem-solving strategy specifying the inference steps the strategy is composed of and the sequence in which these steps have to be carried out. The GT approach is based on the strong interaction problem hypothesis which states that the structure and representation of domain knowledge is completely determined by its use [33]. Therefore, a GT comes with both, a fixed problem-solving strategy and a fixed collection of knowledge structures. Since a GT fixes the type of knowledge which is needed to solve the associated task, a GT provides a task specific vocabulary which can be exploited to guide the knowledge acquisition process. Furthermore, by offering an executable shell for a GT, called a task specific architecture, the implementation of a specific KBS could be considered as the instantiation of the predefined knowledge types by domain specific terms (compare [34]). On a rather pragmatic basis several GTs have been identified including Hierarchical Classification, Abductive Assembly and Hypothesis Matching. This initial collection of GTs was considered as a starting point for building up an extended collection covering a wide range of relevant tasks. However, when analyzed in more detail two main disadvantages of the GT approach have been identified (see [37]): The notion of task is conated with the notion of the PSM used to solve the task, since 7
each GT included a predetermined problem-solving strategy. The complexity of the proposed GTs was very different, i.e. it remained open what the appropriate level of granularity for the building blocks should be. Based on this insight into the disadvantages of the notion of a GT, the so-called Task Structure approach was proposed [37]. The Task Structure approach makes a clear distinction between a task, which is used to refer to a type of problem, and a method, which is a way to accomplish a task. In that way a task structure may be defined as follows (see Figure 3): a task is associated with a set of alternative methods suitable for solving the task. Each method may be decomposed into several subtasks. The decomposition structure is refined to a level where elementary subtasks are introduced which can directly be solved by using available knowledge.
diagnosis
Task
Statistical Classication
Heuristic Classication
Decision Tree
ProblemSolving Method
abstract
match
rene
Subtasks
Task / Subtasks
Fig. 3
Problem-Solving Method
Sample Task Structure for Diagnosis
As we will see in the following sections, the basic notion of task and (problem-solving) method, and their embedding into a task-method-decomposition structure are concepts which are nowadays shared among most of the knowledge engineering methodologies.
Modeling Frameworks
In this section we will describe three modeling frameworks which address various aspects of model-based KE approaches: CommonKADS [129] is prominent for having defined the structure of the Expertise Model, MIKE [6] puts emphasis on a formal and executable specification of the Expertise Model as the result of the knowledge acquisition phase, and PROTG-II [51] exploits the notion of ontologies. It should be clear that there exist further approaches which are well known in the KE community, like e.g VITAL [130], Commet [136], and EXPECT [72]. However, a discussion of all these approaches is beyond the scope of this paper.
task: diagnosis goal: nd causes which explain the observed symptoms; observables: set of observed input: symptoms; output: solutions: set of identied causes task body: control: abstract() match() rene() abstract observables heuristic match
abstract
rene
observables
Fig. 4
Domain layer: At the domain layer all the domain specic knowledge is modeled which is needed to solve the task at hand. This includes a conceptualization of the domain in a domain ontology (see Section 6), and a declarative theory of the required domain knowledge. One objective for structuring the domain layer is to model it as reusable as possible for solving different tasks. Inference layer: At the inference layer the reasoning process of the KBS is specied by exploiting the notion of a PSM. The inference layer describes the inference actions the generic PSM is composed of as well as the roles, which are played by the domain knowledge within the PSM. The dependencies between inference actions and roles are specied in what is called an inference structure. Furthermore, the notion of roles provides a domain independent view on the domain knowledge. In Figure 4 (middle part) we see the inference structure for the PSM Heuristic Classication. Among others we can see that patient data plays the role of observables within the inference structure of Heuristic Classication. Task layer: The task layer provides a decomposition of tasks into subtasks and inference actions including a goal specication for each task, and a specication of how these 10
goals are achieved. The task layer also provides means for specifying the control over the subtasks and inference actions, which are dened at the inference layer. Two types of languages are offered to describe an Expertise Model: CML (Conceptual Modeling Language) [127], which is a semi-formal language with a graphical notation, and (ML)2 [79], which is a formal specification language based on first order predicate logic, meta-logic and dynamic logic (see Section 4). Whereas CML is oriented towards providing a communication basis between the knowledge engineer and the domain expert, (ML)2 is oriented towards formalizing the Expertise Model. The clear separation of the domain specific knowledge from the generic description of the PSM at the inference and task layer enables in principle two kinds of reuse: on the one hand, a domain layer description may be reused for solving different tasks by different PSMs, on the other hand, a given PSM may be reused in a different domain by defining a new view to another domain layer. This reuse approach is a weakening of the strong interaction problem hypothesis [33] which was addressed in the GT approach (see Section 2). In [129] the notion of a relative interaction hypothesis is defined to indicate that some kind of dependency exists between the structure of the domain knowledge and the type of task which should be solved. To achieve a flexible adaptation of the domain layer to a new task environment, the notion of layered ontologies is proposed: Task and PSM ontologies may be defined as viewpoints on an underlying domain ontology. Within CommonKADS a library of reusable and configurable components, which can be used to build up an Expertise Model, has been defined [29]. A more detailed discussion of PSM libraries is given in Section 5. In essence, the Expertise Model and the Communication Model capture the functional requirements for the target system. Based on these requirements the Design Model is developed, which specifies among others the system architecture and the computational mechanisms for realizing the inference actions. KADS aims at achieving a structurepreserving design, i.e. the structure of the Design Model should reflect the structure of the Expertise Model as much as possible [129]. All the development activities, which result in a stepwise construction of the different models, are embedded in a cyclic and risk-driven life cycle model similar to Boehms spiral model [21]. The basic structure of the expertise model has some similarities with the data, functional, and control view of a system as known from software engineering. However, a major difference may be seen between an inference layer and a typical data-flow diagram (compare [155]): Whereas an inference layer is specified in generic terms and provides - via roles and domain views - a flexible connection to the data described at the domain layer, a data-flow diagram is completely specified in domain specific terms. Moreover, the data dictionary does not correspond to the domain layer, since the domain layer may provide a complete model of the domain at hand which is only partially used by the inference layer, whereas the data dictionary is describing exactly those data which are used to specify the data flow within the data flow diagram (see also [54]).
provides a development method for KBSs covering all steps from the initial elicitation through specification to design and implementation. MIKE proposes the integration of semiformal and formal specification techniques and prototyping into an engineering framework. Integrating prototyping and support for an incremental and reversible system development process into a model-based framework, is actually the main distinction between MIKE and CommonKADS [129]: MIKE takes the Expertise Model of CommonKADS as its general model pattern and provides a smooth transition from a semiformal representation, the Structure Model, to a formal representation, the KARL Model, and further to an implementation oriented representation, the Design Model. The smooth transition between the different representation levels of the Expertise Model is essential for enabling incremental and reversible system development in practice. In MIKE the executability of the KARL Model enables validation of the Expertise Model by prototyping. This considerably enhances the integration of the expert in the development process. The different MIKE development activities and the documents resulting from these activities are shown in Figure 5. In MIKE, the entire development process is divided into a number of subactivities: Elicitation, Interpretation, Formalization/Operationalization, Design, and Implementation. Each of these activities deals with different aspects of the system development. The knowledge acquisition process starts with Elicitation. Methods like structured interviews [48] are used for acquiring informal descriptions of the knowledge about the specific domain and the problem-solving process itself. The resulting knowledge expressed in natural language is stored in so-called knowledge protocols. During the Interpretation phase the knowledge structures which may be identified in the knowledge protocols are represented in a semi-formal variant of the Expertise Model: the Structure Model [112]. All structuring information in this model, like the data dependencies between two inferences, is expressed in a fixed, restricted language while the basic building blocks, e.g. the description of an inference, are represented by unrestricted texts. This representation provides an initial structured description of the emerging knowledge structures and can be used as a communication basis between the knowledge engineer and the expert. Thus the expert can be integrated in the process of structuring the knowledge. The Structure Model is the foundation for the Formalization/Operationalization process which results in the formal Expertise Model: the KARL Model. The KARL Model has the same conceptual structure as the Structure Model while the basic building blocks which have been represented as natural language texts are now expressed in the formal specification language KARL (cf. [53], [55]). This representation avoids the vagueness and ambiguity of natural language descriptions and thus helps to get a clearer understanding of the entire problem-solving process. The KARL Model can be directly mapped to an operational representation because KARL (with some small limitations) is an executable language. The result of the knowledge acquisition phase, the KARL Model, captures all functional requirements for the final KBS. During the Design phase additional non-functional requirements are considered. These non-functional requirements include e.g. efficiency and maintainability, but also the constraints imposed by target software and hardware environments. Efficiency is already partially covered in the knowledge acquisition phase, but only to the extent as it determines the PSM. Consequently, functional decomposition is 12
already part of the knowledge acquisition phase. Therefore, the design phase in MIKE constitutes the equivalent of detailed design and unit design in software engineering approaches. The Design Model which is the result of this phase is expressed in the language DesignKARL [89]. DesignKARL extends KARL by providing additional primitives for structuring the KARL Model and for describing algorithms and data types. DesignKARL additionally allows to describe the design process itself and the interactions between design decisions [90]. The Design Model captures all functional and non-functional requirements posed to the KBS. In the Implementation process the Design Model is implemented in the target hardware and software environment.
Elicitation
expert knowledge protocols
Interpretation
KBS
Structure Model
Implementation
Design Model
RC R E
Formalization Operationalization
KARL Model
Design
activity
evaluation
Fig. 5
document
The result of all phases is a set of several interrelated refinement states of the Expertise Model. The knowledge in the Structure Model is related to the corresponding knowledge in the knowledge protocols via explicit links. Concepts and inference actions are related to protocol nodes, in which they have been described using natural language. The Design Model refines the KARL Model by refining inferences into algorithms and by introducing additional data structures. These parts of the Design Model are linked to the corresponding inferences of the KARL Model and are thus in turn linked to the knowledge protocols. Combined with the goal of preserving the structure of the Expertise Model during design, the links between the different model variants and the final implementation ensure traceability of (non-)functional requirements. The entire development process, i.e. the sequence of knowledge acquisition, design, and implementation, is performed in a cycle guided by a spiral model [21] as process model. 13
Every cycle produces a prototype of the KBS which may be evaluated by testing it in the real target environment. The results of the evaluation are used in the next cycle to correct, modify, or extend this prototype. The development process of MIKE inherently integrates different types of prototyping [7]: The executability of the language KARL allows the Expertise Model to be built by explorative prototyping. Thus the different steps of knowledge acquisition are performed cyclically until the desired functionality has been reached. In a similar way, experimental prototyping can be used in the design phase for evaluating the Design Model since DesignKARL can be mapped to an executable version. The Design Model is refined by iterating the subphases of the design phase until all non-functional requirements are met. Recently, a new version of the specification language KARL has been developed [5] which integrates the notion of task and thus provides means to formally specify task-to-method decomposition structures. The MIKE approach as described above is restricted to modeling the KBS under development. To capture the embedding of a KBS into a business environment, the MIKE approach is currently extended by new models which define different views on an enterprise. Main emphasis is put on a smooth transition from business modeling to the modeling of problem-solving processes [45].
relations [71]: Renaming mappings are used for translating domain specic terms into method specic terms. Filtering mappings provide means for selecting a subset of domain instances as instances of the corresponding method concept. Class mappings provide functions to compute instances of method concepts from application concept denitions rather than from application instances. Mappings are, on the one hand, similar to the schema translation rules as discussed in the area of interoperable database systems (see e.g. [126]), and, on the other hand, they correspond e.g. to the lift operator in (ML)2 [79] or the view denitions in KARL [53] (compare Section 4). PROTG-II aims at providing a small collection of rather simple mappings to limit the reuse effort needed to specify these mappings. Thus PROTG-II recommends to reuse domain knowledge only in situations where the required mappings can be kept relatively simple [71]. method input problem-solving method method output
application ontology
Ontologies in PROTG-II
A feature of PROTG-II is that it can generate knowledge-acquisition tools from domain or application ontologies [50]. Therefore, the PROTG-II system includes the DASH component which takes an ontology as input and generates as output a knowledge-acquisition tool that allows domain specialists to enter instances of the domain concepts. Thus domain facts can be specified by the domain expert himself. Recently, the PROTG-II approach has been extended to CORBA-based PSMs and ontologies which enables the reuse of these components in an Internet environment [70].
Over the last ten years a number of specification languages have been developed for describing KBSs. These specification languages can be used to specify the knowledge required by the system as well as the reasoning process which uses this knowledge to solve 15
the task which is assigned to the system. On the one hand, these languages should enable a specification which abstracts from implementation details. On the other hand, they should enable a detailed and precise specification of a KBS at a level which is beyond the scope of specifications in natural language. This area of research is quite well documented by a number of workshops1 and comparison papers that were based on these workshops. Surveys of these languages can be found in [143] and [61]. A short description of their history and usefulness is provided by [80], and [54] provides a comparison to similar approaches in software engineering. In this article, we will focus on the main principles of these languages. Basically, we will discuss the general need for specification languages for KBSs, we show their essence, we sketch some approaches, add a comparison to related areas of research and conclude by outlining lines of current and future research. 4.1 Why did the Need Arise for Specication Languages in the Late 1980s
As mentioned above, we can roughly divide the development of knowledge engineering into the knowledge transfer and the knowledge modelling period. During the former period, knowledge was directly encoded using rule-based implementation languages or frame-based systems. The (implicit) assumption was that these representation formalisms are adequate to express knowledge, reasoning, and functionality of a KBSs in a way which is understandable for humans and for computers. However, as mentioned in Section 2, severe difficulties arose [40]: different types of knowledge were represented uniformly, other types of knowledge were not presented explicitly, the level of detail was too high to present abstract models of the KBS, and knowledge level aspects got constantly mixed with aspects of the implementation. As a consequence, such systems were hard to build and to maintain when they become larger or used over a longer period. In consequence, many research groups worked on more abstract description means for KBSs. Some of them were still executable (like the generic tasks [37]) whereas others combined natural language descriptions with semiformal specifications. The most prominent approach in the latter area is the KADS and CommonKADS approach [129] that introduced a conceptual model (the Expertise Model) to describe KBSs at an abstract and implementation independent level. As explained above, the Expertise Model distinguishes different knowledge types (called layers) and provides for each knowledge type different primitives (for example, knowledge roles and inference actions at the inference layer) to express the knowledge in a structured manner. A semiformal specification language CML [127] arose that incooperates these structuring mechanisms in the knowledge level models of KBSs. However, the elementary primitives of each model were still defined by using natural language. Using natural language as a device to specify computer programs has well known advantages and disadvantages. It provides freedom, richness, easiness in use and understanding, which makes it a comfortable tool in sketching what one expects from a program. However, its inherent vagueness and implicitness make it often very hard to answer questions whether the system really does what is expected, or whether the model is consistent or correct (cf. [2], [80]). Formal specification techniques arose that overcome these shortcomings. Usually they were not meant as a replacement of semiformal specifications but as the possibility to improve the preciseness of a specification when required.
1. See ftp://swi.psy.uva.nl/pub/keml/keml.html at the World Wide Web.
16
Meanwhile, around twenty different approaches can be found in the literature ([61], [54]). Some of them aim mainly at formalization. A formal semantics is provided that enables the unique definition of knowledge, reasoning, or functionality along with manual or automated proofs. Other approaches aim at operationalization, that is, the specification of a system can be executed which enables prototyping in the early phase of system development. Here, the evaluation of the specification is the main interest. They help to answer the question whether the specification really specifies what the user is expecting or the expert is providing. Some approaches aim at formalizing and operationalizing, however they have to tackle conflicting requirements that arise from these two goals. In the following subsection, we will discuss the main common features of these approaches. 4.2 The Essence of Specication Languages for KBSs
We identify three key features of specification languages for KBSs. First, most languages make use of a strong conceptual model to structure formal specifications. This reflects the fact that these languages were motivated by formalizing semiformal notations to describe KBSs. Therefore, these languages offer more than just a mathematical notation to define a computer program as an input-output relationship. Second, these languages have to provide means to specify the dynamic reasoning of a KBSs because this establishes a significant piece of the expertise required by such systems. Third, a KBS uses a large body of knowledge requiring structured and rich primitives for representing it. In the following, we will discuss these different aspects. Formalising a Conceptual Model Specification languages for KBSs arose to formalize conceptual models of KBSs. They use the structuring principles of semiformal specifications and add formal semantics to the elementary primitives and their composition (cf. [46], [47], [154]). As introduced above, the Expertise Model [129] describes the different types of knowledge required by a KBS as well as the role of this knowledge in the reasoning process of the KBS. Based on this, specification languages provide formal means for precisely defining: the goals and the process to achieve them, the functionality of the inference actions, and the precise semantics of the different elements of the domain knowledge. Definitions in natural language are supplemented with formal definitions to ensure disambiguity and preciseness. The structure of the conceptual models naturally organizes the formal specification, improving understandability and simplifying the specification process. The What and the How: Specication of Reasoning In Software Engineering, the distinction between a functional specification and the design/ implementation of a system is often discussed as a separation of what and how. During the specification phase, what the system should do is established in interaction with the users. How the system functionality is realized is defined during design and implementation (e.g., which algorithmic solution can be applied). This separation - which even in the domain of Software Engineering is often not practicable in the strict sense - does not work in the same way for KBSs: A high amount of the problem-solving knowledge, i.e. knowledge about how 17
to meet the requirements, is not a question of efficient algorithms and data structures, but exists as domain-specific and task-specific heuristics as a result of the experience of an expert. For many problems which are completely specifiable it is not possible to find an efficient algorithmic solution. Such problems are easy to specify but it is not necessarily possible to derive an efficient algorithm from these specifications (cf. [32], [110]); domainspecific heuristics or domain-specific inference knowledge is needed for the efficient derivation of a solution. In simple terms this means that analysis is not simply interested in what happens, as in conventional systems, but also with how and why [31]. One must not only acquire knowledge about what a solution for a given problem is, but also knowledge about how to derive such a solution in an efficient manner [60]. Already at the knowledge level there must be a description of the domain knowledge and the problem-solving method which is required by an agent to solve the problem effectively and efficiently. In addition, the symbol level has to provide a description of efficient algorithmic solutions and data structures for implementing an efficient computer program. As in Software Engineering, this type of knowledge can be added during the design and implementation of the system. Therefore a specification language for KBSs must combine non-functional and functional specification techniques: On the one hand, it must be possible to express algorithmic control over the execution of substeps. On the other hand, it must be possible to characterize the overall functionality and the functionality of the substeps without making commitments to their algorithmic realization. The former is necessary to express problem solving at the knowledge level. The latter is necessary to abstract from aspects that are only of interest during implementation of the system. Representing Rich Knowledge Structures Most of the specification languages provide at least epistemological primitives like constants, sorts/types, functions, predicates/relations, and some mathematical toolkit as a means to specify the static aspects of a system. However, richer languages provide additional syntactical sugar on top of these mathematical primitives. In particular, object-oriented or frame-based specification languages provide a rich variety of appropriate modelling primitives for expressing static system aspects: values, objects, classes, attributes with domain and range restrictions, set-valued attributes, is-a relationships with attribute inheritance, aggregation, grouping etc. For example, KARL [53] provides such modelling primitives which enables a smooth transformation from semiformal specification languages like CML to formal specifications of these models. 4.3 Some Approaches And Their Technical Means
In the following, we briey sketch three different specication languages for KBSs: (ML)2, KARL, and DESIRE. A more detailed discussion and comparison can be found in [54]. (ML)2 [79], which was developed as part of the KADS projects, is a formalization language for KADS Expertise Models. The language provides a formal specification language for the KADS Expertise Model by combining three types of logic: order-sorted first-order logic extended by modularization for specifying the domain layer, first-order meta-logic for specifying the inference layer, and quantified dynamic logic [77] for specifying the task layer. KARL [53] is an operational language which restricts the expressive power of the object logic by using a variant of Horn logic. It was developed as part of the MIKE project [6] and 18
provides a formal and executable specification language for the KADS Expertise Model by combining two types of logic: L-KARL and P-KARL. L-KARL, a variant of Frame Logic [84], is provided to specify domain and inference layers. It combines first-order logic with semantic data modelling primitives (see [30] for an introduction to semantic data models). A restricted version of dynamic logic is provided by P-KARL to specify a task layer. Executability is achieved by restricting Frame logic to Horn logic with stratified negation and by restricting dynamic logic to regular and deterministic programs. The language DESIRE (DEsign and Specification of Interacting REasoning components [91],[92]) relies on a different conceptual model for describing a KBS: the notion of a compositional architecture. A KBS is decomposed into several interacting components. Each component contains a piece of knowledge at its object-layer and has its own control defined at its internal meta-layer. The interaction between components is represented by transactions and the control flow between these modules is defined by a set of control rules. DESIRE extensively uses object-meta relationships to structure specifications of KBSs. At the objectlevel, the system reasons about the world state. Knowledge about how to use this knowledge to guide the reasoning process is specified at the meta-level. The meta-level reasons about controlling the use of the knowledge specified at the object-level during the reasoning process. The meta-level describes the dynamic aspects of the object-level in a declarative fashion. A module may reason on its object-level about the meta-level of another module. From a semantic point of view a significant difference between DESIRE on the one hand and (ML)2 and KARL on the other hand lies in the fact that the former uses temporal logics for specifying the dynamic reasoning process whereas the latter use dynamic logic. In dynamic logic, the semantics of the overall program is a binary relation between its input and output sets. In DESIRE, the entire reasoning trace which leads to the derived output is used as semantics [142]. 4.4 Comparison with Related Work
In the following, we investigate the relationships with three other areas. We take a look at the work that is done on validation and verification of KBSs, we discuss related work on specification languages in Software Engineering, and finally we discuss a recent trend in Software Engineering that makes use of the knowledge level for specifying other types of software products than KBSs only. Comparison with V&V The work on validation and verification of KBSs only partially follows the paradigm shift in Knowledge Engineering. Most of the work is still oriented to specific implementation formalisms like rule-based languages or languages stemming from the knowledge representation area. Surveys can be found in [96], [120]. A typical example is the work of [122] that provides algorithms to verify knowledge that is represented with production rules. Examples of such properties are: Unsatisable rule: the precondition of a rule cannot be true for any legal input. Unusable rule: the postcondition of the rule cannot be used to infer the goal nor any preconditions of another rule. Subsumed rule: there exists are more general rule in the rule set.
19
Other approaches apply testing as a means to validate KBSs. Most approaches do not rely on a formal specification in addition to the implemented knowledge. As a consequence, verification of the functionality of a system with respect to its formal specification is not possible. Only recently, the need was recognized for formal specifications and conceptual models that guide and structure the validation and verification process ([103],[78], [59]). Comparison with Traditional Specication Approaches in SE Work in Software Engineering on formal specification languages has a more than thirty year old tradition which has lead to a large number of approaches. An important line of research are algebraic specifications (cf. [46], [154]) that provide well-studied means for the mathematical definitions of the functionality of software systems. Some of the existing specification languages rely directly on these techniques (e.g., [134], [117]), other use similar ideas. The main extensions that are necessary to use them for KBSs are: Algebraic specication formalisms provide generic mathematical notations that have to be integrated into existing conceptual modelling techniques for KBSs to enable smooth transitions from semiformal to formal specications. Algebraic specication formalisms provide a means to dene the functionality of a software system abstracting from how it is computed. We explained earlier that such an abstraction is too strong for the purpose of specifying KBSs as part of a specication is to dene how the output can be derived. Specification approaches like the Vienna Development Method (VDM) [82] and Z [135], which describe a system in terms of states and operations working on these states, are closer in spirit to most specification approaches customized for KBSs. However, their vocabulary for expressing control over these state transitions is rather restricted by their purpose. Algorithmic control is regarded as an aspect that is added during design and implementation of a system. Also, the need for a rich conceptual model that structures a specification of a KBSs becomes apparent through the large case study of [105]. There the architecture of Soar is specified with more than hundred pages filled with formal specification code in Z. Still, work on specification languages for KBSs must always be aware of approaches from Software Engineering or Information Systems Engineering because at a technical level many problems are basically the same. Comparison with Recent Approaches in SE
USER: I need a software system that will help me manage my factory. SOFTWARE ENGINEER: Well, lets see. I can put together components that do sorting, searching, stacks, queues, and so forth. USER: Hmm, thats interesting. But how would those t into my system?[132]
Recently, the knowledge level has been encountered in Software Engineering (cf. [132]). Work on software architectures establishes a much higher level to describe the functionality and the structure of software artefacts. The main concern of this new area is the description of generic architectures that describe the essence of large and complex software systems. Such architectures specific classes of application problems instead of focusing on the small and generic components from which a system is built up. 20
The conceptual models developed in Knowledge Engineering for KBSs fit nicely in this recent trend. They describe an architecture for a specific class of systems: KBSs. Work on formalizing software architectures characterizes the functionality of architectures in terms of assumptions over the functionality of its components [118], [119]. This shows strong similarities to recent work on problem-solving methods that define the competence of problem-solving methods in terms of assumptions over domain knowledge (which can be viewed as one or several components of a KBSs) and the functionality of elementary inference steps (cf. [59]). 4.5 Recent Issues
One of the goals of formal specification approaches is to provide a clear formal notation. However, some constructs in the existing languages are rather ad-hoc and work is going on to provide well-defined formal ground for these languages. These activities also enable another goal of formal specifications: semiautomatic or automated proof support which helps to deal with the overwhelming amount of details that arise in formal proofs of specifications. Above, we mentioned recent work that integrate conceptual models into existing verification tools similar to earlier activities that integrates conceptual models into mathematical notations to provide high-level specification support. In addition to formal support in verification, one could also wish to provide automated support in deriving refined specifications or implementations of KBSs with techniques of automated program development (cf.[133]). In the following sections we will discuss the subfields problem-solving methods and ontologies which are fruitful research areas generalizing the original notions of inference engines and domain knowledge bases. Clearly, these fields introduce new interesting requirements for formal approaches. The competence of problem-solving methods (i.e., their functionality and utility) need to be characterized in terms of assumptions on available knowledge. Ontologies need to provide meta-level characterizations of knowledge bases to support their reuse and to solve the interaction problem between the problem-solving process and the domain knowledge. Building large knowledge bases requires sophisticated structuring of, and mediation facilities between different ontologies.
Problem-Solving Methods
Originally, KBSs used simple and generic inference mechanisms to infer outputs for provided cases. The knowledge was assumed to be given declaratively by a set of Horn clauses, production rules, or frames. Inference engines like unification, forward or backward resolution, and inheritance captured the dynamic part of deriving new information. However, human experts have exploit knowledge about the dynamics of the problem-solving process and such knowledge is required to enable problem-solving in practice and not only in principle [60]. [38] provided several examples where knowledge engineers implicitly encoded control knowledge by ordering production rules and premises of these rules that together with the generic inference engine, delivered the desired dynamic behaviour. Making this knowledge explicit and regarding it as an important part of the entire knowledge contained by a KBS, is the rationale that underlies Problem-Solving Methods (PSMs). PSMs refine generic inference engines mentioned above to allow a more direct control of the reasoning process. PSMs describe this control knowledge independent from the application 21
domain enabling reuse of this strategical knowledge for different domains and applications. Finally, PSMs abstract from a specific representation formalism as opposed to the general inference engines that rely on a specific representation of the knowledge. Meanwhile, a large number of such PSMs have been developed and libraries of such methods provide support in reusing them for building new applications. In general, software and knowledge engineers agree that reuse is a promising way to reduce development costs of software systems and knowledge-based systems. The basic idea is that a KBS can be constructed from ready-made parts instead of being built up from scratch. Research on PSMs has adopted this philosophy and several PSM libraries have been developed. In the following, we will discuss several issues involved in developing such libraries.
applying such a PSM in a particular application requires considerable refinement and adaptation. This phenomenon is known as the reusability-usability trade-off [85]. Recently, research has been conducted to overcome this dichotomy by introducing adapters that gradually adapt task-neutral PSMs to task-specific ones [58] and by semi-automatically constructing the mappings between task-neutral PSMs and domain knowledge [19]. Libraries with informal PSMs provide above all support for the conceptual specification phase of the KBS, that is, they help significantly in constructing the reasoning part of the Expertise Model of a KBS [128]. Because such PSMs are informal, they are relatively easy to
!
66#6 7 7 6! 6#
021 "# !
7
6
1%
66 1 1#4!
77
6/
Libraries with formal PSMs are particularly important if the KBS needs to have some guaranteed properties, e.g. for use in safety-critical systems such as nuclear power plants. Their disadvantage is that they are not easy to understand for humans [24] and limit the expressiveness of the knowledge engineer. Apart from the possibility to prove properties, formal PSMs have the additional advantage of being a step closer to an implemented system. Libraries with implemented PSMs allow the construction of fully operational systems. The other side of the coin is, however, that the probability that operational PSMs exactly match the requirements of the knowledge engineer, is lower. Developing a KBS using libraries with coarse-grained PSMs, amounts to selecting the most suitable PSM and then adapt it to the particular needs of the application [107]. The advantage is that this process is relatively simple as it involves only one component. The disadvantage is, however, that it is unlikely that such a library will have broad coverage, since each application might need a different (coarse-grained) PSM. The alternative approach is to have a library with fine-grained PSMs, which are then combined together (i.e. configured) into a reasoner, either manually ([123], [137]) or automatically ([14], [9]).
%11-
7 %1
!
#
-1
#" ($7 &%'(*/1 #" )!
*%1
1
specify the assumptions under which the PSM can be applied correctly, such as its requirements on domain knowledge ([17], [18]). Selection of PSMs from such libraries first considers the competence of PSMs (selecting those whose competences match the task at hand), and then the assumptions of PSMs (selecting those whose assumptions are satisfied). Libraries can also be organised, based on the functionality of PSMs, in which case PSMs with similar functionality are stored together. In addition, the functionality of PSMs can be 23
configured from pre-established parameters and values [139]. An opposite way to organize PSMs is proposed in [124] where PSMs are indexed according to the algorithms they use. Another criterion to structure libraries of PSMs is based on solely assumptions, which specify under what conditions PSMs can be applied. Assumptions can refer to domain knowledge (e.g. a certain PSM needs a causal domain model) or to task knowledge (a certain PSM generates locally optimal solutions). To our knowledge, there does not exist a library organised following this principle, but work is currently being performed to shed more light on the role of assumptions in libraries for knowledge engineering ([17], [18], [56], [58]). A last proposal to organise libraries of PSMs is based on a suite of so-called problem types ([27], [28]) (or tasks, for the purpose of this article, tasks and problem types are treated as synonyms). The suite describes problem types according to the way that problems depend on each other. The solution to one problem forms the input to another problem. For example, the output of a prediction task is a certain state, which can form the input to a monitoring task that tries to detect problems, which on their turn can be the input to a diagnosis task. It turns out that these problem dependencies recur in many different tasks. According to this principle, PSMs are indexed under the problem type they can solve. Selection of PSMs in such a library would first identify the problem type involved (or task), and then look at the respective PSMs for this task.
24
Ontologies
Since the beginning of the nineties ontologies have become a popular research topic investigated by several Artificial Intelligence research communities, including knowledge engineering, natural-language processing and knowledge representation. More recently, the notion of ontology is also becoming widespread in fields such as intelligent information integration, information retrieval on the Internet, and knowledge management. The reason for ontologies being so popular is in large part due to what they promise: a shared and common understanding of some domain that can be communicated across people and computers. The main motivation behind ontologies is that they allow for sharing and reuse of knowledge bodies in computational form. In the Knowledge Sharing Effort (KSE) project [111], ontologies are put forward as a means to share knowledge bases between various KBSs. The basic idea was to develop a library of reusable ontologies in a standard formalism, that each system developer was supposed to adopt.
7 66 1 !7 6! # 7 16" 1
#/# 76
)! 6
751 1
25176
%1 !
)6 5# 7 1
beings in the world. Artificial Intelligence (AI) deals with reasoning about models of the world. Therefore, it is not strange that AI researchers adopted the term ontology to describe what can be (computationally) represented of the world in a program. Many definitions of ontologies have been given in the last decade, but one that characterises best, in our opinion, the essence of an ontology is based on the related definitions in ([74], [22]): An ontology is a formal, explicit specification of a shared conceptualisation. A conceptualisation refers to an abstract model of some phenomenon in the world by having identified the relevant concepts of that phenomenon. Explicit means that the type of concepts used, and the constraints on their use are explicitly defined. For example, in medical domains, the concepts are diseases and symptoms, the relations between them are causal and a constraint is that a disease cannot cause itself. Formal refers to the fact that the ontology should be machine readable, which excludes natural language. Shared reflects the notion that an ontology captures consensual knowledge, that is, it is not private to some individual, but accepted by a group. Almost all ontologies that are nowadays available are concerned with modelling static domain knowledge, as opposed to dynamic reasoning knowledge (e.g., domain models in medicine, power plant, cars, mechanic machines). In its strongest form, an ontology tries to capture universally valid knowledge, independent of its use, a view closely related to its philosophical origin. However, AI researchers quickly gave up this view, because it turned out that specific use of knowledge influenced its modelling and representation. A weaker, but still strong, notion of ontology is CYC's aim to capture human common-sense knowledge [93]. Other researchers aim at capturing domain knowledge, independent of the task or method that might use the knowledge [76]. Within a given domain, an ontology is not just a representation - in a computer - of that domain. An ontology also claims to reflect a certain rate of consensus about the knowledge in that domain. However, in practical ontological engineering research, the definition of ontology has been somewhat diluted, in the sense that taxonomies are considered to be full ontologies. Ontologies differ in two respects from taxonomies as such: they have richer 25
7 6 "##"&
"
internal structure and reflect some consensus. The question is then of course, consensus between whom? In practice, this question does not have one unique answer; it depends on the context. For example, if a hospital is building up an ontology with knowledge about a
1
# &# 6 66
1!6 1 1
consensus should be between the doctors involved. If, on the other hand, a government wants to setup a nation-wide network of bibliographic, ontology-based databases that can be consulted from nearly every terminal with an Internet connection in the country, then the consensus should be nation-wide (i.e. everybody should accept this ontology as workable). In the library example, consensus should be reached about terms such as books, authors, journals, etc., and about axioms such as: if two articles appear in the same journal, they should have the same volume and issue number. Especially because ontologies aim at consensual domain knowledge, their development is often a cooperative process involving different people, possibly at different locations. People who agree to accept an ontology are said to commit themselves to that ontology.
26
27
6 !6 )
1
7#
1
6!# ) #
6 1 . )
!6 6
"
1..
! " 1
!6
*
7
%1 # #
!
of strings. The KACTUS project [16] was concerned with constructing large ontologies for technical devices through incremental refinement of general ontologies into technical ontologies.
ontology for integer arithmetic (where + only is applied to integers, since all numbers are integers in that world). The last way to assemble ontologies that we discuss here is polymorphic refinement, known from object-oriented approaches. Suppose we want to
7 # %7
7 (7 6!!
!
#
1
*1 !
6
% 11
66 17!
#
6
# 7 6!
1 ! #
"!
#
However, as we stated earlier, ontologies are more than taxonomies. Ontologies include all relevant constraints between classes, attribute values, instances and relations (i.e. axioms). For example, in some domains the part-of relation is fundamental, and an ontology of such domain has to include and define this relation (see the mereological super theory of the PHYSSYS ontology [22]). Typical AI languages that can be used for implementing ontologies are description logics for representing declarative knowledge. Most of such languages support subsumption checking, automatic classification and consistency maintenance. Examples of such description logics include KL-ONE [26], LOOM [97], KRYPTON [25], CYCL [94]. A dedicated language for specifying [74], which has become quite 6! 1 6
# 6
ontologies #"# is
Ontolingua
1%%6# 7
1
7/ "
$
1 1 !
#66 7
7 # #! 6! %1 !
# 6!#4 " #"
about relations. Ontolingua's main aim is to allow sharing and communication of ontologies, and not so much reasoning with them. Therefore, Ontolingua comes with several translators able to generate code in Prolog, CORBA's IDL [114], CLIPS, LOOM [97], KIF, Epikit [66].
During the last decade research in Knowledge Engineering (KE) resulted in several important achievements which are also relevant for other disciplines like Software Engineering, Information Integration or Knowledge Management. Notable developments are: Within the framework of model-based KE, model structures have been dened which clearly separate the different types of knowledge which are important in the context of Knowledge-based Systems (KBSs). The Expertise Model is the most prominent example of these models.
29
The clear separation of the notions of task, problem-solving method, and domain knowledge provides a promising basis for making the reuse-oriented development of KBSs more feasible. The integration of a strong conceptual model is a distinctive feature of formal specication languages in Knowledge Engineering. Subsequently, we will discuss some relationships between methods in Knowledge Engineering and other disciplines.
mediators. In [68], the Infomaster system is described which is a generic tool for integrating different types of information sources, like e.g. relational databases or Web pages. Each information source is associated with a wrapper that hides the source specific information structure of the information sources. Internally, Infomaster uses the Knowledge Interchange Format for representing knowledge (compare Section 6). In Infomaster so-called base relations are used for mediating between the conceptual structure of the information sources and the user applications. The collection of base relations may be seen as a restricted domain ontology for integrating the different heterogeneous information sources. In the meantime ontologies are also used for supporting the semantic retrieval of information from the World-Wide Web. The SHOE approach [95] proposes to annotate Web pages with ontological information which can then be exploited for answering queries. Thus, the syntactic based retrieval of information from the Web as known from the various Web search engines is replaced by a semantic based retrieval process. A further step is taken in the Ontobroker project [57] which proposes to use a more expressive ontology combined with a corresponding inference mechanism. Thus, the search metaphor of SHOE is replaced by an inference metaphor for retrieving information from the Web since the inference mechanism can use complex inferences as part of the query answering process.
31
Acknowledgement
Thanks are due to Stefan Decker for valuable comments on a draft version of the paper. Rainer Perkuhn provided valuable editorial support. Richard Benjamins was partially supported by the Netherlands Computer Science Research Foundation with financial support from the Netherlands Organisation for Scientific Research (NWO), and by the European Commission through a Marie Curie Research Grant (TMR).
References
[1]
[9]
!
"#$
%&
(')* +,
-/.&0123$1546738$%9:
;<.&.&=&.&(>?%@
AB
C'%9=&
C/')* +,
-/.&AD%&.9:E12%&
.&:/>?0F0C.9G4 "#
=&-H?IJ=&K)%9=&.90K.&L =98$
-MI=&%9:NJOP0=&;=&.9J
J
=&%Q?
::R4SUTS;
.&
semiotic View, International Journal of Human-Computer Studies, 43(2/3):929-944, 1995. "# =9-/
V:.9.&W
X5).&
%&.&W>?KZY
%&.&[
U\$LL
]V
%&W.9G4"#^_HV
0W
.&=& ')*
+,-H.&0 ')* +,
-/.&0`12%&-H0%9a.&:bV=9KE@$c.&
:0.&E
b>?K.&$)%9W.&b38%&
:W;
<.&0.&=9.& International Journal of Expert Systems: Research and Applications, 8(2):93-120, 1995. ' )*
+,
-/.&d
W eT2=&;
=&
L
0a.&:W3 Tfc
a%9L-/=6>6K:KW)
0
%&?.&
.9#42 Case-study in Diagnosis, International Journal of Human-Computer Studies, 47:259-288, 1997. "# +,
-/.&g1Q 0=&g$FF
-/
0B
hi.&
%&.&:
BI=&%&:N]O0=&;
=&.&i@H>?
;K.9;
%c-/
.&0 ')*
+,
-/.&02cS
0
%&2)dTS=&
=&-H200-/L=&.9B@Q12%9-/0%&a.9:!D=&KB
>?K.9Q)%9 .&J38%&
:kj?:.&
.&:6.&G4ShlhiK%90=&6
.&=& Proc. ECAI-96, pages 408-412. J. Wiley & Sons, Ltd., 1996. ' )* +,
-/.&$
mO*12.&
=&Y%&
.&;
K$0-/L=&.&$@1S%&
-/0%&a.&:mV
=&K0.&G4C!T2K%9=6=
al., eds., Lecture Notes in Artificial Intelligence, 1076, 9th European Knowledge Acquisition Workshop, in: N. Shadbolt et al., eds., Advances in Knowledge Acquisition, Lecture Notes in Artificial Intelligence 1076, Springer-Verlag, Berlin, 1996. Working Papers European Conference on Artificial Intelligence ECAI'94 Workshop on Implemented Ontologies, pages 35-45, Amsterdam, 1994. Problem Solving Methods for Diagnosis. PhD Thesis, University of Amsterdam, Amsterdam, The Netherlands, 1993. Building Practical Planning System, in: Proc. of the 15th International Joint Conference on Artificial Intelligence (IJCAI 97), pages 1246-1251, Japan, 1997, Morgan Kaufmann Publishers, Inc. Third International Conference on Artificial Intelligence Planning Systems (AIPS-96), pages 11-18. American Association of Artificial Intelligence (AAAI), 1996.
A. Abecker, S. Decker, K. Hinkelmann, and U. Reimer, Proc. Workshop Knowledge-based Systems for Knowledge Management in Enterprises, 21st Annual German Conference on AI (KI97), Freiburg, 1977. URL: http://www.dfki.uni-kl.de/km/ws-ki-97.html ormally Specifying Re-usable Knowledge Model Components, Knowledge Acquisition 5:119-141, 1993. M. Aben, Formal Methods in Knowledge Engineering, PhD Thesis, University of Amsterdam, 1995. Multiple Domain Models in Diagnostic Reasoning, PhD Thesis, University of Amsterdam, Amsterdam 1994. J. Angele, S. Decker, R. Perkuhn, and R. Studer, Modeling Problem-Solving Methods in NewKARL, in: Proc. of the 10th Knowledge Acquisition for Knowledge-based Systems Workshop (KAW96), Banff, 1996. J. Angele, D. Fensel, and R. Studer, Developing Knowledge-Based Systems with MIKE, to appear in Journal of Automated Software Engineering, 1998. J. Angele, D. Fensel, and R. Studer, Domain and Task Modeling in MIKE, in: A. Sutcliffe et al., eds., Domain Knowledge for Interactive System Design, Chapman & Hall, 1996. H. Akkermans, B. Wielinga, and A.Th. Schreiber, Steps in Constructing Problem-Solving Methods, in: N. Aussenac et al., eds., Knowledge Acquisition for Knowledge-based Systems, 7th European Workshop (EKAW93), Toulouse, Lecture Notes in AI 723, Springer-Verlag, 1993.
[10]
[11] [12]
[17]
[18]
32
[19]
h 0=9 0=`
J"#AFF "#15 8
Y!6.&;K
N>6
O-/-/
W@/S-/
%dV
=&K0 )*" ;
K-/
?'!15CY.&%&
=&?
"#C?a
0<
?
j60=&.9%QN.&
)
0.&:
TSN0=9-
423$8%&
: and Symbol Level Accounts of KRYPTON, in: Proceedings IJCAI-85, 1985. )* " ;
K-/
" TS;
K-/%&[
$ Ia
a.&
8 @!=&K(3$?IjU3$8%&
:()
L
0
=&=9.& TSN0=9-/ Cognitive Science, 9(2), 1985. "#
FGO-/L
=&$@12%&- TS%9a.&:m
m>?NL
$@12%9-/0G.&G4CTS=&
%9$
=6
%, G# A Future of
Knowledge Acquisition, Proc. 8th European Knowledge Acquisition Workshop (EKAW94), Hoegaarden, Lecture Notes in Artificial Intelligence 867, Springer-Verlag, 1994. J. Breuker, A Suite of Problem Types, in: J. A. Breuker and W. van de Velde, eds., The CommonKADS Library For Expertise Modelling, IOS Press, Amsterdam, 1994. J. A. Breuker and W. van de Velde, eds., The CommonKADS Library For Expertise Modelling, IOS Press, Amsterdam, 1994. M. L. Brodie, On the Development of Data Models, in: Brodie et al., eds., On Conceptual Modeling, Springer-Verlag, Berlin, 1984. A. G. Brooking, The Analysis Phase in Development of Knowledge-Based Systems, in: W. A. Gale, ed., AI and Statistic, Addison-Wesley Publishing Company, Reading, Massachusetts, 1986. T. Bylander, D. Allemang, M.C. Tanner, and J.R. Josephon, The Computational Complexity of Abduction, Artificial Intelligence 49, 1991. T. Bylander and B. Chandrasekaran, Generic Tasks in Knowledge-based Reasoning: The Right Level of Abstraction for Knowledge Acquisition, in: B. Gaines and J. Boose, eds., Knowledge Acquisition for Knowledge Based Systems, Vol. 1, Academic Press, London, 1988. T. Bylander and S. Mittal, CSRL, A Language for Classificatory Problem Solving, AI Magazine 8, 3, 1986, 66-77. AI Magazine, 11:59-71, 1990. B. Chandrasekaran, Generic Tasks in Knowledge-based Reasoning: High-level Building Blocks for Expert System Design, IEEE Expert 1, 3 1986, 23-30. B. Chandrasekaran, T. R. Johnson, and J. W. Smith, Task Structure Analysis for Knowledge Modeling, Communications of the ACM 35, 9 (1992), 124137. W.J. Clancey, The Epistemology of a Rule-Based Expert System - a Framework for Explanation, Artificial Intelligence 20 (1983), 215-251. W.J. Clancey, Heuristic Classification, Artificial Intelligence 27 (1985), 289-350. W.J. Clancey, From Guidon to Neomycin and Heracles in Twenty Short Lessons, in: A. van Lamsweerde, ed., Current Issues in Expert Systems, Academic Press, 1987. W.J. Clancey, The Knowledge Level Reinterpreted: Modeling How Systems Interact, Machine Learning 4, 1989, 285-291. F. Cornelissen, C.M. Jonker, and J. Treur: Compositional Verification of Knowledge-based Systems: A Case Study for Diagnostic Reasoning, in: E. Plaza and R. Benjamins, eds., Knowledge Acquisition, Modeling, and Management, 10th European Workshop (EKAW97), Sant Feliu de Guixols, Lecture Notes in Artificial Intelligence 1319, Springer-Verlag, 1997.
OPK
0F
c
0.&:*1S%&
-lTS%9a.&:G4>?0F*
%&N0.&0
33
[46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56]
S
<K )*2.&F
0
"#).&;
>?KMI=&%9.&:7TS
a
_4b >?% @lO%&%9
=&.&a7I=&%&:N
Construction, International Journal of Human-Computer Studies, 46:707-728, 1997. D. Fensel, The Knowledge Acquisition and Representation Language KARL, Kluwer Academic Publ., Boston, 1995. D. Fensel, Formal Specification Languages in Knowledge and Software Engineering, The Knowledge Engineering Review 10, 4, 1995. D. Fensel, J. Angele, and R. Studer, The Knowledge Acquisition and Representation Language KARL, to appear in IEEE Transactions on Knowledge and Data Engineering.
J.-M. David, J.-P. Krivine, and R. Simmons, eds., Second Generation Expert Systems, Springer-Verlag, Berlin, 1993. R. Davis, B. Buchanan, and E.H. Shortcliffe, Production Rules as a Representation for a Knowledge-base Consultation Program, Artificial Intelligence 8 (1977), 15-45. S. Decker, M. Daniel, M. Erdmann, and R. Studer, An Enterprise Reference Scheme for Integrating Model-based Knowledge Engineering and Enterprise Modeling, in: E. Plaza and R. Benjamins, eds., Knowledge Acquisition, Modeling, and Management, 10th European Workshop (EKAW97), Sant Feliu de Guixols, Lecture Notes in Artificial Intelligence 1319, Springer-Verlag, 1997. H. Ehrig and B. Mahr, eds., Fundamentals of Algebraic Specifications 1, Springer-Verlag, Berlin, 1985. H. Ehrig and B. Mahr, eds., Fundamentals of Algebraic Specifications 2, Springer-Verlag, Berlin, 1990. H. Eriksson, A Survey of Knowledge Acquisition Techniques and Tools and their Relationship to Software Engineering, Journal of Systems and Software 19, 1992, 97-107. Epistemics, PCPACK Portable KA Toolkit, 1995. H. Eriksson, A. R. Puerta, and M. A. Musen, Generation of Knowledge Acquisition Tools from Domain Ontologies, Int. J. Human-Computer Studies 41, 1994, 425-453. H. Eriksson, Y. Shahar, S.W. Tu, A.R. Puerta, and M.A. Musen, Task Modeling with Reusable ProblemSolving Methods, Artificial Intelligence 79 (1995), 293-326.
cS
0
%H
')d
+,
-/.&0$00-/L=&.&J.& V%9
0
c.&
:0.&0$.&G4 )*`Y$
.&
D
! cS
0
%
)*Y-H$TSL
;
.&@N.9: 38%&:
0
T2N0=&
-/J8.&=&K)
0%&(OP-/L=90$.&G4
Proceedings 9th Int. Conference on Software Engineering and Knowledge Engineering (SEKE '97), pages 349-357, Madrid, 1997. D. Fensel and A. Schnegge, Using KIV to Specify and Verify Architectures of Knowledge-Based Systems, in: Proceedings of the 12th IEEE International Conference on Automated Software Engineering (ASEC-97), Incline Village, Nevada, November 1997. D. Fensel and R. Straatman, The Essence of Problem-Solving Methods: Making Assumptions for Efficiency Reasons, in: N. Shadbolt et al., eds., Advances in Knowledge Acquisiiton, Lecture Notes in Artificial Intelligence (LNAI) 1076, Springer-Verlag, Berlin, 1996. D. Fensel and F. van Harmelen, A Comparison of Languages which Operationalize and Formalize KADS Models of Expertise, The Knowledge Engineering Review 9, 2, 1994. Proceedings of the Industrial Engineering Research Conference, 1993. AI Magazine, 18(3):53-74, 1997. B. Gaines et al., eds., Working Notes AAAI-97 Spring Symposium Artificial Intelligence in Knowledge Management, Stanford, March 1977. B. Gaines and M.L.G. Shaw, New Directions in the Analysis and Interactive Elicitation of Personal Construct Systems, Int. J. Man-Machine Studies 13 (1980), 81-116. M.R. Genesereth, ed., The Epikit Manual, Epistmemics, Inc, Palo Alto, CA, 1992. M.R. Genesereth and R.E. Fikes, Knowledge Interchange Format, Version 3.0, Reference Manual. Technical Report, Logic-92-1, Computer Science Dept., Stanford University, 1992. http://
[57] [58]
Musen, editors, Proceedings of the 10th Banff Knowledge Acquisition for Knowledge-Based Systems Workshop, pages 5.1-5.18, Alberta, Canada, 1996. SRDG Publications, University of Calgary. http:// ksi.cpsc.ucalgary.ca:80/KAW/KAW96/KAW96Proc.html. D. Fensel, S. Decker, M. Erdmann, and R. Studer, Ontobroker: Transforming the WWW into a Knowledge Base, submitted for publication.
[59]
[60]
T5RS "#OK.&:%9e
WQ2
%& O -/-H0
0EV
%5@g=9KEj?=&L.&0
e.&G4 S.&-/
Nk!O* c
@
>?K TS&=
=& @`=&K =e.&!I$=9%&:Nkc.&:
34
Y-/
[1S
[
W!2
[
WXc '.&;
www.cs.umbc.edu/kse/. M.R. Genesereth, A.M. Keller, and O.M. Duschka, Infomaster: An Information Integration System, in: Proc. ACM SIGMOD Conference, Tucson, 1997. J.H. Gennari, R.B. Altman, and M.A. Musen, Reuse with PROTG-II: From Elevators to Ribosomes, in: Proceedings of the Symposium on Software Reuse, Seattle, 1995. J.H. Gennari, A.R. Stein, and M.A. Musen, Reuse for Knowledge-based Systems and CORBA Components, in: Proc. of the 10th Knowledge Acquisition for Knowledge-based Systems Workshop, Banff, 1996. J.H. Gennari, S.W. Tu, T.E. Rothenfluh, and M.A. Musen, Mappings Domains to Methods in Support of Reuse, Int. J. on Human-Computer Studies 41 (1994), 399-424. Y. Gil and C. Paris, Towards Method-independent Knowledge Acquisition, Knowledge Acquisition 6, 2 (1994), 163-178.
[74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87]
Journal of Human-Computer Studies, 43(2/3):625-640, 1995. D. Harel, Dynamic Logic, in: D. Gabby et al., eds., Handook of Philosophical Logic, vol. II, Extensions of Classical Logic, Publishing Company, Dordrecht (NL), 1984. F. van Harmelen and M. Aben, Structure-preserving Specification Languages for Knowledge-based Systems, International Journal of Human-Computer Studies 44, 1996. F. van Harmelen and J. Balde, (ML)2, A Formal Language for KADS Conceptual Models, Knowledge Acquisition 4, 1, 1992. F. van Harmelen and D. Fensel, Formal Methods in Knowledge Engineering, The Knowledge Engineering Review 9, 2, 1994. F. Hayes-Roth, D.A. Waterman, and D.B. Lenat, Building Expert Systems, Addison-Wesley, New York, 1983 C. B. Jones, Systematic Software Development Using VDM, 2nd ed., Prentice Hall, 1990. G.A. Kelly, The Psychology of Personal Constructs, Norton, New York, 1955. M. Kifer, G. Lausen, and J. Wu, Logical Foundations of Object-Oriented and Frame-Based Languages, Journal of the ACM 42 (1995), 741-843.
Y3%&.&F
O* K%9 Yc
%&%&
-/:
cV
<0 " V;c
-H=&=& \0
%& )
0%9 Programming Constructs, Knowledge Acquisition, 3:117-136, 1991. 3 3.&:K=`
JTQ?F .&%&.&:J!?
:!38%&:
0!@V
;
K.&k>?
0%&
=&.&6.&G4 Proc. AAAI-94, Seattle, 1994. I 3 KiI=&%&:Ni@/=&KVOP0a
=&.&i@/OPL
=&V38%&:D=$OP
F0K
@=c
0.&:B.&G4 "#^_ V0A.9=&Working Papers European Conference on Artificial Intelligence ECAI'94 Workshop
[88]
[89]
[90]
[91]
on Implemented Ontologies, pages 141-152, Amsterdam, 1994. O. Khn and A. Abecker, Corporate Memories for Knowledge Management in Industrial Practice: Prospects and Challenges, J. of Universal Computer Science 3, 8 (August 1977), Special Issue on Information Technology for Knowledge Management, Springer Science Online. URL: http:// www.iicm.edu/jucs_3_8/corporate_memories_for_knowledge. D. Landes, DesignKARL - A Language for the Design of Knowledge-based Systems, in: Proc. 6th International Conference on Software Engineering and Knowledge Engineering (SEKE94), Jurmala, Lettland, 1994, 78-85. D. Landes and R. Studer, The Treatment of Non-Functional Requirements in MIKE, in: W. Schaefer et al., eds., Proc. of the 5th European Software Engineering Conference (ESEC95), Sitges, Lecture Notes in Computer Science 989, Springer-Verlag, 1995. I. van Langevelde, A. Philipsen, and J. Treur, Formal Specification of Compositional Architectures, in:
35
[92]
[93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107]
c ?
=R*)*'YK
)*D;
Y
:^0.9=&KB?II$ O%&
00.&@.&
YEV.&%&%&
Jh I)cj?>B4k I%&.9 ? .&;
%bc
=&
0
Proceedings of the 10th Banff Knowledge Acquisition for Knowledge-Based Systems Workshop, pages 9.1-9.20, Alberta, Canada, 1996. SRDG Publications, University of Calgary. http:// ksi.cpsc.ucalgary.ca:80/KAW/KAW96/KAW96Proc.html. [108] M.A. Musen, An Overview of Knowledge Acquisition, in: J.-M. David et al., eds., Second Generation Expert Systems, Springer-Verlag, 1993. [109] J. Myplopoulos and M. Papazoglu, Cooperative Information Systems, Guest Editors Introduction, IEEE Intelligent Systems 12, 5 (September/October 1997), 28-31. [110] B. Nebel, Artificial Intelligence: A Computational Perspective, in: G. Brewka, ed., Principles of Knowledge Representation, CSLI Publications, Studies in Logic, Language and Information, Stanford, 1996. [111] Knowledge Sharing, AI Magazine, 12(3):36-56, 1991. [112] S. Neubert, Model Construction in MIKE, in: N. Aussenac et al., eds., Knowledge Acquisition for Knowledge-based Systems, Proc. 7th European Workshop (EKAW93), Toulouse, Lecture Notes in Artificial Intelligence 723, Springer-Verlag, 1993. [113] A. Newell, The Knowledge Level, Artificial Intelligence 18, 1982, 87-127. [114] The Essential Distributed Objects Survival Guide, John Wiley & Sons, New York, 1996. [115] Knowledge Modelling with Libraries of Task Decomposition Methods, PhD Thesis, Swedish Institute of Computer Science, 1996. [116]
jBV=&=9*B?
K
%&S12
-/
=&.9;*c0.9:!12%&
- TS%&a.9:S.&G4 )* Y
.&
!#V0
S
.&=&0
Proceedings of the 10th European Conference on Artificial Intelligence (ECAI-92), Vienna, Austria, August, 1992. I. van Langevelde, A. Philipsen, and J. Treur, A Compositional Architecture for Simple Design Formally Specified in DESIRE, in: J. Treur and Th. Wetter, eds., Formal Specification of Complex Reasoning Systems, Ellis Horwood, New York, 1993. D. Lenat and R.V. Guha, Building Large Knowledge-Based Systems: Representation and Inference in the CYC Project, Addison-Wesley Publ. Co., 1990. Representation and Inference in the Cyc Projec, Addison-Wesley, 1990. S. Luke, L. Spector, D. Rager, and J. Hendler, Ontology-based Web Agents, in: Proc. 1st Int. Conf. on Autonomous Agents, 1977. T. J. Lydiard, Overview of Current Practice and Reserach Initiatives for the Verification and Validation of KBS, The Knowledge Engineering Review, 7(2), 1992. SIGART Bulletin, 2(3):70-76, June 1991. N.A.M. Maiden, Acquiring Requirements: a Domain-specific Approach, in: A. Sutcliffe et al., eds., Domain Knowledge for Interactive System Design, Chapman & Hall, 1996. S. Marcus, ed., Automating Knowledge Acquisition for Experts Systems, Kluwer Academic Publisher, Boston, 1988. S. Marcus, SALT: A Knowledge Acquisition Tool for Propose-and-Revise Systems, in: S. Marcus, ed., Automating Knowledge Acquisition for Experts Systems, Kluwer Academic Publisher, Boston, 1988. S. Marcus, J. Stout, and J. McDermott, VT: An Expert Elevator Configurer that Uses Knowledge-based Backtracking, AI Magazine 9 (1), 1988, 95-112. J. McDermott, Preliminary Steps toward a Taxonomy of Problem-solving Methods, in: S. Marcus, ed., Automating Knowledge Acquisition for Experts Systems, Kluwer Academic Publisher, Boston, 1988. P. Meseguer and A. D. Preece, Verification and Validation of Knowledge-Based Systems with Formal Specifications, The Knowledge Engineering Review 10(4), 1995. International Journal of Lexicography, 3(4):235-312, 1990. B. G. Milnes, A Specification of the Soar Cognitive Architecture in Z, Research Report CMU-CS-92169, School of Computer Science, Carnegie Mellon University, Pittsburg, PA, 1992. K. Morik, Underlying Assumptions of Knowledge Acquisition as a Process of Model Refinement, Knowledge Acquisition 2, 1, March 1990, 21-49.
)*
;K
06)djCS.&F
06>BS.&.96>B)*AY6>BTS
=96
Jh )*AT28$
=&=&6j?
%&.&:J>?
;K%&:N
@
)*I@%&.9gc!
FNg
]"#j?80g
.&=&0 3I a 3I a G1S.&;
.&L%&
$@P?.&
.&
$@P>?Fmc
;-HL0.&=&.& V
=&KOP;%&.&@- /OP0
0=&N
36
[117] [118]
[119]
[120] [121]
[122] [123]
"#)P-/:KV %&
KVhl12
-/
%&.9bQj?Nb
Mh ?0
[127]
[138]
in: N. Shadbolt et al., eds., Advances in Knowledge Acquisiiton, Lecture Notes in Artificial Intelligence 1076, Springer-Verlag, Berlin, 1996. C. Pierret-Golbreich and X. Talon: An Algebraic Specification of the Dynamic Behaviour of Knowledge-Based Systems, The Knowledge Engineering Review, 11(2), 1996. J. Penix and P. Alexander, Toward Automated Component Adaption, in: Proceedings of the 9th International Conference on Software Engineering & Knowledge Engineering (SEKE-97), Madrid, Spain, June 18-20, 1997. J. Penix, P. Alexander, and K. Havelund, Declarative Specifications of Software Architectures, in: Proceedings of the 12th IEEE International Conference on Automated Software Engineering (ASEC-97), Incline Village, Nevada, November 1997. R. Plant and A. D. Preece, Special Issue on Verification and Validation, International Journal of HumanComputer Studies (IJHCS), 44, 1996. K. Poeck and U. Gappa, Making Role-Limiting Shells More Flexible, in: N. Aussenac et al., eds., Knowledge Acquisition for Knowledge-Based Systems, Proc. 7th European Knowledge Acquisition Workshop (EKAW93) Toulouse, Lecture Notes in Artificial Intelligence 723, Springer-Verlag, 1993. A. D. Preece, Foundations and Applications of Knowledge Base Verification, International Journal of Intelligent Systems, 9, 1994. A. R. Puerta, J. W. Egar, S. W. Tu, and M. A. Musen, A Multiple-Method Knowledge Acquisition Shell for the Automatic Generation of Knowledge Acquisition Tools, Knowledge Acquisition 4, 1992, 171196. F. Puppe, Systematic Introduction to Expert Systems: Knowledge Representation and Problem-Solving Methods, Springer-Verlag, Berlin, 1993. Object-Oriented Modelling and Design. Prentice Hall, Englewood Cliffs, New Jersey, 1991. F. Saltor, M.G. Castellanos, and M. Garcia-Solaco, Overcoming Schematic Discrepancies in Interoperable Databases, in: D.K. Hsiao et al., eds., Interoperable Database Systems (DS-5), IFIP Transactions A-25, North-Holland, 1993. A.Th. Schreiber, B. Wielinga, H. Akkermans, W. van de Velde, and A. Anjewierden, CML: The CommonKADS Conceptual Modeling Language, in: Steels et al., eds., A Future of Knowledge Acquisition, Proc. 8th European Knowledge Acquisition Workshop (EKAW94), Hoegaarden, Lecture Notes in Artificial Intelligence 867, Springer-Verlag, 1994. A.Th. Schreiber, B. Wielinga, and J. Breuker, eds., KADS. A Principled Approach to Knowledge-Based System Development, Knowledge-Based Systems, vol 11, Academic Press, London, 1993. A.Th. Schreiber, B.J. Wielinga, R. de Hoog, H. Akkermans, and W. van de Velde, CommonKADS: A Comprehensive Methodology for KBS Development, IEEE Expert, December 1994, 28-37. N. Shadbolt, E. Motta, and A. Rouge, Constructing Knowledge-based Systems, IEEE Software 10, 6, 3438. M.L.G. Shaw and B.R. Gaines, The Synthesis of Knowledge Engineering and Software Engineering, in: P. Loucopoulos, ed., Advanced Information Systems Engineering, LNCS 593, Springer-Verlag, 1992. M. Shaw and D. Garlan, Software Architecture: Perspectives on an Emerging Discipline, Prentice Hall, 1996. D. R. Smith, Towards a Classification Approach to Design, in: Proceedings of the 5th International Conference on Algebraic Methodology and Software Technology (AMAST-96), Munich, Germany, July, 1996. J. W. Spee and L. in t Veld, The Semantics of KBSSF: A Language For KBS Design, Knowledge Acquisition 6, 1994. J. M. Spivey, The Z Notation. A Reference Manual, 2nd ed., Prentice Hall, New York, 1992. L. Steels, The Componential Framework and its Role in Reusability, in: David et al., eds., Second Generation Expert Systems, Springer-Verlag, Berlin, 1993. R. Studer, H. Eriksson, J.H. Gennari, S.W. Tu, D. Fensel, and M.A. Musen, Ontologies and the Configuration of Problem-Solving Methods, in: Proc. of the 10th Knowledge Acquisition for Knowledgebased Systems Workshop, Banff, 1996.
)dBY
.&
f
BV .9=&0 Proceedings of the 10th Banff Knowledge Acquisition for
[139] [140]
=&
7>6
. +,
[141] [142]
\0;
K%& Y.&:
I=&%9:. 1S.&;.9L%& D=&K0
LL%&.&;
=&.90 '%9=&]
O* ;F
K@@dI:.9[
=&.&
DY.&
;R4B ?.&N @k$000-H=HV
%&0d.9G4B
Knowledge-Based Systems Workshop, pages 32.1-32.19, Alberta, Canada, 1996. SRDG Publications, University of Calgary. http://ksi.cpsc.ucalgary.ca:80/KAW/KAW96/KAW96Proc.html. Automated Configuration of Problem Solving Methods in Diagnosis, PhD Thesis, University of Amsterdam, Amsterdam, The Netherlands, 1997. P. Terpstra, G. van Heijst, B. Wielinga, and N. Shadtbolt, Knowledge Acquisition Support through Generalised Directive Models, in: J.-M. David et al., eds., Second Generation Expert Systems, SpringerVerlag, Berlin, 1993. Tove, Manual of the Toronto Virtual Enterprise, Technical Report, University of Toronto, 1995. Available at: www.ie.utoronto.ca/EIL/tove/ontoTOC.html. J. Treur, Temporal Semantics of Meta-Level Architectures for Dynamic Control of Reasoning, in: L. Fribourg et al., eds., Logic Program Synthesis and Transformation - Meta Programming in Logic, Proceedings of the 4th International Workshops, LOPSTER-94 and META-94, Lecture Notes in Computer Science 883, Springer Verlag-Berlin, 1994. J. Treur and Th. Wetter, eds., Formal Specification of Complex Reasoning Systems, Ellis Horwood, New York, 1993. Knowledge Engineering Review, 11(2):93-155, 1996.
[146]
15j a
*W'=&`1Q TSL
%&`
f"#^_GV
0P>?Km1S%&.&.9I=&%&:Nb@dOP
-/.&;WV
=&
.9%&0`.&G4e"#^_ Ya
. +,0=& Ya
i. + 0=&B>?KG?TS;
K.&
B
"#?hi.&
%&.&:
B\0.&:ij L%&.9;.&= I=&%&:.&
.&i3 T c
a%&L-H=&
Aussenac et al., eds., Knowledge Acquisition for Knowledge-based Systems, Proc. 7th European Workshop (EKAW93), Toulouse, Lecture Notes in Artificial Intelligence 723, Springer-Verlag, 1993.
Mars, editor, Working Papers European Conference on Artificial Intelligence ECAI'94 Workshop on Implemented Ontologies, pages 187-206, Amsterdam, 1994. [147] The Role of Ontologies in Knowledge Engineering, PhD Thesis, University of Amsterdam, May 1995. [148] International Journal of Human-Computer Studies, 46(2/3):183-292, 1997. [149] G. van Heijst, R. van der Spek, and E. Kruizinga, Organizing Corporate Memories, in: Proc. of the 10th Knowledge Acquisition for Knowledge-based Systems Workshop, Banff, 1996. [150] IEEE Computer, 25(3):3849, 1992. [151] Journal of Intelligent Information Systems, Special Issue on Intelligent Integration of Information, 1996. [152] G. Wiederhold and M. Genesereth, The Conceptual Basis for Mediation Services, IEEE Intelligent Systems 12, 5 (September/October 1997), 38-47. [153] B.J. Wielinga, A.Th. Schreiber, and J.A. Breuker, KADS: A Modelling Approach to Knowledge Engineering, Knowledge Acquisition 4, 1 (1992), 5-53. [154] M. Wirsing, Algebraic Specification, in: J. van Leeuwen, ed., Handbook of Theoretical Computer Science, Elsevier Science Publ., 1990. [155] E. Yourdon, Modern Structured Analysis, Prentice-Hall, Englewood Cliffs, 1989.
Yhi.&
K%&RV
.&=9`.&H=&K ;
K.&=&
;=9 @eS=& ^@-/
=&.&HT2N0=&
-/0 Yhi.&
K%&^=9%&%9.&:=R^=&
:=9.&*@^@-H=&.9
38