Using Knowledge in Its Context: Report on the IJCAI-93 Workshop
Brezillon, Patrick, Abu-Hakima, Suhayya
The Workshop on Using Knowledge in Its Context was held in Chambery, France, on 28 August 1993, preceding the Thirteenth International Joint Conference on Artificial Intelligence (IJCAI-93). This article provides a summary of the discussions between the participants before (by e-mail) and during the one-day workshop. It is clear from these discussions that the notion of context is far from defined and is dependent in its interpretation on a cognitive science versus an engineering (or system building) point of view. In identifying the two points of view, this workshop permitted us to go one step further than previous workshops (notably Maskery and Meads [1992] and Maskery, Hopkins, and Dudley [1992]). Once a distinction is made on the viewpoint, one can achieve a surprising consensus on the aspects of context that the workshop addressed -- mainly, the position, the elements, the representation, and the use of context. Despite this consensus on the aspects of context, agreement on the definition of context was not yet achieved.
Routine Design for Mechanical Engineering
Brinkop, Axel, Laudwein, Norbert, Maasen, Rudiger
COMIX (configuration of mixing machines) is a system that assists members of the EKATO Sales Department in designing a mixing machine that fulfills the requirements of a customer. It is used to help the engineer design the requested machine and prepare an offer that's to be submitted to the customer. comix integrates more traditional software techniques with explicit knowledge representation and constraint propagation. During the process of routine design, some design decisions have to be made with uncertainty. By including knowledge from process technology and company experience in the mechanical design, a sufficiently high degree of flexibility is achieved that the system can even assist in difficult design situations. The success of the system can be measured by the increase in the quantity and the quality of the submitted offers.
The Simon Newcomb Awards
Hayes, Patrick J., Ford, Kenneth M.
We also was the Stopping Problem argument. Achievement Award' in recognition aeroplane," suggested Newcomb sarcastically, This means, in effect, sentence. Most versions of this argument that the hardware must be - Joseph Rychlak of Loyola University miss these subtle aspects and accommodated as information is of Chicago (for his exclusive-OR are therefore simply invalid, but Penrose processed. Hubert has been accessible via the intellect only. Much of the content of with it...mathematicians interpreted in binary fashion: these books consists of the kind of communicate...by each one having "either x or y, but not both."
On the Informativeness of the DNA Promoter Sequences Domain Theory
The DNA promoter sequences domain theory and database havebecome popular for testing systems that integrate empirical andanalytical learning. This note reports a simple change andreinterpretation of the domain theory in terms of M-of-N concepts,involving no learning, that results in an accuracy of 93.4% on the 106items of the database. Moreover, an exhaustive search of the space ofM-of-N domain theory interpretations indicates that the expectedaccuracy of a randomly chosen interpretation is 76.5%, and that amaximum accuracy of 97.2% is achieved in 12 cases. This demonstratesthe informativeness of the domain theory, without the complications ofunderstanding the interactions between various learning algorithms andthe theory. In addition, our results help characterize the difficultyof learning using the DNA promoters theory.
Cost-Sensitive Classification: Empirical Evaluation of a Hybrid Genetic Decision Tree Induction Algorithm
This paper introduces ICET, a new algorithm for cost-sensitive classification. ICET uses a genetic algorithm to evolve a population of biases for a decision tree induction algorithm. The fitness function of the genetic algorithm is the average cost of classification when using the decision tree, including both the costs of tests (features, measurements) and the costs of classification errors. ICET is compared here with three other algorithms for cost-sensitive classification - EG2, CS-ID3, and IDX - and also with C4.5, which classifies without regard to cost. The five algorithms are evaluated empirically on five real-world medical datasets. Three sets of experiments are performed. The first set examines the baseline performance of the five algorithms on the five datasets and establishes that ICET performs significantly better than its competitors. The second set tests the robustness of ICET under a variety of conditions and shows that ICET maintains its advantage. The third set looks at ICET's search in bias space and discovers a way to improve the search.
Truncating Temporal Differences: On the Efficient Implementation of TD(lambda) for Reinforcement Learning
Temporal difference (TD) methods constitute a class of methods for learning predictions in multi-step prediction problems, parameterized by a recency factor lambda. Currently the most important application of these methods is to temporal credit assignment in reinforcement learning. Well known reinforcement learning algorithms, such as AHC or Q-learning, may be viewed as instances of TD learning. This paper examines the issues of the efficient and general implementation of TD(lambda) for arbitrary lambda, for use with reinforcement learning algorithms optimizing the discounted sum of rewards. The traditional approach, based on eligibility traces, is argued to suffer from both inefficiency and lack of generality. The TTD (Truncated Temporal Differences) procedure is proposed as an alternative, that indeed only approximates TD(lambda), but requires very little computation per action and can be used with arbitrary function representation methods. The idea from which it is derived is fairly simple and not new, but probably unexplored so far. Encouraging experimental results are presented, suggesting that using lambda > 0 with the TTD procedure allows one to obtain a significant learning speedup at essentially the same cost as usual TD(0) learning.
A Domain-Independent Algorithm for Plan Adaptation
The paradigms of transformational planning, case-based planning, and plan debugging all involve a process known as plan adaptation - modifying or repairing an old plan so it solves a new problem. In this paper we provide a domain-independent algorithm for plan adaptation, demonstrate that it is sound, complete, and systematic, and compare it to other adaptation algorithms in the literature. Our approach is based on a view of planning as searching a graph of partial plans. Generative planning starts at the graph's root and moves from node to node using plan-refinement operators. In planning by adaptation, a library plan - an arbitrary node in the plan graph - is the starting point for the search, and the plan-adaptation algorithm can apply both the same refinement operators available to a generative planner and can also retract constraints and steps from the plan. Our algorithm's completeness ensures that the adaptation algorithm will eventually search the entire graph and its systematicity ensures that it will do so without redundantly searching any parts of the graph.
Solving Multiclass Learning Problems via Error-Correcting Output Codes
Multiclass learning problems involve finding a definitionfor an unknown function f(x) whose range is a discrete setcontaining k > 2 values (i.e., k ``classes''). Thedefinition is acquired by studying collections of training examples ofthe form [x_i, f (x_i)]. Existing approaches tomulticlass learning problems include direct application of multiclassalgorithms such as the decision-tree algorithms C4.5 and CART,application of binary concept learning algorithms to learn individualbinary functions for each of the k classes, and application ofbinary concept learning algorithms with distributed outputrepresentations. This paper compares these three approaches to a newtechnique in which error-correcting codes are employed as adistributed output representation. We show that these outputrepresentations improve the generalization performance of both C4.5and backpropagation on a wide range of multiclass learning tasks. Wealso demonstrate that this approach is robust with respect to changesin the size of the training sample, the assignment of distributedrepresentations to particular classes, and the application ofoverfitting avoidance techniques such as decision-tree pruning.Finally, we show that---like the other methods---the error-correctingcode technique can provide reliable class probability estimates.Taken together, these results demonstrate that error-correcting outputcodes provide a general-purpose method for improving the performanceof inductive learning programs on multiclass problems.
Probabilistic Anomaly Detection in Dynamic Systems
Padhraic Smyth Jet Propulsion Laboratory 238-420 California Institute of Technology 4800 Oak Grove Drive Pasadena, CA 91109 Abstract This paper describes probabilistic methods for novelty detection when using pattern recognition methods for fault monitoring of dynamic systems. The problem of novelty detection is particularly acutewhen prior knowledge and training data only allow one to construct an incomplete classification model. Allowance must be made in model design so that the classifier will be robust to data generated by classes not included in the training phase. For diagnosis applications one practical approach is to construct both an input density model and a discriminative class model. Using Bayes' rule and prior estimates of the relative likelihood of data of known and unknown origin the resulting classification equations are straightforward.