Abstracts of PhD theses related to ILP


Author:

C.H.Bryant

PhD Thesis Title:

Data Mining for Chemistry: the Application of Three Machine Induction Tools to a Database of Enantioseparations

University:

University of Manchester Institute of Science and Technology (UMIST), UK

Year:

1996

Abstract:

A validated first step is taken towards meeting the need for a symbolic expert system for enantioseparations by High-Performance Liquid Chromatography on chiral stationary phases (CSPs). Three data mining tools are applied to data from a database of published enantioseparations performed on commercially available Pirkle CSPs with the aim of inducing generalisations that recommend particular CSP chiral selectors based on the structural features of an enantiomer pair. Two of the tools, Golem and Progol, are from the field of inductive logic programming (ILP) and are available in the public domain. The other, DataMariner, is a commercially available tool which generates essentially propositional rules.
The ability of each of these tools to induce useful generalisations from the database is revealed by investigating whether each one can make generalisations about the values of attributes, attributes themselves, relationships between attributes and negations of literals. The usefulness of generalisations is measured by their accuracy, size, ease of interpretation and chemical justification.
The ability of DataMariner to make generalisations about the values of attributes and negations of literals is shown to be sufficient for the induction of accurate rules from the database: the accuracy of the most useful rule-set induced by DataMariner is 63% +/-3% which is more than ten times greater than the accuracy that would have resulted from a random choice. However it is difficult to provide a chemical justification for the rules. Golem can induce clauses that represent knowledge about relationships between attributes in the database which can not be represented by the essentially propositional rules generated by DataMariner. A scheme devised for Progol, which allows generalisations to be made about both the attributes in the database and their values, enables Progol to induce clauses which are more concise and easier to justify chemically than those which can be induced by DataMariner. Progol uses a substantial amount of general chemical knowledge when making these generalisations. This supports the claim made by the proponents of ILP that one of the advantages of ILP over classical machine induction is that ILP allows substantial background knowledge to be used during induction.


Author:

Gunther Sablon

PhD Thesis Title:

Iterative versionspaces with an application in inductive logic programming

University:

Katholieke Universiteit Leuven, Belgium

Year:

1995

Abstract:

This thesis consists of two parts: in the first part we develop a language-independent framework for efficiently solving concept learning problems; in the second part we apply this framework to Inductive Logic Programming.
We view concept learning as a search problem. In the well-known framework of Versionspaces a bi-directional depth-first search algorithm that learns a maximally general and maximally specific concept representation is presented, and contrasted to the breadth-first approach of Mellish's Description Identification algorithm. In this context, we identify redundant information elements, in order to reduce the memory needed for storing information elements. We describe how automatically generated information elements can replace less informative ones.
Next, we extend this framework to describe the more complex versionspaces that originate from introducing disjunctions. To be practically useful, we gradually restrict these disjunctive versionspaces by imposing preference criteria, based on notions of minimality. This leads to extensions of the non-disjunctive algorithms to the disjunctive case.
In the second part of the thesis we show in detail how this general framework can be instantiated to Inductive Logic Programming. In this respect we also discuss the integration of machine learning in a planning system based on Horn clause logic. This illustrates that the use of a logical representation allows a smooth integration of Machine Learning and Problem Solving.
In summary, the thesis contributes to the understanding and the development of search algorithms for concept learning in general, by developing a language independent framework, and by introducing several novel and generally applicable concept learning techniques. The application in the second part of the thesis shows that the framework is practically useful, and that it contributes to the field of Inductive Logic Programming.


Author:

Patrick van der Laag

Title:

An analysis of refinement operators in inductive logic programming

University:

Erasmus Universiteit, Rotterdam, the Netherlands

Year:

1995

Abstract:

This thesis develops a classification theory for search operators employed in incremental learning systems. It characterizes specialization and generalization operators in terms of their effectiveness and efficiency, and investigates the problems of combining these desirable properties. Applying this theory to the relational, first-order representations of Inductive Logic Programming (ILP), it provides positive and negative results concerning effective and efficient learning in practical ILP-systems.


Author:

Aram Karalic

PhD Thesis Title:

First order regression

University:

University of Ljubljana, Slovenia

Year:

1995

Abstract:

A system named FORS (First Order Regression System) is developed, capable of inducing numerical concepts in first order logic. FORS utilizes some useful ILP concepts (such as background knowledge), alleviates some shortcomings of the existing propositional regression systems, and integrates some of their advantages.
Our goals were do develop a system which can: FORS was successfully applied in several synthetic and real-world domains where the requirements proved necessary and useful.
FORS constructs a model in the form of a Prolog program. Covering approach, similar to the one of FOIL is used. The clause building part of the algorithm uses a top-down approach. The algorithm starts with the most general candidate clause, covering the entire example set and then specializes the clause by adding literals. Clause construction uses beam search to guide the algorithm through the space of possible clauses.
As a part of the system, the pruning based on the Minimum description length principle was developed that can handle also continuous variables. It turned out that MDL pruning helps to build more comprehensible models, while at the same time preserves model's performance in terms of its prediction power.
In experiments with physical domains FORS rediscovered Kepler's third law of planetary motion and ideal gas law (including rediscovery of the gas constant and the absolute temperature scale).
In all the real-world domains, models using linear regression appealed most to the experts, since linear regression largely increased the expressive power of the models in comparison with pure first order logic. The experts were able to thoroughly analyse the induced models and carefully exploit the selected regions of attribute space and more reliably evaluate model's quality inside the regions.
Modelling of water behavior in surge tank proved that FORS can successfully handle times series data. Additionally, FORS' ability to partition the attribute space coupled with its linear regression capabilities proved to be crucial in inducing a useful model without prior knowledge of ``absolute value'' function.
Models induced from the Lake of Bled data adequately well describe the growth of algae in the lake. Newly induced background literals defining seasons help in better comprehensibility of the induced models. For the expert this was not the first approach to modelling the Lake of Bled behavior and experiments with FORS ~have confirmed his opinion that not much more can be done with current data without additional more precise and frequent measurements.
A domain expert in the domain of steel grinding claims to be satisfied with the results of the machine learning, since our models enabled him to grasp some additional process properties, which he wouldn't be able to discover only with classical statistical tools. However, he suggests that the machine learning approach should not be used alone, but should be considered as a powerful supplement to already existent instruments. In this domain it also turned out that simple background knowledge can sometimes significantly improve the usefulness of induced models.
FORS was also used in the domains of finite element mesh design and for construction of rules for the prediction of the mutagenic activity of nitroaromatic compounds, where its performance was at the level of other machine learning algorithms, used in those domains.
During the work in the electrical discharge machining process, a data acquisition environment was established which enabled us to monitor and record crucial process parameters as well as the operator's control actions. Models were induced which, according to the expert, capture the main behaviour patterns of the operator. During the knowledge acquisition process several important guidelines for knowledge acquisition, concerning mainly the process of interaction with the domain experts, emerged, confirming that comprehensibility of the induced model plays an important role in the process of behaviour cloning, and again confirming, that for successful modelling of the expert behavior one can not rely only on interviews with an expert.


Author:

Saso Dzeroski

Title:

Numerical constraints and learnability in inductive logic programming

University:

University of Ljubljana, Slovenia

Year:

1995

Abstract:

Inductive logic programming is a part of machine learning, a subarea of artificial intelligence, concerned with the induction of first-order clausal theories from examples and background knowledge. Two main approaches can be distinguished within ILP: the normal and nonmonotonic semantics. While several ILP systems have been developed for each of the two semantics, efficiency problems and problems with handling real numbers, among others, have hindered practical applications.
We first present a new formalization of the nonmonotonic ILP setting, where examples are interpretations and concepts are clausal theories. Examples may be also given in the form of definite logic programs, from which interpretations are obtained as minimal Herbrand models. A background theory may be taken into account. This formalization makes nonmonotonic ILP a natural extension of propositional learning where a first-order representation and background knowledge are used.
We then present a study of the computational complexity of inductive logic programming. We introduce PAC-learnability models for both semantics and prove positive learnability results in both. Results concerning learning in the presence of noise are also presented. The results for normal ILP are based on the transformation of ILP problems to propositional form performed by DINUS. The results for nonmonotonic ILP are based on our new formalization. The concept classes considered in our results can be induced by existing ILP systems, such as GOLEM and CLAUDIEN.
Finally, we develop methods for handling real numbers and numerical constraints in both semantics. The DINUS system can be used for learning relational rules with real numbers and relational regression in normal ILP. We extend number handling in CLAUDIEN, adding the capability of using linear equations, thus combining quantitative and qualitative discovery. We finally present LAGRANGE, a machine discovery system tightly related to the two ILP semantics. From example behaviors, LAGRANGE discovers differential equations that describe the dynamics of a given system.