Stage-oe-small.jpg

Inproceedings618

Aus Aifbportal
Wechseln zu:Navigation, Suche


Are Substitutions the Better Examples? Learning Complete Set of Clauses with Frog


Are Substitutions the Better Examples? Learning Complete Set of Clauses with Frog



Published: 1995
Herausgeber: L. De Raedt
Buchtitel: Proceedings of the 5th International Workshop on Inductive Logic Programming ILP-95, Leuven, September 4-6, 1995, Technical Report, Department of Computer Science, Katholieke Universiteit Leuven, Belgium

Referierte Veröffentlichung

BibTeX

Kurzfassung
The paper presents an approach for machine learning in a restricted first-order language with finite minimal Herbrand models by means of a search through a propositional representation space. The learning target is to find a set of goal clauses which can be used to define a target predicate. That is, we deal with single-predicate learning. For the search process we use the learning algorithm JoJo/Frog which provides a flexible search strategy. The transition from the first-order representation to the representation in propositional logic is achieved by ground substitutions which transform clauses into ground clauses. Taking a closer look at this transition makes clear that the sufficiency condition which is used by algorithms like FOIL and LINUS as a criterion for judging the achieved learning results does not correspond to the completeness condition in the propositional case. Therefore, we use an extended completeness condition which captures all information given by the example knowledge. As a consequence we get a new definition of positive and negative examples. Instead of ground facts we regard ground substitutions as examples.

Download: Media:1995_618_Fensel_Are_Substitutio_1.ps.gz



Forschungsgruppe

Wissensmanagement


Forschungsgebiet