Return to search

Learning by Failing to Explain

Explanation-based Generalization requires that the learner obtain an explanation of why a precedent exemplifies a concept. It is, therefore, useless if the system fails to find this explanation. However, it is not necessary to give up and resort to purely empirical generalization methods. In fact, the system may already know almost everything it needs to explain the precedent. Learning by Failing to Explain is a method which is able to exploit current knowledge to prune complex precedents, isolating the mysterious parts of the precedent. The idea has two parts: the notion of partially analyzing a precedent to get rid of the parts which are already explainable, and the notion of re-analyzing old rules in terms of new ones, so that more general rules are obtained.

Identiferoai:union.ndltd.org:MIT/oai:dspace.mit.edu:1721.1/6850
Date01 May 1986
CreatorsHall, Robert Joseph
Source SetsM.I.T. Theses and Dissertation
Languageen_US
Detected LanguageEnglish
Format140 p., 15467251 bytes, 5755509 bytes, application/postscript, application/pdf
RelationAITR-906

Page generated in 0.0022 seconds