9
Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory Wolfson Building, Parks Road Oxford, OX1 3QD. Inductive Logic Progrdng (ILP) involves the construction of first-order definite clause theories from examples and background knowledge, Un- like both traditional Machine Learning and Com- putational Learning Theory, ILP is based on lock- step development of Theory, Implementations and Applications. ILP systems have successful appli- cations in the learning of structure-activity rules for drug design, semantic grammars rules, finite element mesh design rules and rules for prediction of protein structure and mutagenic molecules. The strong applications in ILP can be contrasted with relatively weak PAC-learning results (even highly- restricted forms of logic programs are known to be prediction-hard). It has been recently argued that the mismatch is due to distributional assump- tions made in application domains. These assump- tions can be modelled as a Bayesiau prior prob- ability representing subjective degrees of belief. Other authors have argued for the use of Bayesian prior distributions for reasons different to those here, though this has not lead to a new model of polynomial-time learnability. Incorporation of Bayesian prior distributions over time-bounded hy- potheses in PAC leads to a new model called U- Learnability. It is argued that U-learnability is more appropriate than PAC for Universal (Tur- ing computable) languages. Time-bounded logic programs have been shown to be polynomially U-learnable under certain distributions. The use of time-bounded hypotheses enforces decidability and allows a unified characterisation of speed-up learning and inductive learning. U-learnability has as special cases PAC and Natarajan’s model of speed-up learning. Permission to copy without fee all or part of this material is granted provided that the copies are not made or distributed for direct commercial advantage, the ACM copyright notice and the title of the publication and its date appear, and notice is given that copying is by permission of the Association of Computing Machinery. To copy otherwise, or to republish, requires a fee and/or specific permission. COLT 94- 7/94 New Brunswick, N..J. USA @ 1994 ACM 0-89791-655-7/94/0007..$3.50 1 INTRODUCTION This paper is written for two separate audiences: 1) a Ma- chine Learning audience, interested in implementations and applications and 2) a Computational Learning Theory audl- ence interested in models and results in learnability. The paper should thus be taken as two loosely connected papers with a common theme involving the use of distributional information in Inductive Logic Programming (ILP). In the last few years ILP [23, 24,20, 31] has developed from a theoretical backwater to a rapidly growing area in Machine Learning. This is in part due to the fact that pure logic programs provide Machine Learning with a representation which is general purpose (Turing-computable), and has a simple, and rigorously defined, semantics. In addition, logic programs tend to be easy to comprehend as hypotheses in scientific domains. ILP is based on interaction between Theory, Implementations and Applications. The author has argued [25] that the devel- opment of a formal semantics of ILP (see Section 2) allows the direct derivation of ILP algorithms. Such derivational techniques have been at the centre of specific-to-general ILP techniques such as Plotkin’s least general generalisation [37, 36], Muggleton and Buntine’s inverse resolution [27, 23] and Idestam-Almquist’s inverse implication [ 15]. TWO con- tending semantics of ILP have been developed [31]. These are the so-called ‘open-world’ semantics [31 ] and ‘closed- world’ semantics [13]. A number of efficient ILP systems have been developed, in- cluding FOIL [38], Golem [28], LINUS [21], CLAUDIEN [39] and Progol [26]. The use of a relational logic formalism has allowed successful application of ILP systems in a num- ber of domains in which the concepts to be learned cannot easily be described in an attribute-value, propositional-level, language. These applications (see Section 3) include struc- ture activity prediction for drug design [19, 44, 43], protein secondary-structure prediction [29], finite-element mesh de- sign [9] and learning semantic grammars [45]. A variety of positive and negative PAC-learnability results ex- ist for subsets of definite clause logic [11, 34, 10, 18, 7, 40]. However, in contrast to experimentally demonstrated abili- ties of ILP systems in applications, the positive PAC results are rather weak, and even highly restricted forms of logic 3

Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

Bayesian Inductive Logic Programming

Abstract

Stephen MuggletonOxford University Computing Laboratory

Wolfson Building, Parks Road

Oxford, OX1 3QD.

Inductive Logic Progrdng (ILP) involves theconstruction of first-order definite clause theories

from examples and background knowledge, Un-like both traditional Machine Learning and Com-putational Learning Theory, ILP is based on lock-step development of Theory, Implementations and

Applications. ILP systems have successful appli-cations in the learning of structure-activity rulesfor drug design, semantic grammars rules, finiteelement mesh design rules and rules for predictionof protein structure and mutagenic molecules. Thestrong applications in ILP can be contrasted withrelatively weak PAC-learning results (even highly-restricted forms of logic programs are known to

be prediction-hard). It has been recently arguedthat the mismatch is due to distributional assump-tions made in application domains. These assump-tions can be modelled as a Bayesiau prior prob-ability representing subjective degrees of belief.Other authors have argued for the use of Bayesianprior distributions for reasons different to thosehere, though this has not lead to a new modelof polynomial-time learnability. Incorporation ofBayesian prior distributions over time-bounded hy-potheses in PAC leads to a new model called U-Learnability. It is argued that U-learnability ismore appropriate than PAC for Universal (Tur-ing computable) languages. Time-bounded logicprograms have been shown to be polynomiallyU-learnable under certain distributions. The useof time-bounded hypotheses enforces decidabilityand allows a unified characterisation of speed-uplearning and inductive learning. U-learnability hasas special cases PAC and Natarajan’s model ofspeed-up learning.

Permission to copy without fee all or part of this material isgranted provided that the copies are not made or distributed fordirect commercial advantage, the ACM copyright notice and thetitle of the publication and its date appear, and notice is giventhat copying is by permission of the Association of ComputingMachinery. To copy otherwise, or to republish, requires a feeand/or specific permission.COLT 94- 7/94 New Brunswick, N..J. USA@ 1994 ACM 0-89791-655-7/94/0007..$3.50

1 INTRODUCTION

This paper is written for two separate audiences: 1) a Ma-chine Learning audience, interested in implementations andapplications and 2) a Computational Learning Theory audl-ence interested in models and results in learnability. Thepaper should thus be taken as two loosely connected paperswith a common theme involving the use of distributionalinformation in Inductive Logic Programming (ILP).

In the last few years ILP [23, 24,20, 31] has developed froma theoretical backwater to a rapidly growing area in MachineLearning. This is in part due to the fact that pure logicprograms provide Machine Learning with a representationwhich is general purpose (Turing-computable), and has asimple, and rigorously defined, semantics. In addition, logicprograms tend to be easy to comprehend as hypotheses inscientific domains.

ILP is based on interaction between Theory, Implementationsand Applications. The author has argued [25] that the devel-opment of a formal semantics of ILP (see Section 2) allowsthe direct derivation of ILP algorithms. Such derivationaltechniques have been at the centre of specific-to-generalILP techniques such as Plotkin’s least general generalisation[37, 36], Muggleton and Buntine’s inverse resolution [27, 23]and Idestam-Almquist’s inverse implication [ 15]. TWO con-tending semantics of ILP have been developed [31]. Theseare the so-called ‘open-world’ semantics [31 ] and ‘closed-world’ semantics [13].

A number of efficient ILP systems have been developed, in-cluding FOIL [38], Golem [28], LINUS [21], CLAUDIEN

[39] and Progol [26]. The use of a relational logic formalismhas allowed successful application of ILP systems in a num-ber of domains in which the concepts to be learned cannoteasily be described in an attribute-value, propositional-level,language. These applications (see Section 3) include struc-ture activity prediction for drug design [19, 44, 43], proteinsecondary-structure prediction [29], finite-element mesh de-sign [9] and learning semantic grammars [45].

A variety of positive and negative PAC-learnability results ex-ist for subsets of definite clause logic [11, 34, 10, 18, 7, 40].However, in contrast to experimentally demonstrated abili-ties of ILP systems in applications, the positive PAC resultsare rather weak, and even highly restricted forms of logic

3

Page 2: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

programs have been shown to be prediction hard [7]. Likemany other results in PAC-learning, positive results are onlyachieved for ILP by setting language parameters to constantvalues (eg. k-clauses and l-literals). This provides ‘hard’boundaries to the hypothesis space, which often reducesthe representation to a propositional level. An alternativemodel, U-learnability [30], is better suited to Universal (Tur-ing computable) representations, such as logic programs. U-learnability provides a ‘soft’ boundary to hypothesis spaces inthe form of a prior probability distribution over the completerepresentation (eg. time-bounded logic programs). Other

authors have argued for the use of Bayesian prior distribu-tions for reasons different to those here, though this has notlead to a new model of polynomial-time learnability. U-Iearnability allows standard distributional assumptions, suchas Occam’s razor, to be used in a natural manner, withoutparameter settings. Although Occam algorithms have beenstudied in PAC-learning [3], this has not led to algorithmscapable of learning in universal representations, as it has inInductive Logic Programming. In the early 1980’s a distribu-tion assumption very different to Occam’s razor was implic-itly implemented by Arbab, Michie [1] and Bratko [4] in analgorithm for constructing decision trees. The algorithm con-structs a decision list (linear decision tree) where one existsand otherwise returns the most linear decision tree which canbe constructed from the data. This distributional assumptionis based on the fact that decision lists are generally easier tocomprehend than arbitrary decision trees. Other interestingkinds of distributions along these lines can be imagined; as-signing higher prior probabilities to grammars that are regularor almost so, logic programs that are deterministic or almostso and logic programs that run in linear time or almost so.One can imagine very different kinds of prior distributionson hypotheses. For example, when learning concepts whichmimic human performance in skill tasks [41] predictive ac-curacy is dependent on hypotheses being evaluable in timesimilar to that of human reaction, and so such hypothesesshould be preferred a priori.

This paper is organised as follows. Section 2 gives a formaldefinition of ILP in terms of Bayesian inference. Section 3provides an overview of ILP applications in molecular bi-

ology and discusses some of the distributional assumptionsused in these applications. Section 4 defines U-learnabilityand describes some general results on U-learnable distribu-tions.

2 BAYES’ ILP DEFINITIONS

Familiarity with standard definitions from Logic Program-ming [22, 14] is assumed in the following. A Bayesian ver-sion of the usual (open world semantics) setting for ILP is asfollows. Suppose T, 7 and V are sets of predicate symbols,function symbols and variables and cd and Ch are the classesof definite and Horn clauses constructed from T, 7 and V.The symbol En denotes SLDNF derivation bounded by n res-olution steps. An ILP learner is provided with backgroundknowledge B z Cd and examples E = (E+, E- ) in whichE+ C cd are positive examples and E- C (Ch – cd) arenegat~ve examples. Each hypothesis is a p~r H. = (H, n)

where H Q cd, H 1= B and n is a a natural number. Hnis said to hold for examples E, or holds(Hn, E), when boththe following are truel.

Sufficiency: H h. E+

Satisfiability: H A E- tjn ❑

The prior probability, p(Hn), of an hypothesis is defined bya given distribution D. According to Bayes’ theorem, theposterior probability is

P(~nl~)=p(EIHn).p(Hn)

p(E)

where p(E[Hn) is 1 when holds(Hn, E) and O otherwiseand

p(E) = ~ p(H:)

holds (H& ,E)

Well known strategies for making use of distributional infor-mation include a) maximisation of posterior probability b)class prediction based on posterior weighted sum of predic-tions of all hypotheses (Bayes optimal strategy) c) randomlysampling an hypothesis from the posterior probability (Gibbsalgorithm). The sample complexities of strategies b) and c)are analysed in [12].

Although no existing ILP system takes an hypothesis distri-

bution D as an input parameter, such distributions are implicitin FOIL’s [38] information gain measure and Golem’s [32]compression measure. Additionally, search strategies suchas general-to-specific or specific-to-general, use an implicitprior distribution which favours more general hypotheses or,conversely, more specific ones.

Bayesian approaches to Machine Learning have been dis-cussed previously in the literature. For instance, Buntine[6] develops a Bayesian framework for learning, which heapplies to the problem of learning class probability trees.Also Haussler et al. [12] analyse the sample complexity oftwo Bayesian algorithms. This analysis focuses on averagecase, rather than worst case, accuracy. On the other handU-learnability uses average case time complexity and worstcase accuracy. Also unlike U-learnability (Section 4) neither

Buntine nor Haussler et al, develop a new learning modelwhich allows significantly larger polynomially-learnable rep-resentations, such as logic programs. The applications in thefollowing section show that ILP systems are effective at learn-ing logic programs in significant real-world domains. Thisis consistent with the fact that time-bounded logic programsare U-learnable under certain distributions (see Section 4).

3 APPLICATIONS IN MOLECULARBIOLOGY

ILP applications in Molecular Biology extend a long tradi-tion in approaches to scientific discovery that was initiated byMeta-Dendral’s [5] application in mass spectrometry. Molec-ular biology domains are particularly appropriate for ILP due

1Though most ILp systems can deal with noise, this is ignored

in the above definition for the sakeof simplicity.

4

Page 3: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

to the rich relational structure of the data. Such relationshipsmake it hard to code feature-based representations for thesedomains. ILP’s applications to protein secondary structureprediction [29], structure-activity prediction of drugs [19] andmutagenicity [42] of molecules have produced new knowl-edge, published in respected scientific journals in their area.

3.1 PROTEIN SECONDARY STRUCTUREPREDICTION

When a protein (amino acid sequence) shears from the RNAthat codes it, the molecule convolves in a complex thoughdeterministic manner. However, it is largely the final shape,known as secondary and tertiary structure, which determinesthe protein’s function. Determination of the shape is a longand Iabour intensive procedure involving X-ray crystallogra-phy. On the other hand the complete amino acid sequence isrelatively easy to determine and is known for a large numberof naturally occurring proteins. It is therefore of great in-terest to be able to predict a protein’s secondary and tertiarystructure directly from its amino acid sequence.

Many attempts at secondmy structure prediction have beenmade using hand-coded rules, statistical and pattern matchingalgorithms and neural networks. However, the best results forthe overall prediction rate are still disappointing (65-70%).In [29] the ILP system Golem was applied to learning sec-ondary structure prediction rules for the sub-domain of a/aproteins. The input to the program consisted of 12 non-

homologous proteins (1612 amino acid residues) of knownsecondary structure, together with background knowledgedescribing the chemical and physical properties of the nat-urally occurring amino acids. Golem learned a set of 21rules that predict which amino acids are part of the a-helixbased on the positional relationships and chemical and phys-ical properties. The rules were tested on four independentnon-homologous proteins (416 amino acid residues) givingan accuracy of 8190 (with 2% expected error). The best pre-viously reported result in the literature was 76%, achievedusing a neural net approach. Although higher accuracy ispossible on sub-domains such as a/a than is possible in thegeneral domain, ILP has a clear advantage in terms of com-prehensibility over neural network and statistical techniques.

However, according to the domain expert, Mike Sternberg,Golem’s secondary-structure prediction rules were hard tocomprehend since they tended to be overly specific and highlycomplex. In many scientific domains, comprehensibility ofnew knowledge has higher priority than accuracy. Such apriori preferences can be modelled by a prior distribution onhypotheses.

A

THz

NH

k,

A-\ \’

N/ /NH

Hz

Figure 1: The family of analogues in the first ILP study. A)Template of 2,4-diamino-5(substituted-benzyl)pyrimidinesR3, R4, and R5 are the three possible substitution positions.B) Example compound: 3 – Cl, 4 – iViZZ, 5 – C’HJ

drug development projects. However, the in vitro activity ofmembers of a family of drugs can be determined experimen-tally in the laboratory. This data is usually analysed by drugdesigners using statistical techniques, such as regression, topredict the activity of untested members of the family of

drugs. This in turn leads to directed testing of the drugspredicted to have high activity.

An application of ILP to structure-activity prediction was re-ported in [19]. The ILP program Golem [28] was appliedto the problem of modelling the structure-activity relation-ships of trimethoprim analogues binding to dihydrofolatereductase. The training data consisted of 44 trimethoprimanalogues and their observed inhibition of Escherichia colidihydrofolate reductase. A further 11 compounds were usedas unseen test data. Golem obtained rules that were statisti-cally more accurate on the training data and also better on thetest data than a previously published linear regression model.Figure 1 illustrates the family of analogues used in the study.

Owing to the lack of variation in the molecules, a highlydeterministic subset of the class of logic programs was suffi-cient to learn in this domain. Unlike the protein domain, thedomain expert found the rules to be highly compact and thuseasy to comprehend.

3.2 DRUG STRUCTURE-ACTIVITY PREDICTION3.3 MUTAGENESIS PREDICTION

Proteins are large macro-molecules involving thousands ofatoms. Drugs tend to be small, rigid molecules, involvingtens of atoms. Drugs work by stimulating or inhibiting theactivity of proteins. They do so by binding to specific siteson the protein, often in competition with natural regulatorymolecules, such as hormones. The 3-dimensional structureof the protein binding site is not known in over 90% of all

In a third molecular biology study ILP has been applied topredicting mutagenicity of molecules [42]. Mutagenicity ishighly correlated to carcinogenicity. Unlike activity of adrug, mutagenicity is a property which pharmaceutical com-panies would like to avoid. Also, unlike drug developmentprojects, the data on mutagenicity forms a highly heteroge-

5

Page 4: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

:.B ,cl

Figure 2: Examples of the diverse set of aromatic andheteroaromatic nitro compounds used in the mutagene-sis study. A) 3,4,4’-trinitrobiphenyl B) 2-nitro-l,3,7,8-tetrachlorodibenzo-1 ,4-dioxin C) 1,6,-dinitro-9,10, 11,12-tetrahydrobenzo [e]pyrene D) nitrofurantoin

neous set with large variations in molecular form, shape andchame distribution. The data comes from a wide number ofsources, and the molecules involved are almost arbitrary inshape and bond connectivity (see Figure 2).

The 2D bond-and-atom molecular descriptions of 229 aro-matic and heteroaromatic nitro compounds taken from [8]were given to the ILP system Progol [26]. The study wasconfined to the problem of obtaining structural descriptionsthat discriminate molecules with positive mutagenicity fromthose which have zero or negative mutagenicity.

A set of 8 compact rules were discovered by Progol. Theserules suggested 3 previously unknown features leading tomutagenicity. Figure 3 shows the structural properties de-scribed by the rules in the Progol theory. The descriptionsof these patterns are as follows. 1) Almost half of all muta-genic molecules contain 3 fused benzyl rings. Mutagenicityis enhanced if such a molecule also contains a pair of atomsconnected by a single bond where one of the atoms is con-nected to a third atom by an aromatic bond. 2) Another strongmutagenic indicator is 2 pairs of atoms connected by a singlebond, where the 2 pairs are connected by an aromatic bond asshown. 3) The third mutagenic indicator discovered by Pro-gol was an aromatic carbon with a partial charge of +0. 191 inwhich the molecule has three N02 groups substituted onto abiphenyl template (two benzine rings connected by a singlebond).

An interesting feature is that in the original regression anal-ysis of the molecules by [8], a special ‘indicator variable’was provided to flag the existence of three or more fusedrings (this variable was then seen to play a vital role in theregression equation). The first rule (pattern 1 in Figure 3), ex-

NY +

3

2

u x

v—Y... . .. .

x

Y— z...... .

Q\

NO,— /N02—

N02—

i“)\Figure 3: Structural properties discovered by Progol

pressing precisely this fact, was discovered by Progol withoutaccess to any specialist chemical knowledge. Further, there-gression analysis suggested that electron-attracting elementsconjugated with nitro groups enhance mutagenicity. A par-ticular form of this pattern was also discovered by Progol(pattern 3 in Figure 3).

Previous published results for the same data [8] used lineardiscriminant techniques and hand-crafted features. The ILPtechnique allowed prediction of 49 cases which were statis-tical outliers not predictable by the linear discriminant algo-rithm. In addition, the ILP rules are small and simple enoughto provide insight into the molecular basis of mutagenicity.Such a low-level representation opens up a range of arbitrarychemical structures for analysis. This is of particular interestfor drug discovery, since all compounds within a database ofcompounds have known atom and bond descriptions, whilevery few have computed features available.

Owing to the large range of molecules the rules were highlynon-deterministic (of the form “there exists somewhere inmolecule M a structure S with property P“). Owing to sim-plicity, the mutagenicity rules were the most highly favoured

by the expert among the Molecular Biology domains studied.For all the domains in Molecular Biology comprehensibility,rather than accuracy, provides the strongest incentive for anOccam distribution (ie. shorter theories preferred a priori).

4 U-LEARNABILITY

The learnability model defined in this section allows for theincorporation of a priori distributions over hypotheses. Thedefinition of this model, U-learnability, is motivated by thegeneral problems associated with learning Universal (Turingcomputable) representations, such as logic programs. Never-theless U-learnability can be easily applied to other represen-tations, such as decision trees. U-learnability is defined using

6

Page 5: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

the traditional notation from computational learning theoryrather than that typically used in ILP (see Section 2).

4.1 BACKGROUND FOR U-LEARNING

Let (ZE, X, Zc, R, c) be the representation of a learning

problem (as in PAC-learning), where X is the domain ofexamples (finite strings over the alphabet ZE), R is the classof concept representations (finite strings over the alphabetXc), and c : R + 2X maps concept representations to con-cepts (subsets of X). Hence the conce t class being learned

$is the actual range of c (a subset of 2 ), Let P be a subsetof the class of polynomial functions in one variable, Fromthis point on, a concept representation will be a pair (r, p),forr ERandp EP.

Let A(r, p, z) be an algorithm that maps any tuple (r, p, z),forrg R,p CP, andz EX, toO orl. Let Arunin time bounded by q( lrl, P(IZ [)), where q is a polynomialin two variables. Furthermore, let A have the followingproperties. First, for all r 6 R and z c X, if z @ c(r)then for every p E P: A(r, p, z) = O. Second, for allr E R and z E X, if z E c(r) then there exists p E P suchthat for every p’ E P with p’(lzl) z p(lxl): A(r, p’, ~) =1. Intuitively, p specifies some form of time bound thatis polynomially related to the size of the example. It mightspecify the maximum derivation length (number of resolutionsteps) for logic programs or the maximum number of stepsin a Turing Machine computation. Greater derivation lengths

or more computation steps are allowed for larger inputs.

Let F be any family of probability distributions over R x P,where the distributions in F have an associated parametern >0. Let G be any family of probability distributions overx.

4.2 PROTOCOL FOR U-LEARNING

In U-learning, a Teacher randomly chooses a pair (r, p), forr E R and p c P, according to some distribution D1 inF. The Teacher presents to a learning algorithm L an infi-nite stream of labeled examples {(z1, 11), (Z2, 12), ...) where:each example xi is drawn randomly, independently of thepreceding examples, from X according to some fixed, un-known distribution D2 in G and labeled by li = A(r, p, ~i).After each example Zj in the stream of examples, L outputsa hypothesis Hi = (ri, pi), where ri E R and pi c P.

4.3 DEFINITION OF U-LEARNABILITY

Definition 1 Polynomial U-learnability. Let F be a familyof distributions (with associatedparameters) over R x P, andlet G be a family of distributions over X. The pair (F,G)is polynomially U-learnable just if there exist polynomialfunctions pI (y) = y’, for some constant c, P2(V), p3(yI, y2),and a learning algorithm L, such that for every distributionD1 (with parameter n) in F, and every distribution D2 in G,~hefollowing hold.

● The average-case time complexi~ of L at any point ina run is bounded by p] (M), where M is the sum of

As

q(lr[, P( Iwl)) over the examples xi seen m that point,2(Recall that (r, p) is the target concept chosen randomly

according to D1, and that q describes the time compkx-ity of the evaluation algorithm A.)

For all m > p2(n), there aist values E and d, O <

c, 8< 1, such that: p3 ( ~, ~) > m, and with probabilityat least 1 – J the hypothesis Hm = (rm, pm) is (1 –C)-accurate. By (1 - t)-accurate, we mean that theprobability (according to D2) that A(rm, pm, Xm+l ) #lm+l is less than e.

with PAC-learning, we can also consider a predictionvariant, in which the hypotheses llm need not be ‘built fromR (in addition to P), but can come from a different class R’and can have a different associated evaluation algorithm A’.

4.4 INITIAL RESULTS AND PROPOSEDDIRECTIONS

This section presents initial results about polynomial U-Iearnability and suggests directions for further work. Toconserve space the proofs are omitted, but they appear in afull paper on U-learnability [30]. In each of these results, let(X~, X, XC, R, c) be the representation of any given learningproblem. The following theorem shows that PAC-learnabilityis a special case of U-learnability,

Theorem 2 U-learnability generalises PAC. Letp be thepolynomialjimction p(x) = x. Let F be afamily of distribu-tions that contains one distribution Di for each ri c R, suchthat Di assigns probability 1 to (ri, p). Let n = O be the pa-rameterfor each member of F. Let G be thefamily of all dis-tributionsover X. Then the representation (ZE, X, xc, R, c)is PAC-leamable (PAC-predictable) if and only if (F, G] ispolynomially U-learnable (U-predictable).

This observation raises the question of how, exactly, polyno-mial U-learnability differs from PAC-learnability. There is acosmetic difference that polynomial U-learnability has beendefined to look more similar to identi$cation in the limitthan does the definition of PAC-learnability. Ignoring thisleaves the following distinguishing features of polynomialU-learnability.

@~ be rn~re precise, let m be any positive integer. bt

PrD, (r, p) be the probability assigned to (r, p) by DI, and (witha stight abuse of notation) let PrD1,Dz (r, P, (~1, ... . z~)) be theproduct of PrDl (r, p) and the probability of obtaining the sequence(x,, .... z~) when drawing m examples randomly and indepen-dently according to Dz. LetlTME~(r, p, (X I,.... z~)) be the timespent by L until output of its hypothesis I-f~, when the target is(r, p) and the initial sequenceof m examples is (XI, .... z~). Thenwe say that the average-casetime complexity of L to output Ifm isbounded by pI (y) = y= just if the sum (or integral), over all tuples(r,p, (m, ....zm)). of

is less than infinity, where M is the sum of g(lrl, p(lz, l)) over all%, in a 1,.... x~. See [2] for the motivation of this definition ofaverage-casecomplexity.

7

Page 6: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

o

Assumption of a prior distribution on target conceptrepresentations.

Sample size is related to e, d, and (by means of the pa-rameter n) the reciprocal of the probability of the targetrepresentation rather than the target representation size.

Use of average case time complexity, relative to theprior distribution on hypotheses and the distribution onexamples, in place of worst case time complexity.

Concept representations need not bepolynomially evalu-able in the traditional sense, but only in the weaker sensethat some polynomial p must exist for each concept rep-resentation such that the classification of an example zcan be computed in time P( lx 1). This distinction and thenext are irrelevant for concept representations that arealready polynomially evaluable and hence do not needtime bounds.

The learning algorithm runs in (expected) time polyno-mial in the ‘(w&st-case) time r~uired to compute theclassification, by the target, of the examples seen thusfar, rather than in the sum of the sizes of the examplesseen thus far.

Definition 3 The distribution family Dp. Let R be count-

able, let p’(y) be a polynomial function, and let EPI =(EI, E2, ...) bean enumeration of subsets of R such that:(1) for all i ~ 1, the caniinality of E~ is at most p’(i), (2)every r E R ls in some Ei, and (3) the members of E~, foreach i, can be determined eflciently (by an algorithm thatruns in time polynomial in the sum of the sizes of the membersof E~J Let F’ be the family of all distributions D), with finitevariance, over the positive integers, and let the parameter n)

of D’ be the maximum of the mean and standard deviationof D’. For each such distribution D’ in F’, define a corre-sponding distrt”bution D{{ over R as follows: for each i ~ 1,let the probabilities of the r G Ei be unijomtly distributedwith sum PrDl (i). Let the parameter associated with D“also be n’. Let P be the set of all linear functions of theform p(x) = CX, where c is a positive integer Let D~ beany probability distribution over the positive integers thathas apmbabilitydensi~ finction (p.d.$) prD; (x) = $, for

some k >3 (appropriately normalised so that it sums to 1).For each distribution D~, dejine a distribution DL over Pto assign to the function p(z) = cx the probability assignedto c by D~. Finally, for each pair of distributions D“ andDL as de$ned above, define a distribution D over R x Pas follows: for each pair (r, p), where r E R and p ~ P,prD (r, p) = [PrD1/ (r)] [PrDL (p)]. ~t the parameter n of Dbe n’. DP is the family of all such distributions D, each withassociated parameter n.

Theorem 4 Polynomial U-learnability under Dp. Let Fbe the distribution family DP defined by a particular enu-meration of subsets of R. Let G be any family of distributionsover .X. The pair (F, G) is polynomially U-learnable.

The following example relates U-learnability to the ILP def-initions given in Seetion 2.

Example 5 Time-bounded logic programs are U-learnableunder DP. Let R be the set of all logic programs that can be

builtfmm a given alphabet of predicate symbols P, jimctionsymbols F, and variables V. Notice that R is countable.Let P be the set of all bounds p(x) on derivation length ofthe form p(z) = CX, where c is a positive integer and x isthe size of (number of terms in) the goal. Let the domainX of examples be the ground atomic subset of cd, which isthe set of dejinite clauses that can be built from the givenalphabet. Notice also that a concept (r, p) classifies as pos-itive just the dejinite clauses d E X such that r I-P(IdlJ d,

where Idl is the number of terms in d. Let F be the family ofdistributions DP built using some particular enumeration ofpolynomially-growing subsets of R, and let G be thefamily ofall distributions over examples. From Theorem 4, it followsthat (F, G) is polynomially U-learnable.

Definition 6 The distribution family DE.. Let R be count-able, let k be a positive intege~ and let Ee = (El, E2, ...) bean enumeration of subsets of R such that: (1) for all i ~ 1,the cardinality of Ei is at most ki, (2) every (r, p), for r c Rand p G P, is in some Ei, and (3) the members of Ei, foreach i, can be determined efficiently (by an algorhhm thatruns in timepolynomial in the sum of the sizes of the membersof Ei). Let D~ be the discrete exponential distribution withprobability densi~finction

Pr(x) = (k – l)k-x for all integers x ~ 1

and let the parameter n’ of D~ be its mean {~). From D;,

de$ne a corresponding distribution Dk over R asfollows: foreach i ~ 1, let the probabilities of the r c Ei be uniformlydistributed with sum Prl); (i). Let the parameter associated

with Dk also be n’. Let P be the set of all linear functionsof the form p(x) = CX, where c is a positive integer Let thedistributions DL over P be as defined earlier (see definitionof Dp }. Finally, for the distribution Dk (with parameternl) and each distribution DL, dejine a distn”bution D overR x Pas follows: for each pair (r, p), for r E R andp G P,

prD (r, P) = [prD, (r)] [prDL (P)]. Let the parameter n of Dbe the parameter n’ of Dk. DE, is the family of distributionsconsisting of each such distribution D.

Theorem 7 Polynomial U-1earnability under DE.. LetF be the distribution family DE, dejined by a particularenumeration of subsets of R and a particular choice of k,Let G be any family of distributions over X, The pair ofdistributions (F, G) is polynomially U-learnable.

Example 8 Time-bounded logic programs are U-learnableunder DEh. Again let R be the set of all logic programs thatcan be builtfrom a given$nite alphabet ofpredicate symbols

P, function symbols F, and variables V. And again let P bethe set of all bounds p(z) on den’vation length of the formp(x) = CX, where c is a positive integer and x is the size of(number of terms in) the goal. Let the domain X of examplesbe the ground atomic subset of cd. Let (S1, S2, ...) be anenumeration of subsets of R such that Si contains all logicprograms of length i, for all i ~ 1. Let F be the distribu-tion family DE, built fmm this enumeration and fmm thepositive integer k chosen to be the sum of the sizes of P, F,and V. Let G be the family of all distn”butions over exam-ples. From Theorem 7, it follows that (F, G) is polynomiallyU-learnable.

8

Page 7: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

In many cases it may be incorrect to assume that the proba-bilities of target representations vanish as quickly as they doin the distributions considered in the preceding theorems. Inparticular, a more appropriate distribution on target represen-tations may be defined by combining an enumeration Ee, as inDefinition 6, with some distributionin which the probabilitiesdo not decrease exponentially, as in Definition 3. PolynomialU-learnability should be considered with such distributions,perhaps combined with restricted families of distributions onexamples. It can be expected that obtaining positive resultsfor “natural” such distributions will be theoretically challeng-ing, and that such results will have a major impact on practicalapplications of machine learning. It is also expected that itwill be challenging and useful to obtain negative results forsome such distributions. Whereas negative results for PAC-learnability can be obtained by showing that the consistencyproblem for a given concept representation is NP-hard (as-suming RP # NP) [35], negative results for polynomialU-learnability in some cases can be obtained by showing thatthe consistency problem is NP-hard-on-average (or DistNPhard) [2, 16] relative to particular distributions on hypothe-ses and examples. In addition, just as negative results forPAC-predictabiIity can be obtained (based on certain assump-tions) using hard problems from cryptography [17], negativeresults for polynomial U-predictability might possibly be ob-tained in this same way, but would require additional effort.Specifically, obtaining negative results for polynomial U-Unpredictability will require specifying a particular distribution(or family of distributions) to be used with a hard problemfrom cryptography, such as inverting RSA, and it must beverified, in some way, that the problem does not become sub-stantially easier under this distribution (or every distributionin this family). For example, inverting RSA is assumed hardif encryptionldecryption is based on a choice of two largerandom prime numbers according to a uniform distributionover primes of some large size, but it becomes easy if thedistribution over primes of a given size assigns probability1 to one particular prime number; what about for distribu-tions “in between”, where some primes of a given size aremore likely than others? The distributions used in polyno-mial U-predictability would correspond to such distributionsfor RSA.

4.5 INCORPORATING BACKGROUNDKNOWLEDGE

Recall that background knowledge is often used in ILP. Thispaper has not provided for the use of background knowl-edge in polynomial U-learnability. The full paper on U-learnability includes a definition of polynomial U-learnabilitythat allows the teacher to provide the learner with a back-ground theory. A surprising aspect of this definition is thatit not only captures the notion of inductive learning relativeto background knowledge, as desired, but it also providesa generalisation of Natarajan’s PAC-style formalisation ofspeed-up learning [33]. Although Natarajan’s formalisationis quite appealing, few positive have been proven within it be-cause it is also very demanding. The use of a family of priordistributions on hypotheses, as specified in the definition ofpolynomial U-learnability, can in some cases effectively ease

these demands and lead to additional positive results.

5 DISCUSSION

As the ILP applications in Section 3 show, not only is Hornclause logic a powerful representation language for scien-tific concepts, but also one which is tractable in real-worlddomains. This apparently flies in the face of negative PAC-learnability results for logic programs. This may be explainedby the fact that despite logic programs being used as the rep-resentation language throughout the applications in Section3, very different distributional assumptions were necessaryin each application.

ILP has a clear model-theoretic semantics which is inheritedfrom Logic programming. However, without considerationof the probabilistic nature of inductive inference, a purelylogic-based definition of ILP is always incomplete. For thisreason the Bayes’ oriented definitions of ILP found in Section2 provide a more complete semantics for I~P than has pre-viously been suggested. This introduces prior distributionalinformation into the definition of ILP. The introduction ofsuch information is also motivated both by the requirementsof practical experience and the ability to define a model oflearnability which is more appropriate for learning logic pro-grams.

All ILP algorithms make implicit use of distributional in-formation. One promising line of research would involveexplicit provision of distributions to general-purpose ILP sys-tems.

The choice of representation is at the centre of focus in PAC-learning. By contrast, the choice of distribution is the centraltheme in U-learnability. The reason for the refocusing ofattention is that once one is committed to a Universal repre-sentation, as in the case of ILP, differences in distributionalassumptions become paramount.

Some general initial results of U-learnability are stated inthis paper. Much more effort is required to clearly define theboundaries of what is and is not U-learnable. It is believedthat U-learnability will not only be of interest to researchers inILP but also to the wider community within Machine Learn-ing.

Acknowledgements

The author would like to thank David Page and AshwinSrinivasan of Oxford University Computing Laboratory andMichael Sternberg and Ross King of Imperial Cancer Re-search Fund for discussions and input to this paper. Thiswork was supported partly by the Esprit Basic Research Ac-tion ILP (project 6020), the SERC project GR/J46623 andan SERC Advanced Research Fellowship held by the author.The author is also supported by a Research Fellowship atWolfson College, Oxford.

9

Page 8: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

References

[1]

[2]

[3]

[4]

[5]

[6]

[7]

[8]

[9]

[10]

[11]

[12]

[13]

B. Arbab and D. Michie. Generating rules from ex-amples. In ZJCAZ-85, pages 63 1–633, Los Altos, CA,1985. Kaufmann.

S. Ben-David, B. Chor, and O. Goldreich. On the theoryof average case complexity. Journal of Information andSystem Sciences, 44: 193–219, 1992.

R. Board and L. Pitt. On the necessity of Occam algo-rithms. UIUCDCS-R-89- 1544, University of Illinois atUrbana-Champaign, 1989.

I. Bratko. Generating human-understandable decisionrules, Working paper, E. Kardelj University Ljubljana,Ljubljana, Yugoslavia, 1983.

B. Buchanan, E. Feigenbaum, and N. Sridharan. Heuris-tic theory formation: data interpretation and rule for-mation. In B. Meltzer and D. Michie, editors, Machineintelligence 7, pages 267–290. Edinburgh UniversityPress, 1972.

W. Buntine. A Theory of Learning Classi$cation Rules.PhD thesis, School of Computing Science, Universityof Technology, Sydney, 1990.

W. Cohen. Learnability of restricted logic programs.In S. Muggleton, editor, Proceedings of the 3rd Inter-national Workshop on Inductive Logic Programming(Technical report IJS-DP-6707 of the Josef Stefan In-stitute, Ljubljana, Slovenia), pages 41-72, 1993.

A.K. Debnath, R.L Lopez de Compadre, G. Debnath,A.J. Schusterman, and C. Hansch. Structure-activityrelationship of mutagenic aromatic and heteroaromaticnitro compounds. correlation with molecular orbital en-

ergies and hydrophobicity. Journal of Medicinal Chem-istry, 34(2):786 – 797, 1991.

B. Dolsak and S. Muggleton. The application of Induc-tive Logic Programming to finite element mesh design.In S. Muggleton, editor, Inductive Logic Programming.Academic Press, London, 1992.

S, Dzeroski, S. Muggleton, and S. Russell. PAC-learnability of determinate logic programs. In Pro-ceedings of the 5th ACM Workshop on ComputationalLearning Theory, Pittsburg, PA, 1992,

D, Haussler. Applying Valiant’s learning frameworkto AI concept-learning problems. In Y. Kodratoff andR. Michalski, editors, Machine learning: an artificialintelligence approach, volume 3, pages 641-669, Mor-gan Kaufman, San Mateo, CA, 1990.

D. Haussler, M Keams, and R. Shapire. Bounds onthe sample complexity of bayesian learning using in-formation theory and the vc dimension. In COLT-91:Proceedings of the 4th Annual Workshop on Computa-tional Learning Theory, pages 61–74, San Mateo, CA,1991. Morgan Kauffmann.

N. Helft. Induction as nonmonotonic inference. InProceedings of the Ist International Conference onPrinciples of Knowledge Representation and Reason-ing, pages 149–156. Kaufmann, 1989.

[14]

[15]

[16]

[17]

[18]

[19]

[20]

[21]

[22]

[23]

[24]

[25]

[26]

[27]

[28]

[29]

C.J. Hogger. Essentials of logic programming. OxfordUniversity Press, Oxford, 1990.

P. Idestam-Almquist. Generalisation of Clauses. PhDthesis, Stockholm University, 1993.

D. Johnson. The NP-completeness column-an ongo-ing guide. Journal of Algorithms, 4:284-299, 1984.

M, Kearns and L. Valiant. Cryptographic limitationson lemming boolean formulae and finite automata. InProceedings of the 21st ACM Symposium on Theory ofComputing, pages 433444. ACM, 1989.

J.U Kietz. Some lower bounds on the computa-tional complexity of inductive logic programming. InP. Brazdil, editor, Proceedings of the 6th EuropeanConference on Machine Learning, volume 667 of Lec-ture Notes in Artificial Intelligence, pages 115–123.Springer-Verlag, 1993.

R. King, S. Muggleton R. Lewis, and M. Steinberg.Drug design by machine learning: The use of inductivelogic programming to model the structure-activity rela-tionships of trimethoprim analogues binding to dihydro-folate reductase. Proceedings of the National Academyof Sciences, 89(23), 1992.

N. LavraE and S. DZeroski. Inductive Logic Program-ming : Techniques and Applications. Ellis Horwood,1993.

N. Lavra5, S. D2eroski, and M. Grobelnik. Learningnon-recursive definitions of relations with LINUS. InYves Kodratoff, editor, Proceedings of the 5th Eur-opean Working Session on Learning, volume 482 of Lec-ture Notes in Arti@cial Intelligence. Springer-Verlag,1991.

J.W. Lloyd. Foundations of Logic Programming.Springer-VerIag, Berlin, 1984.

S. Muggleton. Inductive logic programming. New Gen-eration Computing, 8(4):295–3 18, 1991.

S. Muggleton, editor. Inductive Logic Programming.Academic Press, 1992.

S. Muggleton. Inductive logic programming: deriva-tions, successes and shortcomings. SIGART Bulletin,5(1):5-1 1, 1994.

S. Muggleton. Mode-directed inverse resolution. InK. Furukawa, D. Michie, and S. Muggleton, editors,Machine Intelligence 24. Oxford University Press, (toappear).

S. Muggleton and W. Buntine. Machine invention offirst-order predicates by inverting resolution. In Pr-oceedings of the Fijlh International Conference on Ma-chine Learning, pages 339–352. Kaufmann, 1988.

S. Muggleton and C. Feng. Efficient induction of logicprograms. In Proceedings of the First Conference onAlgorithmic Learning Theory, Tokyo, 1990. Ohmsha.

S. Muggleton, R. King, and M. Steinberg. Protein sec-ondary structure pred~ction using logic-&sed machinelearning. Protein Engineering, 5(7):647-657, 1992.

10

Page 9: Bayesian Inductive Logic Programmingshm/Papers/bayesian.pdf · 2014-09-26 · Bayesian Inductive Logic Programming Abstract Stephen Muggleton Oxford University Computing Laboratory

[30] S. Muggleton and D. Page. A learnability model for [45] J. Zelle and R. Mooney. Learning semantic grammarsuniversal representations. Technical Report PRG-TR-3- with constructive inductive logic programming. In Pro-

94, Oxford University Computing Laboratory, Oxford, ceedings of the Eleventh National Conference on Ar-1994. ti~cial Intelligence, pages 8 17–822, San Mateo, CA,

[31] S. Muggleton and L. De Raedt, Inductive logic pro-1993, Morgan Kaufmann.

gramming: Theory and methods. Journal of LogicProgramming, 12, 1994. (to appear).

[32] S. Muggleton, A. Srinivasan,and M, Bain. Compres-sion, significanceand accuracy.In Proceedings of theNinth InternationalMachine Learning Conference, SanMateo, CA, 1992. Morgan-Kaufmann.

[33] B. K. Natarajrm, Learning from exercises. In Proceed-ings of the 1989 Workshop on Computational LearningTheory, pages 72–86, San Mateo, CA, 1989. MorganKaufmann.

[34] D. Page and A. Frisch. Generalization and learnabil-ity: A study of constrained atoms. In S. Muggleton,editor, Inductive Logic Programming. Academic Press,London, 1992.

[35] L. Pitt and L, G. Valiant. Computational limitationson learning from examples. Journal of the ACM,35(4):965-984, 1988.

[36] G. Plotkin. A further note on inductive generalization.In Machine Intelligence, volume 6. Edinburgh Univer-sity Press, 1971.

[37] G.D. Plotkin. A note on inductive generalisation. InB. Meltzer and D. Michie, editors, Machine Intelligence5, pages 153–163. Elsevier North Holland, New York,1970,

[38] R, Quinlan. Learning logical definitions from relations.Machine Learning, 5:239-266, 1990.

[39] L. De Raedt and M. Bruynooghe, A theory of clausaldiscovery. In Proceedings of the 13th InternationalJoint Conference on Artificial Intelligence, MorganKaufmann, 1993.

[40] L. De Raedt and S, Dzeroski. First-order jk-clause the-ories are pat-learnable. Technical report, Departmentof Computer Science, Katholieke Universieit Leuven,Heverlee, Belgium, 1994.

[41] C, Sammut, S. Hurst, D. Kedzier, and D. Michie. Learn-ing to fly. In D. Sleeman and P. Edwards, editors, Pro-ceedings of the Ninth International Workshop on Ma-chine Learning, pages 385–393, San Mateo, CA, 1992.Morgan Kaufmann.

[42] A. Srinivasan and S. Muggleton. Discovering rules formutagenesis. Technical Report PRG-TR-4-94, OxfordUniversity Computing Laboratory, Oxford, 1994.

[43] M, Sternberg, R, King, R. Lewis, and S. Muggleton,Application of machine learning to structural molecu-lar biology. Philosophical Transactions of the RoyalSociety B, 1994 (to appear).

[44] M. Steinberg, R. Lewis, R. King, and S, Muggleton,Modelling the structure and function of enzymes bymachine learning, Proceedings of the Royal Society ofChemistry: Faraday Discussions, 93:269-280,1992.

11