title kernel methods in computational biology vert,...

15
Title Kernel methods in computational biology Author(s) Vert, Jean-Philippe; 阿久津, 達也 Citation 物性研究 (2003), 81(1): 142-155 Issue Date 2003-10-20 URL http://hdl.handle.net/2433/97608 Right Type Departmental Bulletin Paper Textversion publisher Kyoto University

Upload: others

Post on 08-Jun-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

Title Kernel methods in computational biology

Author(s) Vert, Jean-Philippe; 阿久津, 達也

Citation 物性研究 (2003), 81(1): 142-155

Issue Date 2003-10-20

URL http://hdl.handle.net/2433/97608

Right

Type Departmental Bulletin Paper

Textversion publisher

Kyoto University

Page 2: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

Kernel methods in computational biology

Jean-Philippe Vert (Ecole des Mines de Paris)

Bt~1:!f&J~~= :t:Ht.Q jJ~*)vit

JJ.R~fF~: ~?\?$J¥1f! (J?:*1~$mJi:PJf)

1. ~~

~.~~$~mJi:§~~-~~.~.~mJi:~~~~~~h~*.~~-~~M~G.~~$~~~fflti

{hU>t~ Eilh~~:~< t-:/it)~Bt.=f-$~1#l§£9~ ~ c -C~ ~o *t-:. ~4m$~:Slt~T~'j:~~~~ti~-~iO!~

~~h~~/it).~hG~~~GT.'5~c~-C~~.$~~m~~~lli9~C~m.ti.e~-~-C~~o

~.~~$~:sltn:·M~cti~~-~~:~'j:.lI~T~~l~-~. 1~$f$j~~-~. lI~T§£m~-~tic

il~~~1J~. ~ ~ -c~'j:. ~hG~~-~~:. '5 ~ c ~PJ~~C9~jJ-*)v$~:~lt~Tm8)j9~o jJ-*)v$~'j:

~ ~ -t1f.rl'l~~:~~~1!l7tJf~:Slt~TJ£JiG T~~=f-$-C~ ~. ~~$~"BtJ'.~< ~rl=l~e~:.ttffl~hTlt~~o

2. Mercer fJ - *)v~~X ~ramiO' G~.~~. K(x,y) il!. ~t~ltt (K(x,y)=K(y,x)) ~frMJt-: G. ~ G~:. iEJE{[(['tt~frMJt-:9

~~~:. ~. K(.,.) ~'j: Mercer tJ-*)vclJ-¥~;fh~o K(.,.) il! Mercer jJ-*)v-C~~~~. ~~ 1::)v/\()v t­~rl'l~<I>. s~lt. XiO\G<I>~~~.¢ (X)1J~ffttG. K(x,yP'j:¢ (x) c<!> (y) ~P'lmctiQo ~~. ~wH:~'j:

RKHS(reproducing kernel Hilbert space) C: lJ*tfh~ 1:::: )v/\()v J--. ~FI'I~~ ffllt\ ~ z: C: ~::J:: I:J. Mercer jJ -*)v

c RKHS ~~t.tt·:Ht9~ ~ c1J~-C~~o *t-:. RKHS ~m.ti'tt'itc GT. RKHS il!~~R*:lG~rl'l~-c35-:>

T't. 35Q~{tf:~-CiEJtrJ1~~ht-:~.~~/J\1~iO~~~R@I~J¢J.~~~~~9Q~ c-Cfrx~clt~ '5 ~ ciO~~~j

GhQo

3. fJ-*)v5!jJ-*)v$~*~tJ.;frJJ¢J.~-~cGT. 1::)v/\()v t- ~FI'l~~~fl9Q ~ c~G ~:~/{~~.iO~frx~ ~ ciO~

~~jGh. ~~;::c~'j:jJ-*)vt- ') '/7clJ-¥~;fh~o AA!jfti{9ilcGT~'j:1::)v/\()vt-~rl'l~~:S~tQ 2J¢J.rl'l~~Im

l1JliO~jJ-*)v~.~AA!jftitJl~~bit-C*/it)~ ~ c1J~-C~ Qo et I:J~ fflti{9ilc GT. ~BtM~~'±.=f-$~

-~-C~Q'±~7.t7.t~(PCA) 1J~. jJ-*)v~ ffllt~t-:~~~: 't. 1::)v/\()v t- ~rl'l~~:S~t ~Bt.ti G ~:t-=rX QojJ -*)v~ ffllt\t-:iE$t13~7.t~ (CCA) ~'j:fi!il~{[([fit.rl=l~e~:9m~9 ~ ;:: ciO~-c~. =~m~~-~ ~~~G

t-:M~~t-=r '5 ~~:~ffl-C~~o

~7f- t- /\( 7 ~-~.~,</ (SVM) ~'j:jJ-*)v$~;:£':J < (~am35 ~) ~~$~~t-:/it)~=f-$-C. iE~~{9il

~4xGh~~.iE~~~~7.tI1JlG.~~.Mm~*-c~ImI1Jl(~-~~)~M*cti~m~w~~.9~o

~llf]H:~'j:. iE~~{9il~~~~;:7.t11Jl9Q;:: ciO~/f'PJ~~-C~Q~~iO~~lt~~'"t'. 7.tmaJ;'H~cImI1Jl~ t- v- ~::t7

Gt-:~~~~)MHt9~oSVM-C~'j:. 7J-*)vt-1)'/7~;:et~.~)jti7.tlUm~W1J! (~<~~~~;:~'j:1>tiv~

~-1A:~) iE~~{9ilO)$7.t~~~:~t9~7J-*)v~m~~bit~;:etI:J~m~h~o

4. ~/I'\?~:T-~I;:M9.QfJ-*)v5!

tJ-*)v$~~~$~-~~::)jffl9~t-:{}). ~ ~J\7'it~~@9~T-~ ~::~t9~jJ-*)v~.iO'q'J!~~

hTlt~~o :M'~;:. rJ2~rj C)(~~U) (:~t9~jJ-*)v~.~'j:et<mJi:~hTlt~Qo *~k~$7.tJt~~l~llim

t~JU~~/\(7 t- )v~ffllt)~;:: c ~:~ I:J. )(~~rjiO' G.::L-71) '/ ~ ~rl'l~~~jJ-*)v~fx:~JE~-C~~1J~. ~ ~

=f-$~'j: spectrum tJ-*)vclJ-¥~;fhTv) ~o *t;:. ~~lM~~;:J.t< ;frJffl~hTv' ~~$1::~)v-C~Q~h~)v::1 7=C~)v (HMM) tic il, G't¥f¥N~fffitfJ9~ ~ c (:~ ~ tJ-*)V~fx:~JE~9Q. Fisher tJ-*)V'tfJ!~~hTlt)~o

~~UT-~ £)3'i-~:~'j:. lI~TJ£m~-~. Phylogenetic Profile tic~. '5t-:/it)O)jJ-*)v~. 7·7 7f$j~

(:~9Q diffusion tJ-*)vctJ-*Jv CCA ~tJl~-@Jbit{~~J\Ar).:r:-1 CJ£m~-~~t13~~fffitfJ9Qm

Ji:tJ. c 1J~frbhT It, Q 0

jJ-*Jv~tJl~~bit(:~9QmJi:'t1-=rbhTS ~. *iEJE{[([BtTIID$~;:~Q. jJ-*Jv~~W*S~~M)i

{l::;tic1J~;p;fJi:~hTV 1Q 0

- 142-

Page 3: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

Kernel methods in computational biology

Jean-Philippe VertEcole des Mines de Paris

35 rue Saint-Honor77300 Fontainebleau

[email protected]

1 Introduction

Computational biology in the post-genomics era aims at providing a computational envi­ronment for biological research, where huge amounts of data generated by high-throughputtechnologies can be integrated, and where biological hypotheses resulting from the auto­matic analysis of these data can be generated. Besides obvious data management issues,core problems in computational biology concern the definition of a coherent and usefulmathematical framework to integrate data.

The purpose of this paper is to propose a mathematical framework to model largesets of objects, such as the set of all genes in a genome or of all chemicals in a cell,and to represent various types of information about these objects (such as the sequences,structures and expression of genes) in a unified framework. Moreover, a number of toolsfor analysis and inference are provided in this framework. It is based on the theory ofMercer kernels and reproducing kernel Hilbert spaces, while analysis and inference toolsare derived from the recently developed theory of kernel methods which has attracted alot of attention during the last decade in the machine learning community.

This framework should obviously neither be considered a unique nor a final solution tothe problem of building a mathematical framework for post-genomics. There is probablyno single answer to this need, but rather a collection of possible frameworks with differentadvantages and drawbacks. The one proposed in this paper is certainly very limited interms of its capacity to represent complex relationships among objects, such as evolution­ary or regulatory relationships among genes. However, the tools for statistical analysisand inference in this framework are reasonably powerful and have been shown to be usefulin many applications, such as predicting the functions of genes. Hence we can say thatthis formalism is slightly biased toward deriving powerful inference tools, at the expenseof modeling complex biological relationships.

The organization of th~ paper is the following. We first describe the mathematicaltheory of Mercer kernels and review some of their properties, as well as the family ofkernel methods they underlie. We then turn our attention to biological systems such asthe set of all genes of an organism, and argue that various types of information about thesystem can be encoded as Mercer kernels and can be considered as various realizations ofa more abstract model of the systems.

The application of kernel methods in computational biology has recently been subjectto much investigation. Even though we mention arbitrary references to illustrate some ofour statements, this short paper is not a review of this field and we apologize for all theinteresting recent contributions which we don't mention here.

- 143-

Page 4: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

2 Mercer kernels

Let X be a set supposed to represent objects one wants to analyze, e.g., the set of allgenes in a given genome. We propose to represent any form of information about theobjects by a Hilbert space structure on the set X. Intuitively, the Hilbert product betweentwo objects will be defined as a mea~ure of similarity of the objects with respect to theinformation available. Obviously, every measure of similarity does not define a validHilbert product. For a function K : X x X -t lR to be a valid product, it must besymmetric (i.e., K(x, y) = K(y, x) for any x, y E X) and positive definite in the ~ense thatfor any n EN, any (al,'" ,an) E lRn and any (Xl, ... ,xn) E X n, the following holds:

n

L aiajK(xi, Xj) 2: O.i,j=l

In that case we say that K is a Mercer kernel. It is easy to see that if X is a Hilbertspace with Hilbert product < .,. >, then K(x, y) =< X, Y > is a Mercer kernel on X.Conversely, if X is a set and K(.,.) a Mercer kernel on X, then there exists a Hilbertspace <I> and a mapping ¢ : X -t <I> such that the kernel between two points be the Hilbertproduct between their images: K(x, y) =< ¢(x), ¢(y) >.

An interesting construction of such a mapping ¢ is the reproducing kernel Hilbert space(RKHS), which we now define. Simply speaking, a (real) RKHS on X is a Hilbert space1i of real-valued functions on X with the property that, for each X E X, the evaluationfunctional Lx which associates any f E 1i with Lxf = f(x), is a bounded linear functional.The boundedness means that for any x EX, there exists a constant M x E lR such that

Vf E 1i, ILxfl = If(x)1 :S Mxllfllrt·

We remark that this definition implies that the functions of the RKH8 must be definedpointwise and thus that the familiar space £2(X), if X is measurable, is not a RKH8.The link with Mercer ~ernels is given in the following theorem which shows an equivalencebetween RKH8 and Mercer kernels [8802]:

Theorem 1 Let (1i,< .,. » be a RKHS on a set X. Then there exists a unique Mercerkernel K on X with the following properties:

1. K spans 1i, i.e., 1i is the completion of span{K(x, .), x EX}.

2. K has the reproducing property, .i.e.,

V(j, x) E 1i x X, < f, K(x,.) >= f(x);

in particular,

V(x, x') E X X X', < K(x, .), K(x' ,.) >= K(x, x'). (1)

Conversely, if X is a set and K is a Mercer kernel on X ,then a these equations defineuniquely a RKHS on X.

Equation 1 shows that when X is a set endowed with a Mercer kernel, then the mapping¢ : x -t ¢(x) = K(x,.) from X to the RKH8 defined by K is a valid mapping from Xto a Hilbert space that satisfies K(x, y) =< ¢(x), ¢(y) >, and justifies the fact that anyMercer kernel defines a Hilbert space structure on X.

- 144-

Page 5: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

A number of interesting properties make Mercer kernels a very powerful tool in real­world applications. First, Mercer kernels can be designed in such a way that the norm inthe RKHS be of interest for various purposes. As an example, suppose X = JR d , and takethe Gaussian RBF kernel:

(Ilx - Y11 2

)K(x, y) = exp 2(j2 '

which is a valid Mercer kernel. Then the norm in the associated RKHS is given by:

(2)

where j is the Fourier transform of f (here H is the set of functions with a Fouriertransform such that (2) be finite). In that case Ilfll7-l is a smoothing functional, in thesense that it is small when f is smooth, and large otherwise. This can typically be usefulas a term for regularization, when one wants to minimize a function 0,(J) on H withsimultaneously imposing some smoothness constraints on f (otherwise the problem mightbe ill-posed). In that case, instead of minimizing 0,(J), one might want to minimize0, (J) + >'11 f 117-l, where >. controls the trade-off between the minimization of n and thesmoothness of f.

A second remarkable propertiy of RKHS is that even though they are typically infinite­dimensional, a large class of regularized minimization problems in H can be solved exactlyin a lower-dimensional space. The precise statement is contained in the following repre­senter theorem [KW71, SS02]

Theorem 2 Let w : [0,(0) ---7 JR a strictly monotonic increasing function, X a set with aMercer kernel K and associated RKHS H. Let (Xl, ... , Xn ) E X n be a set of points in Xand n be a function on H. expressed in terms of the Xi and f(Xi) only, i.e., of the form:

Then each minimizer f E H. of the regularized function:

0,(J) + w(llfll7-l)

admits a representation of the form

n

f(x) = I:CKiK(Xi, .).i=l

The Representer theorem has tremendous consequences on the design of methods andalgorithms, including such techniques as splines [Wah90] and kernel methods [SS02]. Itmeans that minimizing a regularized form of a functional which only depends on the valueof f on a finite number of points can be done by working in a finite-dimensional space,and reduces to finding coefficients CKi associated with each point Xi. A whole variety ofalgorithms can be deduced from this theorem by varying the function to be optimizedand the Mercer kernels, ranging from generalized forms of principal component analysisto support vector machines for classificatiOIl and regression.

- 145-

Page 6: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

3 Kernel methods

In this section we briefly review a number of algorithms known as kernel methods based onthe Representer theorem. They all take as input a finite number of points in a set endowedwith a Mercer kernels, such as the set of genes where a Mercer kernel has been defined asa similarity measure with respect to some sort of information, and provide useful tools toperform various analysis of these points. For each of these methods, two complementarypoints of views can help get an intuition of how they work. First, the direct point ofview consists in viewing each point mapped by a function ¢ to some Hilbert space. Thealgorithms consist them in working implicitly in the resulting Hilbert space. Second, froma dual point of view, the Mercer kernel defines a RKHS. The algorithms consist thenin optimizing a regularized functional in the RKHS. The equivalence between these twopoints of view is ensured by the construction of RKHS and the Representer theorem.

3.1 Computing a Euclidean distance

Suppose you want to define a Euclidean distance between two points x and x' of a spaceX endowed with a Mercer kernel K. Then an obvious solution is to consider the Hilbertdistance between their images in a Hilbert space defined by K, i.e., to consider the followingsquare distance:

d(x, x')2 = 11¢(x) - ¢(x')112

=< ¢(x) - ¢(x'), ¢(x) - ¢(x') >=< ¢(x), ¢(x) > + < ¢(x'), ¢(x') > -2 < ¢(x), ¢(x') >= K(x,x) +K(x', x') - 2K(x,x').

The result is expressed in terms of J( only. This shows than once a kernel is given on aset X, it defines naturally a Euclidean distance which can be obtained without computingthe images of the points in the Hilbert space associated with K. This is an example ofthe kernel trick, which consists in working implicitly in a complex Hilbert space withoutever seeing a point in that space.

3.2 Computing the distance between a point and the center of a cloudof points

Suppose you have a set of points (Xl, . .. ,xn ) in a set X endowed with a Mercer kernel,and you want to compute how each point is "far" from the "average" of all points. Oneway to formalize this intuitive goal is to consider the set of points mapped to a Hilbertspace associated with J( by a mapping ¢. Then a natural definition of "average" of allpoints is the center of mass of all all points, i.e.,

1 n

m = - L¢(Xi),n

i=l

and the distance between a point Xi and the average can be quantified by the squareEuclidean distance between ¢(Xi) and m, which can be computed by:

- 146-

Page 7: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

Once again, the kernel trick enables to perform the computation implicitly in the featurespace, and to obtain the result in terms of the function K between the input points only.

3.3 Kernel principal component analysis

Suppose you want to analyze the variations among points (Xl, ... ,xn ) of a set X witha Mercer kernel K. A classical and useful tool in mathematical statistics is principalcomponent analysis (PCA), which is defined when X = IRd • It is a powerful technique forextracting structure form possibly high-dimensional data sets, by projecting the points onthe so-called principal axes of the data, defined as the directions which capture the largestamounts of variance [Jo196] (see Figure 1).

Figure 1: PCA detects the directions of largest variations among a family of points in avector space, called principal directions.

The standard PCA algorithm is readily performed by solving an eigenvalue problem,i.e., by diagonalizing the covariance matrix:

where the data have been centered first (L~l Xi = 0). Suppose now that X is not avector space, but simply a set endowed with a Mercer kernel. Then a useful way toanalyze variations among the points and extract features is to perform a PCA on thepoints ¢(Xi) mapped to a Hilbert space associated with K. As shown in [SSM99, SS02],this can be performed in a dual form by centering the data in the Hilbert space anddiagonalizing the matrix

K~j =< ¢O(Xi)' ¢o(Xj) >,

where ¢O(Xi) represents the point ¢(xiJ after centering. Observe that while the size of thecovariance matrix C is d x d, where d is the dimension of the vector space X in classicalPCA, the dimension of KO is n x n, where n is the number of points to analyze. Henceit becomes tractable to perform PCA in a possibly infinite-dimensional space, becausethe principal components are in fact always in the finite-dimensional space spanned bythe points. Once again, the centering of the matrix, the computation of the principaldirections and the projections of the points can be computed implicitly by only workingwith the function K.

3.4 Extracting correlations between two kernels

Suppose that two different Mercer kernels K 1 and K 2 are defined on the same set X. Thiscan be the case for instance when different notions of similarity are defined on the sameobjects, such as sequence similarity and co-expression for genes. The two kernels define twodifferent mappings ¢ 1 and ¢2 to two Hilbert spaces. In order to compare the two kernels,and extract correlations betwccn them, a useful tool in statistics is canonical correlation

- 147-

Page 8: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

analysis (CCA). When a set of points is given in simultaneously two finite-dimensionalvector spaces, cla.'3sical CCA consists in extracting pairs of directions, one in each space,such that the projections of the points on these directions be as correlated as possible[Hot36] (see Figure 2). Recently it was pointed out that a regularized form of CCA can be

•• •

Figure 2: CCA detects directions simultaneously in two vector spaces such that the pro­jections of the points be maximally correlated.

performed implicitly between two sets of points mapped to Hilbert spaces associated withMercer kernels [BJ02], and that performing CCA in this context is equivalent to solvingthe following generalized eigenvalue problem:

This provides a useful tool to extract correlations between two forms of informations en­coded in two kernels, and h&') recently been applied in the context of natural languageprocessing [AVC03], to extract language-independent semantic representations of texts,and in computational biology [VK, YVNK03], to compare gene networks and gene expres­sions or detect operons in prokaryote genomes.

3.5 Support vector machines

Support vector machines (SVM) are certainly the best-known kernel methods. Theirpopularity is due to their remarkable performances on many real-world problem, andthey have been a main topic of investigation in the machine learning community duringthe last decade. First introduced by Vapnik and coworkers [BGV92, Vap98], SVM area family of algorithms useful for classification and regression. In the simplest binaryclassification case, points (Xl, ... ,xn ) from a set X with a Mercer kernel K are giventogether with a binary label associated with each point, denoted by (Yl,"" Yn) whereYi E {-I, I}. The goal of classification is to find a discrimination between positive andnegative point,s. The solution implemented in SVM, motivated by theoretical results ofstatistical learning theory, is to take a linear discrimination between the positive andnegative points mapped to the Hilbert space associated with K, such that the distancebetween the separating hyperplane and the closest point (called the margin) be the largestpossible. This problem has a unique solution when the points are linearly separable, i.e.,when there exists at least one hyperplane which separates positive points from negativeones. In the general situation, SVM finds an hyperplane that optimizes a trade-off betweenthe number of misclassified points and the margin (see Figure 3). Once again this problemcan be expressed in a dual formulation and results in the following optimization problem:

{1ll""O I:~~j f1; -0\ I:;~j~j f1;f1jY;YjJ((x;, Xj),

'V'l, - 1, ... ,n O:S O:'I :S C,

2:7=1 O:"iYi = 0,

- 148-

Page 9: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

, •, \0 • •\ \.,\

\ , • •0\,

\

0 0 • \

0\\\ ,, •0 0 ,

©>-,,, ,

0 0, \,

Figure 3: SVM finds a lineal' discrimination rule between two classes of points, by opti­mizing a functions which achieves a trade-off between the number of misclassified points(a small as possible) and the margin of the classification, i.e., the distance between theseparating hyperplane and the closest point (which should be as large as possible).

where C is the parameter which controls the trade-off between large margin and goodclassification. This algorithm is one of the most powerful algorithms availablenowadayson many real-world problems. In particular, as those equations suggest, if X = lRP,then the dimension d does not matter too much on the performance of SVM (thanks tothe choice of the largest margin linear classifier), or at least it seems to matter less forSVM than for many other classification algorithms which often perform poorly in largedimensions (this is part of a phenomenon called the curse of dimensionality).

SVM have been applied with success to many real-world problems. In the field ofcomputational biology, a non-exhaustive list of applications using SVM include gene func­tional classification from microarrays [BGL+00, PWCG01], tissue samples classificationfrom microarrays [MTIVI+9S, FDC+OO, GWBV02], protein family prediction from amino­acid sequences [JDHOO], or protein secondary structure and localization prediction fromamino-acid sequences [HSOlb, HSOla].

4 Kernelizing the proteome

This quick overview of the theory of Mercer kernels, RKHS and kernel methods aimed atconvincing the reader that they provide a framework with sound theoretical foundations,and which is associated with a panoply of powerful methods thanks to the simple Repre­senter theorem. In this section, we now turn our attention to the question of the relevanceof this framework to post-genomiGs.

One of the main challenges of computational biology in the post-genomics era is tointegrate and process a variety of data generated by high-throughput technologies aboutbiological objects, and use them to generate biological hypothesis. In the sequel we fo­cus on problems related to functional genomics and proteomics, where one wants to inferknowledge about genes such as their functions, their regulation and their interactionsfrom several different sources of information. Concretely, genes and proteins can be char­acterized by a number of different points a view: they can be defined by nucleotides oramino-acid sequences, by the structure of their promoter region in DNA, by their expres­sion measured with DNA microarrays, by the secondary, tertiary and quaternary structureof the proteins they encode, by their role in metabolic and signaling pathways, by their in­teractions etc... Each characterization incorporates some partial representation of a moreabstract concept of gene.

One approach to integrate these complementary characteri7;ations in a unified franlc-

- 149-

Page 10: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

work is to encode each type of information into a Mercer kernel defined on the set X ofall genes (which can be considered a finite set when one focuses on the genes of a givengenome, but can be extended to more complex structures such as the set of all finite­lengths nucleotide sequences for instance). This suggests a more abstract representationof a set of genes, as for instance a probability density on the set of kernel functions (e.g.,the empirical measure corresponding to a set of kernel functions obtained by encodingdifferent sources of informations). We won't develop further this avenue in this paper, butfurther investigations of the shape of this empirical measure is likely to provide interestinginformation about the relationships between different types of information. This couldalso provide an interesting notion of complexity of a biological system, as the entropy ofthe probability measure on the set of Mercer kernels used to represent it.

To illustrate this discussion we review in the following sections several recent attemptsto encode biological information into kernel functions, as well as several attempts to explorethe space of Mercer kernels obtained by different sources of information. By lack ofspace we will conclude with these onsiderations, but believe that exploring this frameworkis likely to result in new interesting conceptualizations of biological systems, as well aspowerful tools to handle large amounts of heterogeneous data.

4.1 String kernel

Biological databases are full of sequences. A gene can be characterized as a subsequenceof the DNA molecule (hence as a finite string over a four-letter alphabet), or by thesequence of amino-acids which forms the protein it encodes (hence as a finite string inover 20-letter alphabet). In both cases, the set of genes is a finite set of sequences, witha particular hidden structure. Many genes are known to share common ancestors, whichoften indicates functional or structural relationships. This evolutionary relationship can bedetected to some extent by comparing the sequences of genes, because two genes sharing acommon ancestor were similar a long time ago, but have evolved in different environmentsthrough mutations or insertion/deletion of letters or subsequences. As a result, measuringthe "similarity" between gene sequences gives some information about the evolutionary,functional and structural relationships between genes.

It this context, a number of initiatives have been taken to engineer kernel functions forsequences to transform the sequence similarity relationship into a Hilbert space structureon the set of genes. Examples include the spectrum kernel [LEN02] which explicitlymaps any sequence to its k-spectrum (k > 0) i.e., the number of appearance of each k­mer in the sequence, and defines the inner product between k-spectra as Hilbert productbetween sequences. This is motivated by the idea that even though the global appearanceof a sequence might be strongly modified by several mutations during evolution, its k­spectrum might be more stable. This kernel was later improved in the mismatch kernel[LEWN03] which consists in convolution the k-kernel with a kernel that accepts up tom < k mismatches, and provides a powerful tool to detect remote homology betweenproteins.

Another general approach for string kernel engineering is the Fisher kernel [JH99] whichis a two-step process. First, a parametric statistical model for sequences, i.e., a familyof probability densities {Po(.), f) E 8 C lRk } on the set of finite sequences, is built (e.g.,a hidden Markov models [DEKM98]), and a parameter () E 8 is estimated to fit a set ofsequences of interest. Second, each sequence x is mapped to the score vector of) log Po (x),i.e., to a vector in the tangent space at the point PiJ of the Riemannian manifold formed bythe statistieallllodel. This manifold has a natlll'al Riemannianllletric [ANOI] and a naturalpositive definite quadratic form is therefore available in the tangent space where sequences

- 150-

Page 11: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

are mapped. This ingenious approach is particularly useful because a lot of work has beendevoted in the early days of bioinformatics to the development of parametric statisticalmodels for biological objects, and the Fisher kernel provides a principled way to derivea kernel from virtually any such model. We can also mention other approaches to makekernels from probabilistic models, such as convolution kernels [Hau99, WatOO, SVUA03]which are based on probabilistic models for pairs of sequences, and can be computedefficiently using finite-state automata.

4.2 Expression profiles

With the development of DNA microarrays, it becomes increasingly simple to characterizea gene by an expression profile, i.e., a vector of real numbers which indicate its level ofexpression in different experiments. As genes are supposed to be expressed when thecell needs the proteins they encode, genes with similar profiles are likely to have relatedfunctions. Hence it might be useful to define a Hilbert product from these vectors, whichcan easily be achieved as a large number of kernel for real-valued vector have been knownfor a long time, such as the simple inner product or the Gaussian RBF kernel.

4.3 Other kernels

Here we briefly mention a non-exhaustive list of Mercer kernels which have been developedrecently in the context of computational biology. As the number of sequenced organismsincreases, classical homology detection tools can detect whether a given gene is presentor absent in every fully sequenced organism. Hence a gene can be characterized by avector of 0/1, which indicates its presence or absence in all sequenced organisms. Thisvector is called the phylogenetic profile of the gene [PMT+99]. Two genes with similarphylogenetic profiles are present and absent in the same organisms, hence chances arehigh that they have related function or participate to common structural complexes, forinstance. While the simple dot product between vectors of bits provides a valid Mercerkernel, a more specific kernel for phylogenetic profiles was proposed in [Ver02b] and shownto be more relevant than the simple dot kernel. This kernel uses prior knowledge about theevolutionary relationships among sequenced organisms, and is one instance of a generalapproach to derive a kernel from probabilistic models when structural knowledge existsamong the random variables [Ver02a].

An other interesting approach is the diffusion kernel [KL02] which is a kernel for thenodes of a graph. It is a discrete equivalent of the Gaussian RBF kernel, which can becomputed as the matrix exponential 'Of the opposite of the graph Laplacian. Intuitively,the kernel between two nodes increases when there are many short paths between them;more precisely, the kernel K(x, y) is the probability of reaching the node y by a randomwalk on the graph starting from the node x, and killed with some probability at each step.This kernel is known as the heat kernel [Chu97] in spectral graph theory, and is particularlyrelevant to define a kernel from a network of genes, for example interaction or regulationnetwork. It was recently applied in combination with kernel canonical correlation analysisto extract correlations between metabolic pathways (which can be represented as a networkof genes) and gene expression data [VK, VK03].

This short list is far from being exhaustive, and the list of kernels specifically developedfor particular objects is increasing very quickly, providing a panoply of Mercer kernels torepresent various types of information about genes.

- 151 -

Page 12: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

4.4 Kernel operations

We showed in the previous sections that a number of kernels can be engineered to representspecific knowledges about genes. As the cla.'3s of Mercer kernel is the class of symmetricpositive definite functions, a variety of new kernels can be computed from a family of basickernels using the closure properties of the class of symmetric positive definite functions.For example, this class is a closed convex cone, i.e., it is stable by addition, multiplicationby a positive constant, and pointwise limit. Some kernels, known as infinitely divisible, canbe taken to an arbitrary non-negative exponent and remain positive definite [Hau99, SS02].

As an example, observe that if 1(1 is a string kernel for protein sequences, and J(2 isa Gaussian RBF kernel for the expressions of the corresponding genes, then 1(1 + J(2 isa kernel that incorporates both sequence and expression similarity. Hence the addition oftwo kernels is a simple yet powerful way to combine two types of informations, which hasbeen for example successfully used in the context of gene function prediction [PWCG01].More generally, if 1(1, ... ,J(p is a family of kernels on the same space, then a valid Mercerkernel obtained as a linear combination of the basic kernel which optimizes some linearfunction can be found by semi-definite programming, which has been explored for instancein [LCG+02]. Alternatively, one can use information geometric properties of the set ofpositive definite matrix to generate new kernels from basic kernels [KT].

5 Conclusion

In this short overview of recent advances in kernel methods and their applications incomputational biology, we tried to present kernel methods as a relevant framework tomanipulate biological data and perform computations and inference from them. Thishas been a very active and exciting field or research in the last few years, as it turnsout that a wide range of kernels can be engineered and adapted to particular data, andthat the machinery of kernel methods provide a large family of algorithms to extractinformation from kernel representations. Moreover, mathematical operations on kernelsprovide a interesting starting point to develop more satisfactory frameworks for modelingliving systems. As an example, we mentioned the possibility we intend to study in thefuture to represellt a biological system as a probability distribution on the set of Mercerkernels, in such a way that' the Mercer' kernels built from various sources of informationwould be different realizations of a more abstract random kernels. Such developmentswould certainly require more research or more integration of theoretical results aboutpositive definite kernels, and might lead to new families of algorithms to fit the need ofpost-genomics.

6 Acknowledgments

I am grateful to Kenji Veno and Tsuyoshi Kato who offered me the opportunity to par­ticipate to the workshop "Mathematical aspects of molecular biology: toward new math­ematics" that was held in Nara, Japan, on January 24-27, 2003. This text is based on thetalk I gave there.

References

[ANOl] Shunichi Amari and Hiroshi Nagaoka. Methods of information geometry. AMSvol. 191, 2001.

- 152-

Page 13: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

[AVC03] John Shawe-Taylor Alexei Vinokourov and Nello Cristianini. Inferring asemantic representation of text via cross-language correlation analysis. InSuzanna Becker, Sebastian Thrun, and Klaus Obermayer, editors, Advancesin Neural Information Processing Systems. MIT Press, 2003.

[BGL+00] Michael P. S. Brown, William Noble Grundy, David Lin, Nello Cristianini,Charles Walsh Sugnet, Terence S. Furey, Jr. Manuel Ares, and David Haussler.Knowledge-based analysis of microarray gene expression data by using supportvector machines. Proc. Natl. Acad. Sci. USA, 97:262-267, 2000.

[BGV92] B. E. Boser, 1. M. Guyon, and V. N. Vapnik. A training algorithm for op­timal margin classifiers. In Proceedings of the 5th annual ACM workshop onComputational Learning Theory, pages 144-152. ACM Press, 1992.

[BJ02] Francis R. Bach and Michael 1. Jordan. Kernel independent component anal­ysis. Journal of Machine Learning Research, 3:1--48, 2002.

[Chu97] Fan R.K. Chung. Spectral graph theory, volume 92 of CBMS Regional Con­ference Series. American Mathematical Society, Providence, 1997.

[DEKM98] Richard Durbin, Sean Eddy, Anders Krogh, and Graeme Mitchison. Biolog­ical Sequence Analysis: Probabilistic Models of Proteins and Nucleic Acids.Cambridge University Press, 1998.

[FDC+OO] Terrence S. Furey, Nigel Duffy, Nello Cristianini, David Bednarski, MichelSchummer, and David Haussler. Support vector machine classification andvalidation of cancer tissue samples using rnicroarray expression data. flioin­formatics, 16(10):906-914, 2000.

[GWBV02] Isabelle Guyon, Jason Weston, Stephen Barnhill, and Vladimir Vapnik. Geneselection for cancer classification using support vector machines. MachineLearning, 46(1/3):389--422, Jan 2002.

[Hau99]

[Hot36]

[HSOla]

[HSOlb]

[JDHOO]

[JH99]

[Jo196]

David Haussler. Convolution kernels on discrete structures. Technical report,UC Santa Cruz, 1999.

H. HoteHing. Relation between two sets of variates. Biometrika, 28:322--377,1936.

Sujun Hua and Zhirong Sun. A novel method of protein secondary struc­ture prediction with high segment overlap measure: Support vector machineapproach. Journal of Molecular Biology, 308(2):397-407, April 2001.

Sujun Hua and Zhirong Sun. Support vector machine approach for proteinsubcellular localization prediction. Bioinformatics, 17(8):721--728, 2001.

Tommi Jaakkola, Mark Diekhans, and David Haussler. A discriminativeframework for detecting remote protein homologies. Journal of ComputationalBiology, 7(1,2):95-114, 2000.

Tommi S. Jaakkola and David Haussler. Exploiting generative models indiscriminative classifiers. In Proc. of Tenth Conference on Advances in NeumlInformation Processing Systems, 1999.

LT. Jolliffe. Principal component analysis. Springer-Verlag, New-York, 1996.

- 153-

Page 14: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

[KL02] R. I. KondoI' and .J. Lafferty. Diffusion kernels on graphs and other discreteinput. In ICML 2002, 2002.

[KT] K. Asai K. Tsuda, S. Akaho. Approximating incomplete kernel matrices bythe EM algorithm. Poster presented at the 6th kernel machines workshop,2002.

[KW71] G.S. Kimeldorf and G. Wahba. Some results on tchebycheffian spline functions.Journal of Mathematical Analysis and Applications, 33:82-95, 1971.

[LCG+02] Gert R.G. Lanckriet, Nello Cristianini, Laurent EI Ghaoui, Peter Bartlett, andMichael I. Jordan. Learning the kernel matrix with semidefinite programming.Technical report, UC Berkeley, 2002.

[LEN02] Christina Leslie, Eleazar Eskin, and William Stafford Noble. The spectrumkernel: a string kernel for svm protein classification. In Russ B. Altman,A. Keith Dunker, Lawrence Hunter, Kevin Lauerdale, and Teri E. Klein, edi­tors, Proceedings of the Pacific Symposium on Biocomputing 2002, pages 564­575. World Scientific, 2002.

[LEWN03] Christina Leslie, Eleazar Eskin, Jason Weston, and William Stafford Noble.Mismatch string kernels for svm protein classification. In Suzanna Becker,Sebastian Thrun, and Klaus Obermayer, editors, Advances in Neural Infor­mation Processing Systems 15. MIT Press, 2003.

[MTM+98] S. Mukherjee, P. Tamayo, J.P. Mcsirov, D. Slonim, A. Verri, and T. Poggio.Support vector machinc classification of microarray data. Technical Rcport182, C.B.L.C., 1998. A.I. Memo 1677.

[PMT+99] Mattco Pellcgrini, Edward M. Marcotte, Michael J. Thompson, David Eisen...:berg, and Todd O. Yeates. Assigning protein functions by comparative genomeanalysis: Protein phylogenetic profiles. Proc. Natl. Acad. Sci. USA, 96:4285­4288, April 1999.

[PWCGOl] Paul Pavlidis, Jason Weston, Jinsong Cai, and William Noble Grundy. Gene'functional classification from heterogeneous data. In Proceedings of the FifthAnnual International Conference on Computational Biology, pages 249-255,2001.

[SS02] Bernhard Scholkopf and Alexander J. Smola. Learning with Kernels: Sup­port Vector Machines, Regularization, Optimization, and Beyond. MIT Press,Cambridge, MA, 2002.

[SSM99] Bernhard Scholkopf, Alexander J. Smola, and Klaus-Robert Muller. Kernelprincipal component analysis. In B. Scholkopf, C. Burges, and A. Smola,editors, Advances in Kernel Methods - Support Vector Learning, pages 327­352. MIT Press, 1999.

[SVUA03] Hiroto Saigo, Jean-Philippe Vert, Nobuhisa Ueda, and Tatsuya Akutsu. Pro­tein homology detection using string alignment kernels. To appear, 2003.

[Vap98] Vladimir N. Vapnik. Statist'leal Learning Theory. Wiley, New-York, 1998.

- 154-

Page 15: Title Kernel methods in computational biology Vert, …repository.kulib.kyoto-u.ac.jp/dspace/bitstream/2433/...Kernel methods in computational biology Jean-PhilippeVert Ecole des Mines

[Ver02a]

[Ver02b]

[VK]

[VK03]

[Wah90]

[WatOO]

Jean-Philippe Vert. Support vector machine prediction of signal peptide cleav­age site using a new class of kernels for strings. In Russ B. Altman, A. KeithDunker, Lawrence Hunter, Kevin Lauerdale, and Teri E. Klein, editors, Pro­ceedings of the Pacific Symposium on Biocomputing 2002, pages 649-660.World Scientific, 2002.

Jean-Phlippe Vert. A tree kernel to analyze phylogenetic profiles. Bioinfor­matics, 18:S276~S284, 2002.

Jean-Philippe Vert and Minoru Kanehisa. Graph-driven features extractionfrom lllicroarray data. Preprint arXiv physics/0206055, June 2002.

Jean-Philippe Vert and Minoru Kanehisa. Graph-driven features extractionfrom microarray data using diffusion kernels and kernel cca. In SuzannaBecker, Sebastian Thrun, and Klaus Obermayer, editors, Advances in NeuralInformation Processing Systems. MIT Press, 2003.

G. Wahba. Spline Models for Observational Data, volume 59 of CBMS-NSFRegional Conference Series in Applied Mathematics. SIAM, Philadelphia,1990.

C. Watkins. Dynamic alignment kernels. In A.J. Smola, P.L. Bartlett,B. Scholkopf, and D. Schuunnans, editors, Advances in Large Margin Classi­fiers, pages 39--50. MIT Press, Cambridge, MA, 2000.

[YVNK03] Y. Yamanishi, J.-P. Vert, A. Nakaya, and M. Kanehisa. Extraction of corre­latcd gcne clusters from multiple gcnomic data by gencralized kcrnel canonicalcorrelation analysis. Bioinformatics, 2003. To appear.

- 155-