input space versus feature space in kernel-based methods...

18
1000 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999 Input Space Versus Feature Space in Kernel-Based Methods Bernhard Sch¨ olkopf, Sebastian Mika, Chris J. C. Burges, Philipp Knirsch, Klaus-Robert M¨ uller, Gunnar R¨ atsch, and Alexander J. Smola Abstract—This paper collects some ideas targeted at advancing our understanding of the feature spaces associated with support vector (SV) kernel functions. We first discuss the geometry of feature space. In particular, we review what is known about the shape of the image of input space under the feature space map, and how this influences the capacity of SV methods. Following this, we describe how the metric governing the intrinsic geometry of the mapped surface can be computed in terms of the kernel, using the example of the class of inhomogeneous polynomial kernels, which are often used in SV pattern recognition. We then discuss the connection between feature space and input space by dealing with the question of how one can, given some vector in feature space, find a preimage (exact or approximate) in input space. We describe algorithms to tackle this issue, and show their utility in two applications of kernel methods. First, we use it to reduce the computational complexity of SV decision functions; second, we combine it with the Kernel PCA algorithm, thereby constructing a nonlinear statistical denoising technique which is shown to perform well on real-world data. Index Terms— Denoising, kernel methods, PCA, reduced set method, sparse representation, support vector machines. I. INTRODUCTION R EPRODUCING kernels are functions which forall pattern sets (1) give rise to positive matrices [23]. Here, is some compact set in which the data lives, typically (but not necessarily) a subset of In the support vector (SV) community, reproducing kernels are often referred to as Mercer kernels (Section II-B will show why). They provide an elegant way of dealing with nonlinear algorithms by reducing them to linear ones in some feature space nonlinearly related to input space: Using instead of a dot product in corresponds to mapping the data into a possibly high- dimensional dot product space by a (usually nonlinear) map Manuscript received January 23, 1999; revised May 16, 1999. Part of this work was done while P. Knirsch was with Bell Labs and B. Sch¨ olkopf and A. J. Smola were with the Department of Engineering, Australian National University, Canberra. This work was supported by the ARC and the DFG under Grant Ja 379/52,71,91. B. Sch¨ olkopf was with GMD FIRST, 12489 Berlin, Germany. He is now with Mocrosoft Research Ltd., Cambridge CB2, U.K. S. Mika, K.-R. M¨ uller, G. R¨ atsch, and A. J. Smola are with GMD FIRST, 12489 Berlin, Germany. C. J. C. Burges is with Bell Laboratories, Holmdel NJ, USA. P. Knirsch is with the Max-Planck-Institut f¨ ur biologische Kybernetik, 72076 T¨ ubingen, Germany. Publisher Item Identifier S 1045-9227(99)07268-9. , and taking the dot product there, i.e., [4] (2) By virtue of this property, we shall call a feature map asso- ciated with Any linear algorithm which can be carried out in terms of dot products can be made nonlinear by substituting an a priori chosen kernel. Examples of such algorithms include the potential function method, SV Machines, and kernel PCA [1], [34] [29]. The price that one has to pay for this elegance, however, is that the solutions are only obtained as expansions in terms of input patterns mapped into feature space. For instance, the normal vector of an SV hyperplane is expanded in terms of SV’s, just as the kernel PCA feature extractors are expressed in terms of training examples (3) When evaluating an SV decision function or a kernel PCA feature extractor, this is normally not a problem: due to (2), multiplying with some mapped test point transforms (3) into a kernel expansion which can be evaluated even if lives in an infinite-dimensional space. In some cases, however, there are reasons mandating a more comprehensive understanding of what exactly is the connection between patterns in input space and elements of feature space, given as expansions such as (3). This field being far from understood, the current paper attempts to gather some ideas elucidating the problem, and simultaneously proposes some algorithms for situations where the above connection is important. These are the problem of denoising by kernel PCA, and the problem of speeding up SV decision functions. The remainder of this paper is organized as follows. Section II discusses different ways of understanding the mapping from input space into feature space. Following that, we briefly review two feature space algorithms, SV machines and kernel PCA (Section III). The focus of interest of the paper, the way back from feature space to input space, is described in Section IV, and, in the more general form of constructing sparse approximations of feature space expansions, in Section V. The algorithms proposed in these two sections are experimentally evaluated in Section VI and discussed in Section VII. II. FROM INPUT SPACE TO FEATURE SPACE In this section, we show how the feature spaces in question are defined by choice of a suitable kernel function. Insight into 1045–9227/99$10.00 1999 IEEE

Upload: others

Post on 31-Jul-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1000 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

Input Space Versus Feature Spacein Kernel-Based Methods

Bernhard Sch¨olkopf, Sebastian Mika, Chris J. C. Burges, Philipp Knirsch,Klaus-Robert Muller, Gunnar Ratsch, and Alexander J. Smola

Abstract—This paper collects some ideas targeted at advancingour understanding of the feature spaces associated with supportvector (SV) kernel functions. We first discuss the geometry offeature space. In particular, we review what is known about theshape of the image of input space under the feature space map,and how this influences the capacity of SV methods. Followingthis, we describe how the metric governing the intrinsic geometryof the mapped surface can be computed in terms of the kernel,using the example of the class of inhomogeneous polynomialkernels, which are often used in SV pattern recognition. We thendiscuss the connection between feature space and input space bydealing with the question of how one can, given some vector infeature space, find a preimage (exact or approximate) in inputspace. We describe algorithms to tackle this issue, and show theirutility in two applications of kernel methods. First, we use it toreduce the computational complexity of SV decision functions;second, we combine it with the Kernel PCA algorithm, therebyconstructing a nonlinear statistical denoising technique which isshown to perform well on real-world data.

Index Terms—Denoising, kernel methods, PCA, reduced setmethod, sparse representation, support vector machines.

I. INTRODUCTION

REPRODUCING kernels are functions whichforall pattern sets

(1)

give rise to positive matrices [23]. Here,is some compact set in which the data lives, typically

(but not necessarily) a subset of In the support vector(SV) community, reproducing kernels are often referred to asMercer kernels(Section II-B will show why). They provide anelegant way of dealing with nonlinear algorithms by reducingthem to linear ones in some feature spacenonlinearlyrelated to input space: Using instead of a dot product in

corresponds to mapping the data into a possibly high-dimensional dot product spaceby a (usually nonlinear) map

Manuscript received January 23, 1999; revised May 16, 1999. Part of thiswork was done while P. Knirsch was with Bell Labs and B. Scholkopf andA. J. Smola were with the Department of Engineering, Australian NationalUniversity, Canberra. This work was supported by the ARC and the DFGunder Grant Ja 379/52,71,91.

B. Scholkopf was with GMD FIRST, 12489 Berlin, Germany. He is nowwith Mocrosoft Research Ltd., Cambridge CB2, U.K.

S. Mika, K.-R. Muller, G. Ratsch, and A. J. Smola are with GMD FIRST,12489 Berlin, Germany.

C. J. C. Burges is with Bell Laboratories, Holmdel NJ, USA.P. Knirsch is with the Max-Planck-Institut fur biologische Kybernetik,

72076 Tubingen, Germany.Publisher Item Identifier S 1045-9227(99)07268-9.

, and taking the dot product there, i.e., [4]

(2)

By virtue of this property, we shall call a feature map asso-ciated with Any linear algorithm which can be carried out interms of dot products can be made nonlinear by substituting ana priori chosen kernel. Examples of such algorithms includethe potential function method, SV Machines, and kernel PCA[1], [34] [29]. The price that one has to pay for this elegance,however, is that the solutions are only obtained as expansionsin terms of input patterns mapped into feature space. Forinstance, the normal vector of an SV hyperplane is expandedin terms of SV’s, just as the kernel PCA feature extractors areexpressed in terms of training examples

(3)

When evaluating an SV decision function or a kernel PCAfeature extractor, this is normally not a problem: due to (2),multiplying with some mapped test point transforms(3) into a kernel expansion which can be evaluated evenif lives in an infinite-dimensional space. In some cases,however, there are reasons mandating a more comprehensiveunderstanding of what exactly is the connection betweenpatterns in input space and elements of feature space, given asexpansions such as (3). This field being far from understood,the current paper attempts to gather some ideas elucidatingthe problem, and simultaneously proposes some algorithmsfor situations where the above connection is important. Theseare the problem of denoising by kernel PCA, and the problemof speeding up SV decision functions.

The remainder of this paper is organized as follows.Section II discusses different ways of understanding themapping from input space into feature space. Followingthat, we briefly review two feature space algorithms, SVmachines and kernel PCA (Section III). The focus of interestof the paper, the way back from feature space to inputspace, is described in Section IV, and, in the more generalform of constructing sparse approximations of feature spaceexpansions, in Section V. The algorithms proposed in thesetwo sections are experimentally evaluated in Section VI anddiscussed in Section VII.

II. FROM INPUT SPACE TO FEATURE SPACE

In this section, we show how the feature spaces in questionare defined by choice of a suitable kernel function. Insight into

1045–9227/99$10.00 1999 IEEE

Page 2: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1001

the structure of feature space can then be gained by consideringtheir relation to reproducing kernel Hilbert spaces, by how theycan be approximated by an empirical map, by their extrinsicgeometry, which leads to useful new capacity results, and bytheir intrinsic geometry, which can be computed solely interms of the kernel.1

A. The Mercer Kernel Map

We start by stating the version of Mercer’s theorem givenin [38]. We assume to be a finite measure space.2 By“almost all” we mean except for sets of measure zero.

Theorem 1 (Mercer):Suppose is a sym-metric real-valued kernel3 such that the integral operator

(4)

is positive, i.e., for all , we have

Let be the normalized eigenfunctions ofassociated with the eigenvalues , sorted in nonincreasingorder. Then

1) ;2) and ;3) holds for almost all

Either or in the latter case,the series converges absolutely and uniformly for almostall

From statement 3, it follows that corresponds to a dotproduct in i.e., with

(5)

for almost allIn fact, theuniform convergence of the series implies that

given there exists an such that even if the rangeof is infinite-dimensional, can be approximated withinaccuracy as a dot product in , between images of

B. The Reproducing Kernel Map

We can also think of the feature space as areproducingkernel Hilbert space(RKHS). To see this [2], [23], [36], [24],[13], recall that a RKHS is a Hilbert space of functionsonsome set such that all evaluation functionals, i.e., the maps

, are continuous. In that case, by the Riesz

1Those readers who are chiefly interested in applications and algorithmsmight want to consider skipping over the present section.

2A finite measure space is a setX with a �-algebra defined on it, anda measure defined on the latter, satisfying�(X )<1 (i.e., up to a scalingfactor,� is a probability measure). A�-algebra� onX is a family of subsetsof X ; which is closed under elementary set-theoretic operations (countableunions, intersections, and complements), and which containsX as a member.A measure is a function�: � ! + which is �-additive, i.e., the measureof a set which is the disjoint unions of some other sets equals the sum of thelatter’s measures.

3I.e., a function of two variables which gives rise to an integral operator.

representation theorem, for each there exists a uniquefunction of , call it , such that

(6)

(here, is the function on obtained by fixing the secondargument of to , and is the dot product of the RKHS.In contrast, we use to denote the canonical (Euclidean) dotproduct). In view of this property, is called areproducingkernel.

Note that by (6), for all implies that isidentically zero. Hence the set of functionsspans the whole RKHS. The dot product on the RKHS thusonly needs to be defined on and can thenbe extended to the whole RKHS by linearity and continuity.From (6), it follows that in particular

(7)

for all , (this implies that is symmetric). Note thatthis means that any reproducing kernelcorresponds to a dotproduct in another space.

Let us now consider a Mercer kernel, i.e., one which satisfiesthe condition of Theorem 1, and construct a dot product suchthat becomes a reproducing kernel for the Hilbert spacecontaining the functions

(8)

By linearity, we have

(9)

Since is a Mercer kernel, the can bechosen to be orthogonal with respect to the dot product in

Hence it is straightforward to construct a dot productsuch that

(10)

(using the Kronecker symbol , in which case (9) reducesto the reproducing kernel property (6) [using (8)].

Therefore, (7) provides us with a feature mapassociatedwith .

Proposition 2: For any Mercer kernel , there exists aRKHS such that for

(11)

we have

(12)

C. The Empirical Kernel Map

While giving an interesting alternative theoretical viewpoint,the map does not appear all that useful at first sight. Inpractice, it would seem pointless to first map the inputs intofunctions, i.e., into infinite-dimensional objects. However, fora given dataset, it is possible to approximatefrom (11) byonly evaluating it on these points (cf., [15], [18], [20], [33]).

Page 3: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1002 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

Definition 3: For given patterns , we call

(13)

the empirical kernel map with regard to .Example: Consider first the case where is a Mercer

kernel, and i.e., we evaluateon the training patterns. If we carry out a linear

algorithm in feature space, then everything will take placein the linear span of the mapped training patterns. Therefore,we can represent the of (11) as without losinginformation. However, the dot product to use in that represen-tation is not simply the canonical dot product in, since the

will usually not form an orthonormal system. To turninto a feature map associated with, we need to endowwith a dot product such that

(14)

To this end, we use the ansatz with being apositive matrix.4 Enforcing (14) on the training patterns, thisyields the self-consistency condition (cf., [28], [31])

(15)

Here, we have used to denote the kernel Gram matrix

(16)

The condition (15) can be satisfied for instance by the pseu-doinverse

Equivalently, we could have incorporated this rescalingoperation, which corresponds to a kernel PCA whitening [29],[28], [33], directly into the map, by modifying (13) to

(17)

This simply amounts to dividing the eigenvector basis vectorsof by where the are the eigenvalues of 5 Itparallels the rescaling of the eigenfunctions of the integraloperator belonging to the kernel, given by (10).

For data sets where the number of examples is smallerthan their dimensionality, it can actually be computationallyattractive to carry out explicitly rather than using kernelsin whatever subsequent algorithm (SVM’s, or kernel PCA,say) one wants to use. As an aside, note that in the caseof kernel PCA (to be described in Section III), one doesnot even need to worry about the whitening step: Using thecanonical dot product in (rather than will simplylead to diagonalizing instead of which yields the sameeigenvectors with squared eigenvalues. This was pointed outby [18] and [20].

We end this section with two notes which illustrate why theuse of (13) need not be restricted to the special case we justdiscussed.

• More general kernels. When using nonsymmetric kernelsin (13), together with the canonical dot product, one

4Note that every dot product can be written in this form. Moreover, we donot require definiteness ofA, as the null space can be projected out, leadingto a lower-dimensional feature space.

5It is understood that ifK is singular, we use the pseudoinverse ofK1=2:

will effectively work with a matrix , with generalNote that each positive semidefinite matrix can be

written asIf we wanted to carry out the whitening step, it would

have to be using (cf., footnote 5 concerningpotential singularities).

• Different evaluation sets. Mika [20] has performed exper-iments to speed up kernel PCA by choosingas a proper subset of

Now that we have described the kernel map in some detail,including variations on it, we shall next look at its properties.Specifically, we study its effect on the capacity of kernelmethods (Section II-D) and the induced geometry in featurespace (Section II-E).

D. The Capacity of the Kernel Map

Vapnik [34], [35] gives a bound on the capacity, measuredby the VC-dimension , of optimal margin classifiers. It takesthe form

(18)

where is an upper bound constraining the length of theweight vector of the hyperplane in canonical form, andis theradius of the smallest sphere containing the data in the spacewhere the hyperplane is constructed. The smaller this sphereis, the smaller is also the capacity, with beneficial effects onthe generalization error bounds.

If the data is distributed in a reasonably isotropic way, whichis often the case in input space, then (18) can be fairly precise.6

If, however, the distribution of the data is such that it doesnot fill the sphere, then (18) is wasteful. The argument inthe remainder of the section, which is summarized from [37],shows that using a kernel typically entails that the data in factlies in some box with rapidly decaying sidelengths, which canbe much smaller than the above sphere.

From statement 2 of Theorem 1, there exists some constantdepending on the kernel such that

for all and almost all (19)

Therefore, is essentially contained in an axis parallelparallelepiped in with side lengths [cf., (5)].

To see how this effectively restricts the class of functions weare using, we first note that everything inis done in termsof dot products. Therefore, we can compensate any invertiblelinear transformation of the data in by the correspondinginverse adjoint transformation on the set of admissible weightvectors in , i.e., for any invertible operator on , we have

for all (20)

Hence, we may construct a diagonal scaling operatorwhich inflates the sides of the above parallelepiped as muchas possible, while ensuring that it still lives in a sphere of theoriginal radius in (Fig. 1). This will not change thefactor on the right hand side of (18), but it buys us something

6In terms of entropy numbers, this is due to the tightness of the rate givenby a famous theorem of Maurey (e.g., [10]).

Page 4: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1003

Fig. 1. Since everything is done in terms of dot products, scaling up the data byA can be compensated by scaling the weight vectors withA�1: BychoosingA such that the data is still contained in a ball of the same radiusR; we effectively reduce our function class (parameterized by the weight vector),which leads to better generalization bounds which depend on the kernel inducing the map�:

regarding the second factor: one can show that the functionclass essentially behaves as if it was finite-dimensional, witha cut-off determined by the rate of decay of’s eigenvalues.

The reasoning that leads to the improved bounds is some-what intricate and cannot presently be explained in full detail.In a nutshell, the idea is to compute the capacity (measured interms of covering numbers of the SV function class, evaluatedon an -sample) via the entropy numbers of a suitable linearoperator.7 Using , consider first the operator

(21)

For our purpose, the entropy numbers of applied toa sphere of radius are crucial. These can be computedas the entropy numbers of the operator Usingfactorization properties of entropy numbers, these can be upperbounded taking into account the above scaling operatorin a rather precise way. The faster the eigenvalues of theintegral operator associated with a given kernel decay, thesmaller the entropy numbers (and hence the capacity) of thecorresponding feature space algorithm function class, and thestronger the generalization error bounds that one can prove.

As an example, we now consider how the entropy numbersdepend asymptotically on the eigenvalues of

Proposition 4 (Exponential-Polynomial Decay [38]):Suppose is a Mercer kernel with for some

ThenAn example of such a kernel (for is the Gaussian

This proposition allows the formulation ofa priori gener-alization error bounds depending on the eigenvalues of thekernel. Using similar entropy number methods, it is alsopossible to give rather precise data-dependent bounds in termsof the eigenvalues of the kernel Gram matrix [27].

7Consider two normed spacesE andF: Forn 2 , thenth entropy numberof a setM � E is defined as�n(M) := inf f�> 0: there exists an�-coverfor M in E containingn or fewer pointsg: (Recall that the covering numberN (�;M), being essentially its functional inverse, measures how many balls ofradius� one needs to coverM:) The entropy numbers�n(T ) of an operatorT : E ! F are defined as the entropy numbers of the image of the unit ballunderT: Note that�1(T ) = jjT jj; intuitively, the higher entropy numberallow a finer characterization of the complexity of the image ofT (e.g., [10],[38]). Note that entropy numbers have some nice properties that coveringnumbers are lacking. For instance, scaling a subset of a normed vector spaceby some factor simply scales its entropy numbers by the same factor.

Entropy numbers are a promising tool for studying thecapacity of feature space methods. This is due to the fact thatin the linear case, which is what we are interested in for featurespace algorithms, they can be studied using powerful methodsof functional analysis (e.g., [10]).

E. The Metric of the Kernel Map

Another way to gain insight into the structure of featurespace is to consider the intrinsic shape of the manifold towhich one’s data is mapped. It is important here to distinguishbetween the feature spaceand the surface in that space towhich points in input space actually map, which we willcall In general will be an dimensional submanifoldembedded in For simplicity, we assume here that issufficiently smooth that structures such as a Riemannian metriccan be defined on it. Here we will follow the analysis of [8],to which the reader is referred for more details, although theapplication to the class of inhomogeneous polynomial kernelsis new.

We first note that all intrinsic geometrical properties ofcan be derived once we know the Riemannian metric inducedby the embedding of in The Riemannian metric can bedefined by a symmetric metric tensor Interestingly, wedo not need to know the explicit mappingto constructit can be written solely in terms of the kernel. To see thisconsider the line element

(22)

where indexes correspond to the vector spaceandto input space Now letting represent a small but finitedisplacement in , we have

Thus we can read off the components of the metric tensor

(23)

For the class of kernels which are functions only of dot prod-ucts between points in both covariant and contravariant

Page 5: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1004 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

components of the metric take a simple form

(24)

(25)

where the prime denotes derivative with respect to the argu-ment To illustrate, let us compute the intrinsic geomet-rical properties of the surface corresponding to the class ofinhomogeneous polynomial Mercer kernels

(26)

where is a constant. Properties relating to the intrinsiccurvature of a surface are completely captured by the Riemanncurvature tensor

(27)

where the Christoffel symbols of the second kind are definedby

(28)

Thus we find that for this class of kernels, the Riemann cur-vature, for arbitrary input space dimensionand polynomialorder , is given by

(29)

It is interesting to compare this result with that for thehomogeneous case [8]

(30)

This analysis shows that adding the constantto thekernel results in striking differences in the geometries. Bothcurvatures vanish for , as expected. However for ,

vanishes for all powers whereasdoes not vanish for any Furthermore, all surfaces forhomogeneous kernels with nonvanishing curvature (i.e., thosewith and have a singularity at , whereasnone of the corresponding surfaces for inhomogeneous kernelsdo, provided 8

Thus beyond providing insight into the geometrical structureof the surfaces in generated by choice of kernel, the geomet-rical analysis also gives more concrete results. For example,

8Note that ifc< 0 the kernel is not a Mercer kernel and the above analysisdoes not apply.

one can expect that the density on the surfacewill becomeill behaved for data whose norm is small, for homogeneouspolynomial kernels, but not for the inhomogeneous case.Similar problems may arise in the pattern recognition caseif one’s data lies near the singularity. These considerationscan be extended to compute the intrinsic volume element on

, which may be used to compute the density on , andto give some simple necessary tests that must be satisfied byany kernel if it is to be a Mercer kernel: the reader is referredto [8] for details.

III. FEATURE SPACE ALGORITHMS

We next describe the two algorithms used in this paper: SVmachines, and kernel PCA. The SV algorithm shall not bedescribed in detail, let us only briefly fix the notation.

SV classifiers[34] construct a maximum margin hyperplanein In input space, this corresponds to a nonlinear decisionboundary of the form

(31)

where the are the training examples. Those with arecalled SV’s; in many applications, most of the, which arefound by solving a quadratic program, turn out to be zero.Excellent classification accuracies in both OCR and objectrecognition have been obtained using SV machines [24], [7].A generalization to the case of regression estimation, leadingto similar function expansion, exists [34], [26].

Kernel Principal Component Analysis[29] carries out alinear PCA in the feature space The extracted features takethe nonlinear form

(32)

where, up to a normalization, the are the components ofthe th eigenvector of the matrix

This can be understood as follows. We wish to find eigen-vectors and eigenvalues of the covariance matrix inthe feature space,9 where

(33)

In the case when is very high dimensional this will beimpossible to compute directly. To be still able to solve thisproblem, one uses Mercer kernels. To this end, we need toderive a formulation which only uses in dot products. Wethen replace any occurrence of by Thisway we avoid dealing with the mapped data explicitly, whichmay be intractable in terms of memory and computational cost.

To find a formulation for PCA which uses only dot prod-ucts, we first substitute the covariance matrix (33) into theeigenequation Note that all solutions to this

9Here we assume that the mapped data is centered, too. In general this willnot be true, but all computations can easily be reformulated to perform anexplicit centering inF [29].

Page 6: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1005

equation with must lie in the span of -images of thetraining data. Thus, we may consider the equivalent system

for all (34)

and expand the solution as

(35)

Substituting (33) and (35) into (34), and defining a matrixby we arrive at a problem which is cast

in terms of dot products: solve

(36)

where (for details on the last step see [29]).Normalizing the solution in translates into

To extract features, we compute the projection of the-image of a test point onto the th eigenvector in the featurespace by

(37)

Usually, this will be much cheaper than taking the dot productin the feature space explicitly.

To conclude the brief summary of kernel PCA, we statea characterization which involves the same regularizer (thelength of the weight vector) as the one used in SV machines.

Proposition 5: For all , the th kernel PCA feature extrac-tor, scaled by , is optimal among all feature extractors ofthe form

in the sense that it has the shortest weight vector

subject to the conditions that

1) it is orthogonal to the first kernel PCA featureextractors (in feature space)

2) applied to the training set it leads to a unitvariance set of outputs.

Both SV machines and kernel PCA utilize Mercer kernels togeneralize a linear algorithm to a nonlinear setting; moreover,both use the same regularizer, albeit in different domains oflearning—supervised versus unsupervised. Nevertheless, fea-ture extraction experiments on handwritten digit images usingkernel PCA have shown that a linear hyperplane classifiertrained on the extracted features can perform as well as anonlinear SV machine trained directly on the inputs [29].

Fig. 2. The preimage problem. Not each point in the span of the mappedinput data is necessarily the image of some input pattern. Therefore, not eachpoint� that can be written as an expansion in terms of mapped input patterns(e.g., a kernel PCA eigenvector, or a SVM hyperplane normal vector), cannecessarily be expressed as the image of a single input pattern.

IV. FROM FEATURE SPACE TO INPUT SPACE

Unlike Section II, which described how to get from inputspace into feature space, we now study the way back. Therehas been a fair amount of work on aspects of this problemin the context of developing so-called reduced set methods(e.g., [6], [9], [25], [12], [22]). For pedagogical reasons, weshall postpone reduced set methods to Section V, as they focuson a problem that is already more complex than the one wewould like to start with.

A. The Preimage Problem

As stated in the introduction, feature space algorithmsexpress their solutions as expansions in terms of mapped inputpoints (3). However, since the map into the feature space

is nonlinear, we cannot generally assert that each suchexpansion will have a preimage under, i.e., a pointsuch that (Fig. 2). If the preimage existed, it wouldbe easy to compute, as shown by the following result [24].

Proposition 6: Consider a feature space expansionIf there exists a such that

and if is an invertible function of , then we cancompute as

where is any orthonormal basis of input space.Proof: We expand as

(38)

Page 7: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1006 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

Several remarks on this proposition should be given. First,examples of kernels which are invertible functions ofare polynomial kernels

odd (39)

and sigmoid kernels

(40)

A similar result holds for RBF kernels (using the polarizationidentity)—all we need is a kernel which allows the reconstruc-tion of from , evaluated on some input points whichwe are allowed to choose (for details, cf., [24]).

The crucial assumption, clearly, is theexistenceof thepreimage. Unfortunately, there are many situations where thereare no preimages. To illustrate this, we consider the featuremap (11). Clearly, only points in feature space which canbe written as do have a preimage under this map. Tocharacterize this set of points in a specific example, considerGaussian kernels

(41)

In this case, maps each input into a Gaussian sitting onthat point. However, it is known [19] that no Gaussian can bewritten as a linear combination of Gaussians centered at otherpoints. Therefore, in the Gaussian case, none of the expansions(3), excluding trivial cases with only one term, has an exactpreimage.

The problem that we had initially set out to solve has turnedout to be insolvable in the general case. Let us try to askfor less. Rather than trying to find exact preimages, we nowconsider approximate ones. We call an approximatepreimagefor if

(42)

is small.10

Are there vectors for which good approximate preimagesexist? As described in Section III, kernel PCA is nothingbut PCA in Therefore, for it providesprojections

(43)

with the following optimal approximation property (e.g., [24]).Here, we assume that the are sorted according to nonin-creasing eigenvalues with being the smallest nonzeroeigenvalue.

Proposition 7: is the -dimensional projection minimiz-ing

(44)

Therefore, can be expected tohave a good approx-imate preimage: trivially, already is a good approximate

10Just how small it needs to be in order to form a satisfactory approximationwill depend on the problem at hand. Therefore, we have refrained from givinga formal definition.

preimage. As we shall see in experiments, however, evenbetter preimages can be found, which makes some interestingapplications possible [26], [21]:

Denoising: Given a noisy , map it into , discardhigher components to obtain , and then compute apreimage Here, the hope is that the main structure in thedata set is captured in the firstdirections, and the remainingcomponents mainly pick up the noise—in this sense,can bethought of as a denoised version of

Compression:Given the eigenvectors and a small num-ber of features [cf., (37)] of , but not , compute apreimage as an approximate reconstruction ofThis is usefulif is smaller than the dimensionality of the input data.

Interpretation: Visualize a nonlinear feature extractorby computing a preimage.

In this paper, we focus on the first point. In the next section,we shall develop a method for minimizing (42), which wewill later, in the experimental section, apply to the case where

B. An Algorithm for Approximate Preimages

The present section [25] gives an analysis for the case ofthe Gaussian kernel, which has proven to perform very wellin applications [30], and proposes an iteration procedure forcomputing preimages of kernel expansions.

We start by considering a problem slightly more generalthan the preimage problem: we are seeking to approximate

by First observe that ratherthan minimizing

(45)

we can minimize the distance betweenand the orthogonalprojection of onto span (Fig. 3),

(46)

To this end, we maximize

(47)

which can be expressed in terms of the kernel. The maximiza-tion of (47) over is preferable to the one of (45) overand

, since it comprises a lower-dimensional problem, and sinceand have different scaling behavior. Once the maximum

of (47) is found, it is extended to the minimum of (45) bysetting [cf. (46)] The function(47) can either be minimized using standard techniques (asin [6]), or, for particular choices of kernels, using fixed-pointiteration methods, as shown presently.

For kernels which satisfy for all (e.g.,Gaussian kernels), (47) reduces to

(48)

Page 8: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1007

Fig. 3. Given a vector 2 F , we try to approximate it by a multiple ofa vector�(zzz) in the image of input space( N ) under the nonlinear map� by finding zzz such that the projection distance of� onto span(�(zzz)) isminimized.

For the extremum, we haveTo evaluate the gradient in terms of

, we substitute (54) to get the sufficient conditionFor [e.g., Gaus-

sians, or for , we obtain, leading to

(49)

For the Gaussian kernelwe thus arrive at

(50)

and devise an iteration

(51)

The denominator equals and thus is nonzero in aneighborhood of the extremum of (48), unless the extremumitself is zero. The latter only occurs if the projection ofonthe linear span of is zero, in which case it is pointlessto try to approximate Numerical instabilities related to

being small can thus be approached by restartingthe iteration with different starting values.

Interestingly, (51) can be interpreted in the context ofclustering (e.g., [5]). It determines the center of a singleGaussian cluster, trying to capture as many of thewithpositive as possible, and simultaneously avoids thosewith negative For SV classifiers, the sign of theequals the label of the pattern It is this sign whichdistinguishes (51) from plain clustering or parametric densityestimation. The occurrence of negative signs is related to thefact that we are not trying to estimate a parametric density but

the differencebetween two densities (modulo normalizationconstants).

To see this, we define the setsand and the shorthands

andThe target (48) then

reads i.e., we are trying to find apoint where the difference between the (unnormalized)“probabilities” for the two classes is maximal, and estimatethe approximation to (54) by a Gaussian centered atMoreover, note that we can rewrite (50) as

(52)

where

(53)

V. REDUCED SET (RS) METHODS

A. The Problem

We now move on to a slightly more general problem, firststudied by [6], where we are no longer only looking for singlepreimages, but expansions. It will turn out that one can alsouse the method developed in the last section to design analgorithm for the more general case.

Assume we are given a vector , expanded in imagesof input patterns

(54)

with , Rather than looking for a singlepreimage, we now try to approximate it by areduced setexpansion [6]

(55)

with , , To this end, one canminimize [6]

(56)

The crucial point is that even if is not given explicitly, (56)can be computed (and minimized) in terms of the kernel.

In the NIST benchmark of 60 000 handwritten digits, SVmachines are more accurate than any other single classifier[24]; however, they are inferior to neural nets in run-timeclassification speed [9]. In applications where the latter is an

Page 9: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1008 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

issue, it is thus desirable to come up with methods to speed upthings by making the SV expansion more sparse, i.e., replacing(54) by (55).

B. Finding the Coefficients

Evidently, the RS problem consists of two parts. One hasto determine the RS vectors, and one has to compute theexpansion coefficients We start with the latter; partly, as itis easier, partly, as it is common to different RS methods.

Proposition 8: The optimal coefficientsfor approximating by (forlinearly independent in the 2-norm aregiven by

(57)

Here, andNote that if the are linearly independent, as they

should be if we want to use them for approximation, thenhas full rank. Otherwise, one can use the pseudoinverse,

or select the solution which has the largest number of zerocomponents.

Proof: See [25]. We evaluate the derivative of the dis-tance in

(58)

and set it to zero. Substituting , we obtain(using

(59)

hence

(60)

No RS algorithm using the 2-norm optimality criterion cancircumvent this result. For instance, suppose we are givenan algorithm that computes the and simultaneouslyand comes up with a solution. Then we can always use theproposition to recompute the optimal coefficients to get asolution which is at least as good as the original one. Differentalgorithms can, however, differ in the way they determinethe vectors in the first place. The one dealt with in thenext section simplyselectssubsets of the , while the one inSection V-D uses vectors different from the original

C. Reduced Set Selection

1) Selection via Kernel PCA:The idea for the first algo-rithm arises from the observation that the null space ofthe Gram matrix precisely tells ushow many vectors can be removed from an expansion whilecommitting zero approximation error (assuming we correctlyadjust the coefficients), i.e., how sparse we can make an SVexpansion, say, without changing it the least [24], [12]. (Here,

the Gram matrix can either be computed only from thoseexamples which have a nonzero, or from a larger set whichwe want to use for the expansion.) Interestingly, it will turnout that this problem is closely related to kernel PCA.

Let us start with the simplest case. Assume there exists aneigenvector of with eigenvalue 0, i.e.,Using (2), this reads

for all (61)

hence

(62)

This means that the are linearly dependent, and there-fore any of the which comes with a nonzero canbe expressed in terms of the others. Hence, we may use theeigenvectors with eigenvalue zero to eliminate certain termsfrom any expansion in the

What happens if we do not have nonzero eigenvalues,such as in the case of Gaussian kernels [19]? Intuitively,we would still believe that even though the above is nolonger precisely true, it should give a good approximation.However, the crucial difference is that in order to get the bestpossible approximation, we now need to take into accountthe coefficients of the expansion if wecommit an error by removing , say, then this error willdepend on too. How do we then select the optimal

Clearly, we would like to find coefficients minimizing theerror we commit by replacing with

(63)

To establish a connection to kernel PCA, we make a change ofvariables. First, define for , for

Hence (63) equals Normalizingto obtain , hence , this leads to theproblem of minimizing

(64)

over (note that is invariant when rescaling 11

A straightforward calculation shows that we can recover theapproximation coefficients for , i.e., the values toadd to the for leaving out , as

,Rather than minimizing the nonlinear function (64), we now

devise a computationally attractive approximate solution. It ismotivated by the observation that alone is minimizedfor the eigenvector with minimal eigenvalue, consistent withthe special case discussed above [cf., (62)]. In that case,

11The idea of approximating the support vector expansion by optimallyremoving individual support vectors, and then adjusting the coefficients ofthose that remain to minimize the resulting error, was arrived at independentlyby Olivier Chapelle, who also derived the expression to be minimized, (64)(private communication).

Page 10: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1009

More generally, if is any normalizedeigenvector of , with eigenvalue , then

(65)

This can be minimized in operations by performingkernel PCA and scanning through the matrix Thecomplexity can be reduced to by only considering thesmallest eigenvalues, with chosena priori. Hence,we can eliminate , with chosen in a principled yetefficient way.

Setting all computational considerations aside, theoptimalgreedy solution to the above selection problem, equivalent to(64), can also be obtained by using Proposition 8: computethe optimal solution for all possible patterns that one couldleave out (i.e., use subsets of of size as

and evaluate (58) in each case.The same applies to subsets of any size. If we have the

resources to exhaustively scan through all subsets of size, then Proposition 8 provides theoptimal

way of selecting the best expansion of a given size. Betterexpansions can only be obtained if we drop the restriction thatthe approximation is written in terms of the original patterns,as done in Section V-D.

No matter how we end up choosing, we approximate by

(66)

The whole scheme can be iterated until the expansion ofis sufficiently sparse. If one wants to avoid having to find thesmallest eigenvalue of at each step anew, then approximateschemes using heuristics can be conceived.

In our experiments to be reported below, we did computeall eigenvectors at each step, using the Gram matrix computedfrom the SV’s and then selectedaccording to (65).

2) Selection via Penalization: We next consider amethod for enforcing sparseness which is inspired byshrinkage penalizers (cf., [13]).

Given some expansion , we approximate it byby minimizing

(67)

over all Here, is a constant determining the tradeoffbetween sparseness and quality of approximation. The con-stants can be set to one or (where is the mean ofall say. In the latter case, we are hoping for a sparserdecomposition, since more emphasis is put on shrinking termswhich are already small. This reflects the intuition that it isless promising to try to shrink very large terms. Ideally, wewould like to countthe number of nonzero coefficients, ratherthan sum their moduli; however, the former does not lead toan efficiently solvable optimization problem.

To dispose of the modulus, we rewrite as

(68)

where In terms of the new variables, we end up withthe quadratic programming problem

(69)

subject to

This problem can be solved with standard quadratic pro-gramming tools. The solution (68) could be used directly asexpansion coefficients. For optimal precision, however, wemerely use it to select which patterns to use for the expansion(those with nonzero coefficients), and recompute the optimalcoefficients using Proposition 8.

3) The Multiclass Case:In many applications, we face theproblem of simultaneouslyapproximating a set of featurespace expansions. For instance, in digit classification, a com-mon approach is to train binary recognizers, one foreach digit. To this end, the quadratic programming formulationof Section V-C.2 can be modified to

(70)

subject to

(71)

The indexes and are understood to range over all SV’s (i.e.,expansion vectors which have a nonzero coefficient for at leastone of the recognizers), and over theclasses, respectively.For , we use either one or, with the same rationale as above,

(here, is the mean of all overThe term together with the constraint (71) ensures

that we only penalize the largest of the coefficients pertainingto each individual SV. Therefore, as soon as one coefficientis being used by one the expansions, it does not cost anythingto also use it for another one. This is precisely what we wantfor speed-up purposes: if we have to compute a certain dotproduct anyway, we might just as well reuse it for the otherexpansions.

Using , with , we arrive at thefollowing problem:

(72)

subject to

(73)

Here, ranges over all SV’s.

Page 11: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1010 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

4) A Note on the Utility of Reduced Set Selection:Whyshould we expect these procedures to be useful? First, for SVmachines, where for some positive value of theregularization constant , there is reason to believe that theSV expansion can be made sparser by removing the constrainton [note that (66) does not care about the constraint].12

Second, the number of eigenvalues which are zero (and thusthe number of patterns that can be removed without loss), orat least small, depends on the problem at hand and the kernelused. For instance, Gaussian kernel Gram matrices do not havezero eigenvalues unless some of the patterns are duplicates[19]. Nevertheless, good approximations are possible, sincethe eigenvalues of Gaussian kernels decay rapidly (e.g., [38]).

D. Reduced Set Construction

So far, we have dealt with the problem of how toselectareduced set of expansion vectors from the original one. Wenow return to the originally posed problem, which includesthe constructionof new vectors to reach high reduction rates.To this end, suppose we want to approximate a vector

(74)

by an expansion of the type (55) with To this end,we iterate the procedure of Section IV-B by

Here, denotes the found for , obtained byiterating (51). To apply (51), needs to be utilized in itsrepresentation in terms of mapped input images

(75)

i.e., we need to set ,

The coefficient could be computed asHowever, if the vectors are not

orthogonal in , then the best approximation of in theirspan is not obtained by computing orthogonal projectionsonto each direction. Instead, we need to compute the optimalcoefficients anew in each step, usingProposition 8 (if the discrepancy has not yet reachedzero, then will be invertible).

The iteration is stopped after steps, either specifiedin advance, or by monitoring when (i.e.,

falls below a specified threshold. The solutionvector takes the form (55).

We conclude this section by noting that in many cases, suchas multiclass SV machines, or multiple kernel PCA featureextractors, we may actually want to approximate severalvectors simultaneously. This leads to more complex equations,given in [25].

12Imagine a case where a certain pattern appears twice in the training set,and the SV expansion has to utilize both of the copies only because the upperbound constraint limits the coefficient of each of them toC:

Fig. 4. Kernel PCA toy example (see text): lines of constant featurevalue for the first eight nonlinear principal components extracted withk(xxx; yyy) = exp (�jjxxx � yyyjj2=0:1): The first two principal components (topmiddle/right) separate the three clusters. Components 3–5 split the clusters.Components 6–8 split them again, orthogonal to the above splits.

VI. EXPERIMENTS

To see how the proposed methods work in practice we ranseveral toy and real-world experiments. In Section VI-A, wegive denoising results for the approach of finding approxi-mate preimages presented in Section IV-A. In Section VI-B,we present some experiments for the reduced set methodsdescribed in Sections V-C and V-D.

A. Kernel PCA Denoising

1) Toy Examples:All experiments reported were carriedout with Gaussian kernels (41), minimizing (42) with theiteration scheme given by (51). However, similar resultswere obtained with polynomial kernels. Matlab code forcomputing kernel PCA is available on the web fromhttp://svm.first.gmd.de.

We generated an artificial data set from three point sourcesat ( 0.5, 0.1), (0, 0.7), (0.5, 0.1) (100 points each) withGaussian noise ( ), and performed kernel PCA onit (Fig. 4). Using the resulting eigenvectors, we extractednonlinear principal components from a set of test pointsgenerated from the same model, and reconstructed the pointsfrom varying numbers of principal components. Fig. 5 showsthat discarding higher order components leads to removal ofthe noise—the points move toward their respective sources.

In a second experiment (Table I), we generated a data setfrom 11 Gaussians in with zero mean and variancein each component, by selecting from each source 100 pointsas a training set and 33 points for a test set (centers of theGaussians randomly chosen in [1, 1] Then we appliedkernel PCA to the training set and computed the projections

of the points in the test set. With these, we carried outdenoising, yielding an approximate preimage in for eachtest point. This procedure was repeated for different numbersof components in reconstruction, and for different values of(also used in the kernel). We compared the results providedby our algorithm to those of linear PCA via the mean squareddistance of all denoised test points to their correspondingcenter. Table I shows theratio of these values; here and below,ratios larger than one indicate that kernel PCA performedbetter than linear PCA. For almost every choice ofand

Page 12: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1011

Fig. 5. Kernel PCA denoising by reconstruction from projections onto theeigenvectors of Fig. 4. We generated 20 new points from each Gaussian,represented them in feature space by their firstn = 1; 2; � � � ; 8 nonlinearprincipal components, and computed approximate preimages, shown in theupper nine pictures (top left: original data, top middle:n = 1, top right:n = 2, etc.). Note that by discarding higher order principal components (i.e.,using a smalln), we remove the noise inherent in the nonzero variance�

2 ofthe Gaussians. The lower nine pictures show how the original points “move”in the denoising. Unlike the corresponding case in linear PCA, where weobtain lines (see Fig. 6), in kernel PCA clusters shrink to points.

Fig. 6. Reconstructions and point movements for linear PCA, based on thefirst principal component.

, kernel PCA did better. Note that using all ten components,linear PCA is just a basis transformation and hence cannotdenoise. The extreme superiority of kernel PCA for smallis due to the fact that all test points are in this case locatedclose to the 11 spots in input space, and linear PCA has tocover them with less than 10 directions. Kernel PCA moveseach point to the correct source even when using only a smallnumber of components.

To get some intuitive understanding in a low-dimensionalcase, Fig. 7 depicts the results of denoising a half circleand a square in the plane, using kernel PCA, anonlinearautoencoder, principal curves, and linear PCA. The principalcurves algorithm [14] iteratively estimates a curve captur-ing the structure of the data. The data are projected tothe closest point on a curve which the algorithm tries toconstruct such that each point is the average of all datapoints projecting onto it. It can be shown that the only

TABLE IDENOISING GAUSSIANS IN 10 (SEE TEXT). PERFORMANCERATIOS LARGERTHAN

ONE INDICATE HOW MUCH BETTER KERNEL PCA DID, COMPARED TO LINEAR

PCA, FOR DIFFERENT CHOICES OF THEGAUSSIANS’ STANDARD DEVIATION �;

AND DIFFERENT NUMBERS OF COMPONENTSUSED IN RECONSTRUCTION

Fig. 7. Denoising in 2-D (see text). Depicted are the data set (small points)and its denoised version (big points, joining up to solid lines). For linearPCA, we used one component for reconstruction, as using two components,reconstruction is perfect and thus does not denoise. Note that all algorithmsexcept for our approach have problems in capturing the circular structure inthe bottom example (taken from [21]).

straight lines satisfying the latter are principal components,so principal curves are a generalization of the latter. Thealgorithm uses a smoothing parameter which is annealedduring the iteration. In the nonlinear autoencoder algorithm,a “bottleneck” five-layer network is trained to reproduce theinput values as outputs (i.e., it is used in autoassociative mode).The hidden unit activations in the third layer form a lower-dimensional representation of the data, closely related to PCA(see, for instance, [11]). Training is done by conjugate gradientdescent. In all algorithms, parameter values were selectedsuch that the best possible denoising result was obtained.The figure shows that on the closed square problem, kernelPCA does (subjectively) best, followed by principal curvesand the nonlinear autoencoder; linear PCA fails completely.However, note that all algorithms except for kernel PCAactually provide an explicit one-dimensional parameterizationof the data, whereas kernel PCA only provides us with ameans of mapping points to their denoised versions (in thiscase, we used four kernel PCA features, and hence obtain afour-dimensional parameterization).

2) Handwritten Digit Denoising:To test our approach onreal-world data, we also applied the algorithm to the USPSdatabase of handwritten digits (e.g., [16], [24]) of 7291 trainingpatterns and 2007 test patterns (size 1616). For each of theten digits, we randomly chose 300 examples from the trainingset and 50 examples from the test set. We used the methodof Section IV-B, withThe width 0.5 equals twice the average of the data’s variance

Page 13: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1012 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

Fig. 8. Visualization of eigenvectors (see text). Depicted are the20; � � � ; 2

8th eigenvector (from left to right). First row: linear PCA, second and thirdrow: different visualizations for kernel PCA.

in each dimension. In Fig. 8, we give two possible depictionsof the eigenvectors found by kernel PCA, compared to thosefound by linear PCA for the USPS set. The second rowshows the approximate preimages of the eigenvectors,

, found by our iterative algorithm. In the third roweach image is computed as follows: Pixelis the projectionof the -image of the th canonical basis vector in input spaceonto the corresponding eigenvector in feature space [upper left( ), lower right ]. In the linear case, bothmethods would simply yield the eigenvectors of linear PCAdepicted in the first row; in this sense, they may be consideredas generalized eigenvectors in input space. We see that thefirst eigenvectors are almost identical (except for arbitrarysigns). However, we also see that eigenvectors in linear PCAstart to focus on high-frequency structures already at smallereigenvalue size. To understand this, note that in linear PCA weonly have a maximum number of 256 eigenvectors, contraryto kernel PCA which gives us the number of training examples(here 3000) possible eigenvectors.

This also explains some of the results we found when work-ing with the USPS set (Figs. 9 and 10). In these experiments,linear and kernel PCA were trained with the original data. Tothe test set, we added

1) additive Gaussian noise with zero mean and standarddeviation or

2) ‘speckle’ noise, where each pixel is flipped to black orwhite with probability

For the noisy test sets, we computed the projections ontothe first linear and nonlinear components, and carried outreconstruction for each case. The results were compared bytaking the mean squared distance of each reconstructed digitof the noisy test set to its original counterpart.

For the optimal number of components in linear and kernelPCA, our approach did better by a factor of 1.6 for theGaussian noise, and 1.2 for the “speckle” noise (the optimalnumber of components were 32 in linear PCA, and 512 and256 in kernel PCA, respectively). Taking identical numbers ofcomponents in both algorithms, kernel PCA becomes up toeight times better than linear PCA. However, note that kernelPCA comes with a higher computational complexity.

B. Speeding Up Support Vector Decision Rules

As in Section VI-A, we used the USPS handwritten digitdatabase. We approximated the SV expansions (31) of tenbinary classifiers, each trained to separate one digit from the

Fig. 9. Denoising of USPS data (see text). The left half shows:top: thefirst occurrence of each digit in the test set,second row:the upper digit withadditive Gaussian noise(� = 0:5), following five rows:the reconstruction forlinear PCA usingn = 1, 4, 16, 64, 256 components, and,last five rows:theresults of our approach using the same number of components. In the righthalf we show the same but for ‘speckle’ noise with probabilityp = 0:2:

rest. We used the Gaussian kernel, and the approximation techniques described

in Sections V-C and V-D.The original SV system had on average 254 SV’s per clas-

sifier. Tables II and III show the classification error results forapproximation using the reduced set selection techniques de-scribed in Section V-C, while Table IV gives results for usingthe reduced set construction method described in Section V-D,for varying numbers of RS vectors. Shown in each line is thenumber of misclassified digits for each single classifier and theerror of the combined 10-class machine. In all RS systems, theoptimal SV threshold was recomputed on the training set.

For the method of Section V-C, RSS-means, that foreach binary classifier, we removed support vectors until onaverage were left; for the method of section Section V-D,that we approximated each decision function usingvectors(i.e., steps using the described iteration procedure). For largenumbers , i.e., small reductions of the decision functionscomplexity, the accuracy of the original system can be ap-

Page 14: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1013

Fig. 10. Mean squared error of denoised images versus number of features used, for kernel PCA and linear PCA. Kernel PCA exploits nonlinearitiesand has the potential to utilize more features to code structure rather than noise. Therefore, it outperforms linear PCA denoising if a sufficiently largenumber of features is used.

TABLE IINUMBERS OF TEST ERRORS FOREACH BINARY RECOGNIZER, AND TEST ERROR

RATES FOR 10-CLASS CLASSIFICATION USING THE RS METHOD OF SECTION

SECTION V-C.1. TOP: NUMBERS OF SVS FOR THEORIGINAL SV RBF-SYSTEM.BOTTOM, FIRST ROW: ORIGINAL SV SYSTEM, WITH 254 SV’S ON AVERAGE;

FOLLOWING ROWS: SYSTEMS WITH VARYING AVERAGE NUMBERS OF RSVECTORS. IN THE SYSTEM RSS-n, EQUAL FRACTIONS OF SV’S WERE REMOVED

FROM EACH RECOGNIZER SUCH THAT ON AVERAGE, n RS VECTORSWERE LEFT

proached closely with both techniques. In Tables II and III, wesee that removal of about 40% of the support vectors leaves theerror practically unchanged. Reducing the numbers of supportvectors further can lead to large performance losses.

TABLE IIINUMBERS OF TEST ERRORS FOREACH BINARY RECOGNIZER, AND TEST ERROR

RATES FOR TEN-CLASS CLASSIFICATION USING THE RS METHOD OF SECTION

V-C.2 (ci = �=j�ij): FIRST ROW: ORIGINAL SV SYSTEM, WITH 254 SV’S ON

AVERAGE; FOLLOWING ROWS: SYSTEMS WITH VARYING AVERAGE NUMBERS OF

RS VECTORS. IN THE SYSTEM RSS2-n, � WAS ADJUSTED SUCH THAT THE

AVERAGE NUMBER OF RS VECTORSLEFT WAS n (THE CONSTANT �, GIVEN

IN PARENTHESES, WAS CHOSEN SUCH THAT THE NUMBERS n WERE

COMPARABLE TO TABLE II). THE RESULTS CAN BE FURTHER IMPROVED

USING THE METHOD OF SECTION V-C3 (ci = �=maxj j�ji j): FOR

INSTANCE, USING ABOUT 570 EXPANSION VECTORS(WHICH IS THE SAME

NUMBER THAT WE GET WHEN TAKING THE UNION OF ALL SV’S IN THE

RSS2 � 74 SYSTEM), THIS LED TO AN IMPROVED ERROR RATE OF 5.5%

The reduced set construction method, which is computa-tionally and conceptually more complex, performs better inthis situation as it is able to utilize vectors different from theoriginal support patterns in the expansion. To get a speedupby a factor of ten, we have to use a system with 25 RS vectors(RSC-25). For the method in Table IV, the classificationaccuracy only drops moderately from 4.4% to 5.1%, which

Page 15: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1014 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

Fig. 11. Complete display of reduced set vectors constructed by the iterative approach of Section V-D forn = 20, with coefficients (Top: recognizer ofdigit 0, � � �, bottom: digit 9). Note that positive coefficients (roughly) correspond to positive examples in the classification problem.

TABLE IVNUMBER OF TEST ERRORS FOREACH BINARY RECOGNIZER, AND TEST ERROR

RATES FOR TEN-CLASS CLASSIFICATION USING THE RS CONSTRUCTION

METHOD OF SECTION V-D. FIRST ROW: ORIGINAL SV SYSTEM, WITH

254 SV’S ON AVERAGE (SEE ALSO TABLES II AND III); F OLLOWING

ROWS: SYSTEMS WITH VARYING NUMBERS OF RS VECTORS(RSC-nSTANDS FORn VECTORSCONSTRUCTED) PER BINARY RECOGNIZER,

COMPUTED BY ITERATING ONE-TERM APPROXIMATIONS, SEPARATELY FOR

EACH RECOGNIZER. LAST TWO ROWS: WITH A SUBSEQUENT GLOBAL

GRADIENT DESCENT, THE RESULTS CAN BE FURTHER IMPROVED (SEE TEST)

is still competitive with convolutional neural networks on thatdata base [16]. Moreover, we can further improve this resultby adding the second phase of the traditional RS algorithm,where a global gradient descent is performed in the spaceof all [6], [9] (computationally more expensive than

the first phase by about two orders of magnitude): this ledto an error rate of 4.7%. For the considered kernel, this isalmost identical to the traditional RS method, which yielded5.0% (for polynomial kernels, the latter method led to 4.3%at the same speedup [6]). Note that the traditional RS methodrestarts the second phase many times to make sure that theglobal minimum of the cost function is actually found, whichmakes it plausible that the final results are similar.

Finally, Fig. 11 shows the RS-20 vectors of the ten binaryclassifiers. As an aside, note that unlike the approach of [6],our algorithm produces images which do look meaningful (i.e.,digit-like).

VII. D ISCUSSION

Algorithms utilizing Mercer kernels construct their solutionsas expansions in terms of mapped inputpatterns. However, the map is often unknown, or toocomplex to provide any intuition about the solution Thishas motivated our efforts to reduce the complexity of theexpansion, summarized in this paper.

As an extreme case, we have first studied how to approx-imate by a single (i.e., how to find an approximatepreimage of ), and proposed a fixed-point iteration algorithmto perform the task.

In situations where no good approximate preimage exists,one can still reduce the complexity of byexpressing it as a sparser expansion Wehave proposed methods for computing the optimal coefficients

and for coming up with suitable patterns either byselecting among the or by iterating the above preimagealgorithm.

Page 16: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1015

Both types of approximations are of theoretical interest forfeature space methods; however, they also lead to practicalapplications. In this paper, we have considered two suchapplications, namely, the problem of statistical denoising viakernel PCA reconstruction, and the problem of speeding upSV decision rules. We address them in turn.

Kernel PCA Denoising:In denoising experiments on real-world data, we obtained results significantly better than usinglinear PCA. Our interpretation of this finding is as follows.Linear PCA can extract at most components, where isthe dimensionality of the data. Being a basis transform, allcomponents together fully describe the data. If the data arenoisy, this implies that a certain fraction of the componentswill be devoted to the extraction of noise. Kernel PCA, on theother hand, allows the extraction of up tofeatures, where isthe number of training examples. Accordingly, kernel PCA canprovide a larger number of features carrying information aboutthe structure in the data (in our experiments, we had ).In addition, if the structure to be extracted is nonlinear, thenlinear PCA must necessarily fail, as we have illustrated withtoy examples.

Open questions and problems include the choice of asuitable kernel for a given noise reduction problem, possiblyin conjunction with the regularization properties of the kernel(e.g., [38]), the application of the approach to compression,and the comparison (and connection) to alternative nonlineardenoising methods (cf., [17]).

Speeding Up SV Machines:We have shown experimentallythat our approximation algorithms can be used to speed up SVmachines significantly. Note that in the Gaussian RBF case,the approximation can never be as good as the original, sincethe kernel matrix has full rank [19].

As in [9], goodRS constructionresults were obtained eventhough the objective function did not decrease to zero (inour RS construction experiments, it was reduced by a factorof two–20 in the first phase, depending on how many RSvectors were computed; the global gradient descent yieldedanother factor two–three). We conjecture that this is due to thefollowing: in classification, we are not interested in ,but in

, where is the underlying probability distributionof the patterns (cf., [3]). This is consistent with the fact thatthe performance of a RS SV classifier can be improved byrecomputing an optimal threshold

The previous RS construction method [6], [9] can be usedfor any SV kernel; the new one is limited to

However, it is fast, and it led to interpretable RS imagesand an interesting connection between clustering and approx-imation in feature spaces. It appears intriguing to pursue thequestion whether this connection could be exploited to formmore general types of approximations of SV and kernel PCAexpansions by making use of Gaussians of variable widths.

The RS selectionmethods, on the other hand, are appli-cable for any SV kernel. In our experiments, they led toworse reduction rates than RS construction; however, theyare simpler and computationally faster. Among the first twoRS selection methods, the one described in Section V-C.1was slightly superior at higher reductions; however, the one

given in Section V-C.2 is computationally cheaper since unlikethe former, it does not remove the SV’s one at a time andtherefore it need not be iterated. Moreover, we found thatit can be improved by simultaneously approximating severalvectors, corresponding to the ten binary recognizers in a digitrecognition task.

The proposed methods are applicable to any feature spacealgorithm based on Mercer kernels. For instance, we couldalso speed up SV regression machines or kernel PCA featureextractors. Moreover, we expect further possibilities to open upin the future, as Mercer kernel methods are being applied in anincreasing number of learning and signal processing problems.

ACKNOWLEDGMENT

The authors would like to thank A. Elisseeff for helpfuldiscussions.

REFERENCES

[1] M. Aizerman, E. Braverman, and L. Rozonoer, “Theoretical founda-tions of the potential function method in pattern recognition learning,”Automat. Remote Contr., vol. 25, pp. 821–837, 1964.

[2] N. Aronszajn, “Theory of reproducing kernels,”Trans. Amer. Math. Soc.,vol. 68, pp. 337–404, 1950.

[3] K. L. Blackmore, R. C. Williamson, and I. M. Y. Mareels, “Decisionregion approximation by polynomials or neural networks,”IEEE Trans.Inform. Theory, vol. 43, pp. 903–907, 1997.

[4] B. E. Boser, I. M. Guyon, and V. N. Vapnik, “A training algorithm foroptimal margin classifiers,” inProc. 5th Annu. ACM Wkshp. Comput.Learning Theory, D. Haussler, Ed. Pittsburgh, PA: ACM, July 1992,pp. 144–152.

[5] J. M. Buhmann, “Data clustering and learning,” inThe Handbook ofBrain Theory and Neural Networks, M. A. Arbib, Ed. Cambridge, MA:MIT Press, 1995, pp. 278–281.

[6] C. J. C. Burges, “Simplified support vector decision rules,” inProc. 13thInt. Conf. Machine Learning, L. Saitta, Ed. San Mateo, CA: MorganKaufmann, 1996, pp. 71–77.

[7] , “A tutorial on support vector machines for pattern recognition,”Data Mining and Knowledge Discovery, vol. 2, no. 2, pp. 1–47, 1998.

[8] , “Geometry and invariance in kernel based methods,” inAd-vances in Kernel Methods—Support Vector Learning, B. Scholkopf, C.Burges, and A. Smola, Eds. Cambridge, MA: MIT Press, 1999, pp.89–116.

[9] C. J. C. Burges and B. Scholkopf, “Improving the accuracy and speed ofsupport vector learning machines,” inAdvances in Neural InformationProcessing Systems 9, M. Mozer, M. Jordan, and T. Petsche, Eds.Cambridge, MA: MIT Press, 1997, pp. 375–381.

[10] B. Carl and I. Stephani,Entropy, Compactness, and the Approximationof Operators. Cambridge, U.K.: Cambridge Univ. Press, 1990.

[11] K. I. Diamantaras and S. Y. Kung,Principal Component Neural Net-works. New York: Wiley, 1996.

[12] T. Frieß and R. F. Harrison, “Linear programming support vectormachines for pattern classification and regression estimation; and theSR algorithm: Improving speed and tightness of VC bounds in SV al-gorithms,” University of Sheffield, U.K., Department Automatic Controland Systems Engineering, Res. Rep. 706, 1998.

[13] F. Girosi, “An equivalence between sparse approximation and supportvector machines,”Neural Comput., vol. 10, no. 6, pp. 1455–1480, 1998.

[14] T. Hastie and W. Stuetzle, “Principal curves,”JASA, vol. 84, pp.502–516, 1989.

[15] T. J. Hastie and R. J. Tibshirani,Generalized Additive Models, vol. 43 ofMonographs on Statistics and Applied Probability. London: Chapman& Hall, 1990.

[16] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W.Hubbard, and L. J. Jackel, “Backpropagation applied to handwritten zipcode recognition,”Neural Comput., vol. 1, pp. 541–551, 1989.

[17] S. Mallat and Z. Zhang, “Matching Pursuit in a time-frequency dictio-nary,” IEEE Trans. Signal Processing, vol. 41, pp. 3397–3415, 1993.

[18] D. Mattera, personal communication, 1998.[19] C. A. Micchelli, “Interpolation of scattered data: Distance matrices and

conditionally positive definite functions,”Constructive Approximation,vol. 2, pp. 11–22, 1986.

[20] S. Mika, “Nichtlineare Signalverarbeitung in Feature-Raumen,” Diplom

Page 17: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

1016 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999

degree thesis, Technische Universit¨at Berlin, Germany, 1998 (in Ger-man).

[21] S. Mika, B. Scholkopf, A. Smola, K.-R. Muller, M. Scholz, and G.Ratsch, “Kernel PCA and denoising in feature spaces,” inAdvancesNeural Inform. Processing Syst. 11, 1999.

[22] E. Osuna and F. Girosi, “Reducing run-time complexity in supportvector machines,” inAdvances in Kernel Methods—Support VectorLearning, B. Scholkopf, C. Burges, and A. Smola, Eds. Cambridge,MA: MIT Press, 1999, pp. 271–283.

[23] S. Saitoh,Theory of Reproducing Kernels and its Applications. Harlow,U.K.: Longman, 1988.

[24] B. Scholkopf, Support Vector Learning, R. Oldenbourg Verlag,Munchen, Doktor degree thesis, TU Berlin, 1997 (in German).

[25] B. Scholkopf, P. Knirsch, A. Smola, and C. Burges, “Fast approxi-mation of support vector kernel expansions, and an interpretation ofclustering as approximation in feature spaces,” inMustererkennung1998—20. DAGM-Symp., P. Levi, M. Schanz, R.-J. Ahlers, and F. May,Eds., Informatik aktuell Berlin, Germany: Springer-Verlag, 1998, pp.124–132.

[26] B. Scholkopf, S. Mika, A. Smola, G. Ratsch, and K.-R. Muller, “KernelPCA pattern reconstructionvia approximate preimages,” inProc. 8th Int.Conf. Artificial Neural Networks, Perspectives in Neural Computing, L.Niklasson, M. Bod´en, and T. Ziemke, Eds. Berlin, Germany: Springer-Verlag, 1998, pp. 147–152.

[27] B. Scholkopf, J. Shawe-Taylor, A. J. Smola, and R. C. Williamson,“Kernel-dependent support vector error bounds,” inProc. ICANN, 1999,to be published.

[28] B. Scholkopf, P. Simard, A. Smola, and V. Vapnik, “Prior knowledge insupport vector kernels,” inAdvances in Neural Information ProcessingSystems 10, M. Jordan, M. Kearns, and S. Solla, Eds. Cambridge,MA: MIT Press, 1998, pp. 640–646.

[29] B. Scholkopf, A. Smola, and K.-R. M¨uller, “Nonlinear componentanalysis as a kernel eigenvalue problem,”Neural Comput., vol. 10, pp.1299–1319, 1998.

[30] B. Scholkopf, K. Sung, C. Burges, F. Girosi, P. Niyogi, T. Poggio, andV. Vapnik, “Comparing support vector machines with Gaussian kernelsto radial basis function classifiers,”IEEE Trans. Signal Processing, vol.45, pp. 2758–2765, 1997.

[31] A. Smola and B. Scholkopf, “From regularization operators to supportvector kernels,” inAdvances in Neural Information Processing Systems10, M. Jordan, M. Kearns, and S. Solla, Eds. Cambridge, MA: MITPress, 1998, pp. 343–349.

[32] , “A tutorial on support vector regression,” NeuroCOLT, Rep. 19,1998. Available http://svm.first.gmd.de

[33] K. Tsuda, “Support vector classifier with asymmetric kernel function,”in Proc. ESANN, M. Verleysen, Ed. Brussels: D Facto, 1999, pp.183–188.

[34] V. Vapnik, The Nature of Statistical Learning Theory. New York:Springer-Verlag, 1995.

[35] , Statistical Learning Theory. New York: Wiley, 1998.[36] G. Wahba,Spline Models for Observational Data, vol. 59 ofCBMS-NSF

Regional Conference Series in Applied Mathematics. Philadelphia, PA:SIAM, 1990.

[37] R. Williamson, A. Smola, and B. Scholkopf, “Entropy numbers, op-erators, and support vector kernels,” inAdvances in Kernel Meth-ods—Support Vector Learning, B. Scholkopf, C. Burges, and A. Smola,Eds. Cambridge, MA: MIT Press, 1999, pp. 127–144.

[38] , “Generalization performance of regularization networks andsupport vector machines via entropy numbers of compact operators,”NeuroCOLT, Tech. Rep. 19, 1998. Available http://www.neurocolt.com

Bernhard Scholkopf received the M.Sc. degree inmathematics from the University of London andthe Diplom degree in physics from the Eberhard-Karls-Universitat, Tubingen, Germany. He receivedthe Ph.D. degree in computer science from theTechnical University of Berlin, Germany, in 1997.

He has worked at the Max-Planck-Institute for bi-ological cypernetics, AT&T Bell Labs, GMDFIRST,and at the Australian National University. He hastaught at Humboldt University and the Techniversityof Berlin. His scientific interests include machine

learning and perception.Dr. Scholkopf’s dissertation on support vector learning won the prize of

the German Association for Computer Science (GI) 1997. He also won theLionel Cooper Memorial Prize from the University of London in 1992.

Sebastian Mika received the Diplom degree incomputer science from the Technical University ofBerlin in 1998. He is currently a doctoral student atGMD FIRST, Berlin.

His scientific interests are in the fields of machinelearning and kernel methods.

Chris J. C. Burges is a Distinguished Member ofTechnical Staff at Bell Laboratories, Lucent Tech-nologies. Educated as a physicist, he joined AT&Tin 1986 and worked on network performance androuting alrogithms. He moved to applied neuralnetwork research in 1990 and has worked on hand-writing and machine print recognition and speakeridentification. For the last several years he has con-centrated on the theory and application of SVM’s.

Philipp Knirsch is a computer science student atthe University of Tubingen, Germany.

He has worked for various companies in theIT area during the last 10 years, including butnot limited to AT&T Bell Labs, Hewlett-Packard,21TORR medienDesign GmbH, Festo AG, and theMax-Planck-Institute. He has worked in severalareas of computer science, such as computer graph-ics, support vector learning machines, network pro-gramming, system administration, code optimiza-tion, databases, compression technology, wavelet

analysis, and others. His current work and interest are in the fields of high-performance clustered network servers, real-time presentation of visual stimulion low-end Linux workstations, and optimizations and improvements ofreduced set vector methods for support vector machines.

Klaus-Robert Muller received the Diplom degreein mathematical physics in 1989 and the Ph.D.degree in theoretical computer science in 1992, bothfrom University of Karlsruhe, Germany.

From 1992 to 1994, he was a Postdoctoral Fellowat the Research Institute for Computer Architectureand Software Technology of the German NationalResearch Center for Information Technology (GMDFIRST) in Berlin. From 1994 to 1995, he wasa European Community STP Research Fellow atUniversity of Tokyo in Prof. Amari’s Lab. Since

1995, he has a tenure position and is Department Head of the intelligentdata analysis (IDA) group at GMD FIRST in Berlin. The IDA group hastwice won the price for the best scientific and technical project within GMD(1996 and 1998). He is currently lecturing at Humboldt University and theTechnical University of Berlin, Germany. He has worked on statistical physicsand statistical learning theory of neural networks and time-series analysis. Hispresent interests are expanded to support vector learning machines, boosting,nonstationary blind separation techniques, and recently also medical dataanalysis (MEG, EEG).

Page 18: Input space versus feature space in kernel-based methods ...pdfs.semanticscholar.org/b6a3/e0028d99439ce2741d0e147b6e9a34bc4267.pdfthis means that any reproducing kernel corresponds

SCHOLKOPF et al.: INPUT SPACE VERSUS FEATURE SPACE 1017

Gunnar Ratsch received the Diplom degree incomputer science from the University of Potsdam,Germany, in 1998. He is a doctoral student at GMDFIRST, Berlin.

His scientific interests are in the fields of Boostingand Kernel methods.

Mr. Ratsch received the prize for the best stu-dent from the faculty of Natural Sciences at theUniversity of Potsdam.

Alexander J. Smolareceived the Diplom degree inphysics from the Technical University of Munich,Germany, in 1996. Two years later, he receivedthe Ph.D. degree in computer science from theTechnical University of Berlin, Germany, with athesis on Learning with Kernels.

He is a Postdoctoral Fellow at GMD FIRST,Berlin. During his studies, which were supportedby the Stiftung Maximilaneum, he spent a yearat AT&T Bell Labs and at the Collegio Ghislieriin Pavia (Italy). In addition, he has spent two

spells at the Australian National University. His scientific goal is to makegeneralization error bounds of statistical learning theory applicable in practice.