Reproducing Kernel Hilbert Spaces

In summary, the author is reading about Reproducing Kernel Hilbert Spaces and their application in machine learning. He doesn't have a formal background in topology, but he is making headway. He is unsure of how the "kernel trick" works, but he is hopeful that someone can help him out.
  • #1
joshthekid
46
1
TL;DR Summary
Clarifying the concept of RKHS
I have been reading a lot about Reproducing Kernel Hilbert Spaces mainly because of their application in machine learning. I do not have a formal background in topology, took linear algebra as an undergrad but mainly have encountered things such as, inner product, norm, vector space, orthogonality, and independence with respect to these ideas associated in Physics. I think I am making headway but wanted to clarify before I go further.

-While in physics we generally think of an inner-product as a dot product giving the angle between two vectors (with the exception of quantum mechanics were inner products play a more abstract role in the sense of operators) it is really a more general mathematical concept endowed over a vector space,## \mathbf V ##, such that ##\mathbf V \mathsf x \mathbf V \rightarrow \mathbb R , \mathbb C ## with the properties of conjugate symmetry, linearity, and positive definite.

Now let's say I have a topological space in ## \mathbb R^\infty## additionally I have a vector space ##\mathbf V \subset \mathbb R^\infty## endowed with an inner product which is defined over ##\mathbf V## but not all of ##\mathbb R^\infty##.

Now here is where I have a little bit of uncertainty. Since the inner product is not defined over the whole topological space I cannot use the basis over the topological space to decompose ##\mathbf V ## into an independent combination of basis vector in ##\mathbb R^\infty##. Here is where I think the "magic" happens with RKHS. Given the defined inner product over ##\mathbf V## I can define a unique set of spanning vectors ##\mathbf K## in which I can express and vector ## v \in \mathbf V ## as a linear combination of the vectors in ##\mathbf K## such that any ## v \in \mathbf V=\alpha\mathbf K##. Furthermore there is a orthonormal basis, ##\mathbf U## over ##\mathbf V## such that ##\mathbf K=\mathbf U\mathbf U^t## allowing me to express any vector in ##\mathbf V## as ##v=\beta\mathbf U##

I just want to correct any misinterpretations as I go forward any figuring out how this allows the "kernel trick" to work in SVMs and other ML algorithms. Any help would be great!
 
Physics news on Phys.org
  • #2
I don't quite understand you. ##\mathbb{R}^\infty## is not a good notation, since it doesn't say a lot about the kind of infinity. Then you have an inner product on ##\mathbf{V}##, which one?, but not on ##\mathbb{R}^\infty##, why?

Yes, you can construct an orthonormal basis for ##\mathbf{V}## (Gram-Schmidt), and usually write ##\mathbb{R}^\infty = \mathbf{V}\oplus \mathbf{V}^\perp## but this is orthogonality in ##\mathbb{R}^\infty## where you say you have none.

So the true magic is a Hilbert space embedded in what is usually as well a Hilbert space, but you ripped of the inner product off the outer space, which sounds weird.
 
  • #3
So the book I am reading is "A Primer on Reproducing Kernel Hilbert Spaces". So they initially talk about extrinsic vs intrinsic topology on a finite topology, ##\mathbb R^n##. they claim that ##\mathbf V \subset \mathbb R^n## endowed with an inner product. They say
" the configuration involves three aspects, the vector space ##\mathbf V## , the orientation of ##\mathbf V## in ##\mathbb R^n## and the inner product on ##\mathbf V##. Importantly the inner product is not defined on the whole of ##\mathbb R^n## otherwise ##\mathbf V=\mathbb R^n##."

So somehow you can define an inner-product on the intrinsic topology that is not defined extrinsically.
 
  • #4
I can only imagine some artificial constructions, e.g. ##\mathbb{R}^2 \subseteq \mathbb{R}^3## with the usual inner product in the plane, and as long it isn't extended to space, it will remain an inner product of only the plane. One can also extend it to space but destroy the inner product property in the third component.

I guess the author wanted to stress the fact that an embedding doesn't necessarily imply that structures are compatible. Another example would be square integrable functions in the space of all functions. The usual inner product for square integrable functions doesn't work for arbitrary functions anymore. Or we could embed square integrable function in the space of continuous functions, where both inner products are different.
 

1. What is a reproducing kernel Hilbert space (RKHS)?

A reproducing kernel Hilbert space is a type of mathematical function space that is commonly used in machine learning and statistical learning theory. It is a Hilbert space of functions that is equipped with a reproducing kernel, which is a special type of inner product that allows for efficient computation of function values and derivatives.

2. How is an RKHS different from other function spaces?

Unlike other function spaces, an RKHS has the property that any function in the space can be uniquely determined by its values at a finite set of points. This is known as the reproducing property, which makes RKHSs useful for function approximation and interpolation.

3. What are some applications of RKHSs?

RKHSs have a wide range of applications in machine learning, including support vector machines, Gaussian processes, and kernel methods for regression and classification. They are also used in signal processing, image processing, and control theory.

4. How are RKHSs related to kernel methods?

RKHSs are closely related to kernel methods, which are a type of machine learning algorithm that uses a kernel function to map data into a higher-dimensional feature space. The RKHS is the space in which the kernel function operates, and the reproducing property of the RKHS allows for efficient computation of the kernel function.

5. What are some challenges in working with RKHSs?

One challenge in working with RKHSs is choosing an appropriate kernel function, as different kernels may perform better for different types of data. Another challenge is the computational complexity of working with RKHSs, as the dimension of the feature space can be very large. Additionally, understanding the theoretical properties of RKHSs can be challenging, as they involve advanced concepts from functional analysis and measure theory.

Similar threads

  • Topology and Analysis
Replies
9
Views
2K
  • Topology and Analysis
Replies
8
Views
462
  • Differential Geometry
Replies
10
Views
2K
  • Topology and Analysis
Replies
16
Views
2K
  • Topology and Analysis
Replies
4
Views
2K
Replies
13
Views
2K
  • Topology and Analysis
Replies
2
Views
2K
  • Topology and Analysis
Replies
4
Views
3K
Replies
2
Views
389
  • Topology and Analysis
Replies
1
Views
1K
Back
Top