Is functional analysis relevant to machine learning?
来源:互联网 发布:android电商app源码 编辑:程序博客网 时间:2024/06/04 19:57
From Quora
One place where functional analysis is particularly relevant to machine learning is the study of kernel methods, a notable example of which is the kernel Support Vector Machine, where the theory of Reproducing kernel Hilbert spaces (RKHS) from functional analysis plays a big role.Every positive definite kernel (on any type of data) uniquely defines a Hilbert space , called the RKHS with reproducing kernel , which satisfies a set of properties, and in particular provides a 'feature map' from the original space to the RKHS, for which the kernel corresponds to an inner product: (in fact, we have ).
One can study the properties of such Hilbert spaces, which can be infinite-dimensional even when the input space on which the kernel is defined isn't. For example, the RKHS of a Gaussian kernel is infinite-dimensional, the RKHS of the min kernel is a a Hilbert space similar to a Sobolev space (the inner product between two functions in the space is the integral of the product of their derivatives), and the polynomial kernel gives a space of polynomials.
One interesting property, which can be easily shown using the 'reproducing property' of the RKHS and the Cauchy-Schwarz inequality, is that for a function ,
,
i.e. is Lipschitz with constant : the variations of the function depend on the variations in the input space with respect to the geometry defined by the map . Basically the RKHS norm directly relates to the smoothness of the function (the smaller the norm, the smaller the variations).
These RKHS define function spaces, and it turns out one can optimize certain problems on these spaces (e.g. find the function in the space which gives the smallest error in an empirical risk minimization problem), using a key result called the representer theorem. The theorem states that if the objective only depends on the evaluations of the function on a set of points , and increases with the RKHS norm of the function, the optimal function will be in the linear span of the mapped functions in the RKHS, , thus reducing the problem to an optimization on , which is much easier. The condition on the RKHS norm is easy to verify if you add this norm as a 'regularizer' in your objective, and this will have the benefit of controlling the smoothness (the 'complexity') or your function.
As an example, say you want to learn a regression function in some RKHS from a set of training points , then you can assume the function is of the form , and you're left with an optimization problem on the vector , whose dimensionality is equal to the size of your training set, even though might be of infinite dimensionality! This is typically what happens in kernel SVMs, and it is part of the reason why they've been originally so successful.
0 0
- Is functional analysis relevant to machine learning?
- Applying Machine Learning to Sentiment Analysis
- what is machine learning?
- Machine Learning is Fun!
- What is Machine Learning
- Introduction to Machine Learning
- Introduction to Machine Learning
- Introduction to Machine Learning
- Introduction to Machine Learning
- Introduction to machine learning
- Introduction to Machine Learning
- Coursera Machine Learning 8.2: Principal Component Analysis
- Machine Learning-PCA(Principal Component Analysis)
- 《Machine Learning is Fun Part 6: How to do Speech Recognition with Deep Learning》译文
- Is machine learning Kaggle competitions?
- how-to-learn-machine-learning
- Introduction to Machine Learning (一)
- Introduction to Machine Learning【转】
- TortoiseSVN中Branching和Merging实践
- 腾达N300路由器无线网络,XP电脑连接速度很慢
- 54张手绘PPT看懂Google是如何运作的
- 后台服务播放器
- 缅怀惨痛的处女面
- Is functional analysis relevant to machine learning?
- Java之回溯法实现01背包
- JMX教程整合
- instancetype 和 id 的区别
- 拓扑排序
- 浅拷贝和深拷贝
- gcc 4.8.2编译时找不到头文件curses.h问题的解决
- 编译问题 无法找到zlib.h
- 输出1到最大的N位数 【微软面试100题 第六十五题】