机器学习笔记2-Supervised learning
来源:互联网 发布:eclipse打包java程序 编辑:程序博客网 时间:2024/05/16 15:28
1.1 Probabilistic interpretation
why might the least-squares cost function J, be a reasonable choice?
Let us assume that the target variables and the inputs are related via the equation :
Let us further assume that the
likelihood function:
maximize the log likelihood
Hence, maximizing
summarize: Under the previous probabilistic assumptions on the data, least-squares regression corresponds to finding the maximum likelihood estimate of θ.
1.2 Locally weighted linear regression
The leftmost figure below shows the result of fitting a
Instead, if we had added an extra feature
it might seem that the more features we add, the better. However, there is also a danger in adding too many features: The rightmost figure is the result of fitting a 5-th order polynomial
As discussed previously, and as shown in the example above, the choice of features is important to ensuring good performance of a learning algorithm. (When we talk about model selection, we’ll also see algorithms for automatically choosing a good set of features.) In this section, let us talk briefly talk about the locally weighted linear regression (LWR) algorithm which, assuming there is sufficient training data, makes the choice of features less critical.
In the original linear regression algorithm, to make a prediction at a query point x .
1. Fit θ to minimize
2. Output
the locally weighted linear regression algorithm
1. Fit θ to minimize
2. Output
A fairly standard choice for the weights is:
if |x(i) − x| is small, then w(i) is close to 1; and if |x(i) − x| is large, then w(i) is small. Hence, θ is chosen giving a much higher “weight” to the (errors on) training examples close to the query point x.
The parameter τ controls how quickly the weight of a training example falls off with distance of its x(i) from the query point x; τ is called the bandwidth parameter.
Locally weighted linear regression is a non-parametric algorithm.The (unweighted) linear regression algorithm that we saw earlier is known as a parametric learning algorithm, because it has a fixed, finite number of parameters (the
0 0
- 机器学习笔记2-Supervised learning
- 机器学习笔记1-Supervised learning
- 机器学习笔记1-Supervised learning
- 机器学习笔记1-Supervised learning
- 机器学习笔记1-Supervised learning
- 【机器学习】Supervised Learning
- 【机器学习笔记】一、Supervised learning监督学习
- [机器学习入门] 李宏毅机器学习笔记-13 (Semi-supervised Learning ;半监督学习)
- Coursera机器学习课程笔记(1) Supervised Learning and Unsupervised Learning
- 学习笔记2 Supervised Learning and Optimization 之 Softmax Regression
- Machine Learning机器学习 - Supervised Learning监督学习 - Native Bayes
- [机器学习] Coursera ML笔记 - 监督学习(Supervised Learning) - Representation
- [机器学习02]supervised learning and unsupervised learning
- Stanford机器学习Prolem Set#1:Supervised Learning
- Machine Learning机器学习 - Supervised Learning监督学习 - SVM(Support Vector Machine)
- Andrew Ng 《Machine Learning》第一讲——Supervised Learning & Unsupervised Learning 学习笔记
- 什么是监督学习(supervised learning)
- 监督式学习(Supervised learning)
- 深入源码理解Android Touch事件分发机制(上篇)
- c# 将xml内容转换成DataSet
- React环境配置和基本使用
- Java80商城系统第一节
- JAVA基础再回首(十九)——异常的概述、Try…Catch、多异常处理、Throws、throw、finally、自定义异常及异常的注意事项
- 机器学习笔记2-Supervised learning
- Centos6.5安装Scrapy
- Dlib+opencv实时提取人脸轮廓(windows环境下//Dlib配置入门)
- zoom插件实现图片放大缩小功能
- 饿了么分布式服务治理及优化经验
- Java concurrent包
- 【华为OJ】字符串匹配
- MyDiskTest硬盘检测绿色版
- 模拟实现C++/boost库智能指针auto_ptr scoped_ptr和share_ptr