Review 6

Today we talked about linear regression. In linear regression, the machine learning system predicts a numerical value in R given numerical inputs(features) in real valued space R. In other words, we determine a function or mapping F. In classification, we find y=f(X) that maps X to one of the k classes [c1 c2 ... ck].

K-Nearest Neighbor. Let us consider R^2 as shown below. The points shown in black are training data set(in other words we know the class X0 fir the given points.) First consider a simple case of k=1 where we will determine class of the new circles. In this case, we determine class of the new point as the class of points in the circle, which in our case is 1. 

Determine computational complexity of K-Nearest Neighbor for:

n training points

m dimensions

Euclidean dissimilarity measure

How many mathematical operations will be needed. And we discussed how many mathematical operations will be needed to calculate the distance. 

Then, professor introduced the Iris data set. It has been downloaded for over 1 million times. He showed the KNN- Confusion Matrices. And it has many precision, recall and F1 scores. 

Gradient Descent

A convex function is a function which means that it has a single minimum point. And if a function has multiple local minima, it is not a convex function.

Slope 

positive and negative slope 

and we can use calculus to 

show the update formula the goal is to find out the minimum of the function.  This is achieved by choosing a random point x on the horizon.

©著作权归作者所有,转载或内容合作请联系作者
平台声明:文章内容(如有图片或视频亦包括在内)由作者上传并发布,文章内容仅代表作者本人观点,简书系信息发布平台,仅提供信息存储服务。

推荐阅读更多精彩内容