You need to login before you can make any post submission.

Add New Chapter
## Tutorial

# Introduction to Machine Learning

###### We’ve probably used a learning algorithm dozens of times a day without knowing it.

### Machine Learning

###### Examples:

### Machine Learning Definition

###### But here are some people who have tried to define it:

##### Note:

### Machine learning algorithms

###### Others:

## Supervised Learning Algorithm

### Regression Problems

### Classification Problems

##### Notes:

## Unsupervised Learning Algorithm

### Clustering Problems

### Non-Clustering Problems

Introduction to Machine Learning

Linear Regression with One Variable

Linear Algebra Review

Linear Regression with Multiple Variables

Octave / Matlab Tutorial

Logistic Regression

Regularization

Neural Networks

Applying Machine Learning

Machine Learning System Design

Support Vector Machines

Unsupervised Learning

Dimensionality Reduction

Anomaly Detection

Recommender Systems

Large Scale Machine Learning

Example Application - Photo OCR

Linear Regression with One Variable

Linear Algebra Review

Linear Regression with Multiple Variables

Octave / Matlab Tutorial

Logistic Regression

Regularization

Neural Networks

Applying Machine Learning

Machine Learning System Design

Support Vector Machines

Unsupervised Learning

Dimensionality Reduction

Anomaly Detection

Recommender Systems

Large Scale Machine Learning

Example Application - Photo OCR

Machine learning is the science of getting computers the ability to learn, without being explicitly programmed.

- Every time we use a web search engine like Google or Bing to search the internet, one of the reasons that works so well is because a learning algorithm, one implemented by Google or Microsoft, has learned how to rank web pages.
- Every time we use Facebook or Apple’s photo typing application and it recognizes our friends’ photos, that’s also machine learning.
- Every time we read our email and the spam filter saves us from having to wade through tons of spam email, that’s also a learning algorithm.

- Grew out of work on AI
- New capability for computers

**Database Mining :**Large datasets from growth of automation/web. Ex: Web click data, Medical Records, Biology, Engineering.**Applications can’t programmed by hand :**Autonomous Helicopter, Handwriting Recoginition, NLP, Computer Vision.**Self-customizing Programs :**Amazon, Netflix product recommendations.**Understanding human learning :**Brain, Real AI

Even among machine learning practitioners there isn’t a well accepted definition of what is and what isn’t machine learning.

**Arthur Samuel(1959) :**Field of study that gives computers the ability to learn, without being explicitly programmed.

*Samuel’s claim to fame was that back in the 1950’s, he wrote a checkers playing program. And the amazing thing about this checkers playing program, was that Arthur Samuel himself, wasn’t a very good checkers player. But what he did was, he had to program for it to play 10’s of 1000’s of games against itself. And by watching what sorts of board positions tended to lead to wins, and what sort of board positions tended to lead to losses. The checkers playing program learns over time what are good board positions and what are bad board positions. And eventually learn to play checkers better than Arthur Samuel himself was able to. This was a remarkable result. Although Samuel himself turned out not to be a very good checkers player. But because the computer has the patience to play tens of thousands of games itself. No human, has the patience to play that many games. By doing this the computer was able to get so much checkers-playing experience that it eventually became a better checkers player than Arthur Samuel himself.*

**Tom Mitchell(1998) :**Well-posed Learning Problem: A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E.

Example:playing checkers.E = the experience of playing many games of checkers

T = the task of playing checkers.

P = the probability that the program will win the next game.

- In general, any machine learning problem can be assigned to one of two broad classifications:
- Supervised learning
- Unsupervised learning.

- Supervised learning
- Unsupervised learning

- Reinforcement learning
- Recommender systems.

- In supervised learning, we are given a data set and already know what our correct output should look like, having the idea that there is a relationship between the input and the output.
- Supervised learning problems are categorized into “regression” and “classification” problems.

- In a regression problem, we try to predict results within a continuous output, meaning that we try to map input variables to some continuous function.

Example:Given data about the size of houses on the real estate market, try to predict their price.

Price as a function of size is a continuous output, so this is a regression problem.

- In a classification problem, we instead try to predict results in a discrete output.
- In other words, we try to map input variables into discrete categories.

Example:Given a patient with a tumor and its sizes, we have to predict whether the tumor is malignant or benign.

- We could turn the
**housing price prediction**example into a classification problem by instead making our output about whether the house “sells for more or less than the asking price.” Here we are classifying the houses based on price into two discrete categories. - In the Breast Cancer prediction problem we have done the classification only on the basis one feature or attribute i.e tumor size but there can be more than 1 attribute responsible for it.
- A learning algorithm can deal with not only 2 or 3 or 5 features/attributers but an infinite no. of features.

Practice Problem:

Other Examples:(a) Regression - Given a picture of a person, we have to predict their age on the basis of the given picture

(b) Classification - Given a patient with a tumor, we have to predict whether the tumor is malignant or benign.

- Unsupervised learning allows us to approach problems with little or no idea what our results should look like.
- We can derive structure from data where we don’t necessarily know the effect of the variables.
- We can derive this structure by clustering the data based on relationships among the variables in the data.
- With unsupervised learning there is no feedback based on the prediction results.

- Unsupervised Learning problems can be categorised into
**Clustering**and**Non-clustering**problems.

- Clustering Example: Take a collection of 1,000,000 different genes, and find a way to automatically group these genes into groups that are somehow similar or related by different variables, such as lifespan, location, roles, and so on.

- Non-clustering Example: The “Cocktail Party Algorithm”, allows us to find structure in a chaotic environment. (i.e. identifying individual voices and music from a mesh of sounds at a cocktail party).

Example Problem: