Skip to article frontmatterSkip to article content

Machine Learning Basics

Cornell University

From Lecture: ML Setup

Basic Definitions

Machine Learning

Train/Validate/Test Split

Splitting the Data

No Free Lunch and Making Assumptions

There is no simple ML algorithm that works for all settings. Each model only works in some specific settings/background. You must make assumptions in order to learn. Each model only works under specific assumptions. Assumptions and settings are actually the same thing here.

“Assumptions” just mean our belief of how our data is distributed. Most of the time when we choose a model, we make assumptions about the distribution our data was drawn from. For example, when we try to determine whether an email is spam or not, we count number of appearance of each word in the email. So we are assuming our features xx are drawn from a multimodal distribution. When we use logistic regression, we assume conditional over x has a form like P(yxi)=11+ey(wTxi+b)P(y|\mathbf{x}_i)=\frac{1}{1+e^{-y(\mathbf{w}^T \mathbf{x}_i+b)}}. These are both very bold assumptions and are very likely to fail. When our assumptions fail, our model will do a terrible job in prediction. Therefore, when our algorithm does not perform well, one big reason is that our assumption does not hold.