Skip to article frontmatterSkip to article content

Beyond supervised learning

Cornell University

Data Augmentation

When we don’t have enough data to make the model robust, sometimes we do data augmentation.

You can think of most data augmentation as adding a soft regularization term to the ERM goal so the model we trained can be invariant to some group action (shift, rotation, ...)

Semi-Supervised Learning

This is when we have a lot of unlabeled data and only some labeled data. There are several ways to get around this and we make different assumptions for each:

Weak Supervision Learning

Labelling is either too expensive or requires too much expertise. So we can only get labels that are noisy or imprecise through

Reference: http://ai.stanford.edu/blog/weak-supervision/

Self-Supervised Learning

Extract a supervision label from the unlabeled data itself.

For example, we have a “fill-in-the-blank” setting for computer vision: Take a image. Remove patches from the image. Train a DNN to recover the original image from the version with the patches removed.