Uses labelled data, to derive a mapping between input examples and target variable.
\(D_\text{train} = X, y\)
Unsupervised
Learning from unlabelled data
\(D_\text{train} = X\)
Semi-Supervised
\(\exists\) labelled data and large amount of unlabelled data. Label the unlabelled data using the labelled data.
For example, love is labelled as emotion, but lovely isn’t
Cotraining, Semi-Supervised SVM
Self-Supervised
Supervised learning without explicit labels; labels created from the data
Images: identify correct rotations Sequences: out-of-sequence corrections Text: word embeddings
Lazy/ Instance-Based
Store the training examples instead of training explicit description of the target function.
Output of the learning algorithm for a new instance not only depends on it, but also on its neighbors.
The best algorithm is KNN (K-Nearest Neighbor) Algorithm.
Useful for recommender system.
Active AL
Learning system is allowed to choose the data from which it learns. There exists a human annotator.
Useful for gene expression/cancer classification
Multiple Instance
Weakly supervised learning where training instances are arranged in sets. Each set has a label, but the instances don’t
Transfer
Reuse a pre-trained model as the starting point for a model on a new related task
Reinforcement Learning RL
Learning in realtime, from experience of interacting in the environment, without any fixed input dataset. It is similar to a kid learning from experience.
Best algorithm is Q-Learning algorithm.
\(D_\text{train} = X, \text{Feedback}\)
Game playing
Bayesian Learning
Conditional-probabilistic learning tool Each observed training expmle can incrementally inc/dec the estimated probability that a hypothesis is correct.
Useful when there is chance of false positive. For eg: Covid +ve