Supervised learning
|
|
Regression: Model finds outputs that are real variables |
Classification: The model divides inputs into classes or groups |
Algorithm
|
Task
|
Description
|
Naïve Bayes |
Classification |
A “probabilistic classifier” that determines the probability of the features occurring in each class by treating every feature independently to return the most likely class based on the Bayes rule.
Particularly suited when the dimensionality of the inputs is high.
|
Support vector machines |
Classification |
|
Random Forest |
Classification/Regression |
An ensemble of simple tree predictors that vote for the most popular class for classification problems. In the regression problems, the tree responses are averaged to obtain an estimate of the dependent variable.
Overfitting is less likely to occur as more decision trees are added to the forest.
|
K‐nearest‐neighbors |
Classification/Regression |
A nonparametric algorithm based on feature similarity by assuming that similar things exist in close proximity.
Useful for a classification study when there is little or no prior knowledge about the distribution data.
|
Artificial neural networks |
Classification/Regression |
|
Deep neural network |
Classification/Regression |
A collection of neurons organized in a sequence of multiple layers.
Type of artificial neural network with several advantages (i.e., shared weights [parameter sharing), spatial relations, and local receptive fields
Learning can be supervised, unsupervised, or semisupervised.
End‐to‐end learning and transfer learning are the major approaches performed by the deep neural network.
Autoencoders and generative adversarial networks are the two specific forms of deep neural networks.
|
Multiple regression |
Regression |
|
Unsupervised learning
|
|
Algorithm
|
Task
|
Description
|
K‐means clustering |
Clustering |
|
Fuzzy clustering |
Clustering |
|
Hierarchical clustering |
Clustering |
|
Principal component analysis |
Dimensionality reduction |
|
Independent component analysis |
Dimensionality reduction |
|
Autoencoders |
Dimensionality reduction |
|
Deep belief nets |
Dimensionality reduction |
|
Generative adversarial networks |
Anomaly detection |
|
Self‐organizing map |
Dimensionality reduction |
|
Semisupervised learning
|
|
Active learning
|
A particular case of semisupervised learning, where the algorithm is allowed to query the user for the label of a subset of training instances
Used to construct a high‐performance classifier while keeping the size of the training data set to a minimum by actively selecting the valuable data points
|
Reinforcement learning
|
|
Transfer learning
|
A deep learning technique enables developers to harness a neural network used for one task and apply it to another domain.
It allows the reuse of a pretrained deep neural network on a new task with only a small amount of data.
Useful when the data is insufficient for a new domain to be handled by a neural network, and there is a big preexisting data pool that can be transferred
|
Multitask learning
|
An approach to inductive transfer that improves generalization performance of multiple related tasks by leveraging useful information among them.
Useful when there are multiple related tasks, each of which has limited training samples
|
Multiple kernel learning
|
A flexible learning method that use a predefined set of kernels and learn convex combinations of kernels over potentially different domains.
Used when there are heterogeneous sources of data for the task at hand
|
Ensemble learning
|
|
End‐to‐end learning
|
A deep learning process in which all of the parameters are trained jointly, rather than step by step. It allows the training of a deep neural network based on raw data without descriptors. Since the pipeline is replaced with a single learning algorithm, it goes directly from the input to the desired output and thereby overcome limitations of the traditional approach.
|