|
Random Forest (RF)
|
|
RF is primarily limited by the fact that the algorithm becomes too slow and inefficient for real-time predictions if there are a large number of trees. |
|
K-nearest neighbors algorithm (KNN)
|
During the process of making real-time predictions, it does not have a training period and only makes use of the training dataset while making predictions in the future.
KNN is easy to implement since only two parameters are required, namely K and the distance function.
|
|
|
Support Vector Machine (SVM)
|
|
|
|
Gradient boost decision tree (GBDT)
|
There is a great deal of flexibility—it can be optimized on various loss functions and has several options for tuning hyperparameters to make the function fit as flexibly as possible.
It is able to predict faults with a high degree of accuracy.
|
|
|
Convolutional Neural Network (CNN)
|
It requires little pre-processing, reducing the human effort required to develop its functionalities.
A good performance was achieved when extracting local features from images.
|
|
|
Isolation Forest (IForest)
|
A small sample size is more effective.
A low memory requirement and minimal computational effort.
Scalable to handle extremely large data sets and multidimensional problems with a large number of irrelevant attributes.
|
|
|
Autoencoder (AE)
|
|
Rather than capturing as much relevant information as possible, learns to capture as much information as possible. |
|
Multilayer perceptron (MLP)
|
It is applicable to complex nonlinear problems.
Ability to handle large amounts of data.
Predict after training on time.
|
Model performance depends on the quality of the training data. |