Parameter
|
Value
|
XGBoost |
Base Learner |
Gradient boosted tree |
Tree construction algorithm |
Exact greedy |
Learning rate
|
0.0991 |
Lagrange multiplier
|
0 |
Number of gradient boosted trees |
80 |
Maximum depth of a tree |
6 |
Minimum sum of instance weight |
1 |
Subsample ratio of the training instances |
1 |
Sampling method |
Uniform |
L2 regularization term on weights |
1 |
Tree growing policy |
Depthwise |
Evaluation metrics for validation data |
Negative log likelihood |
SVM |
Kernel |
Linear |
Degree of the polynomial kernel |
3 |
Kernel coefficient
|
Scale |
Maximum iterations |
No constraint |
Shrinking heuristic |
True |
Probability estimates |
False |
Tolerance for stopping criterion |
1 × 10−3
|
Random forest |
Number of trees in the forest |
10 |
Quality of split measure function |
Entropy |
Minimum number of samples to split |
2 |
Minimum number of samples at a leaf node |
1 |
Use bootstrap samples for building trees |
True |
Number of jobs to run in parallel |
1 |
CatBoost |
Number of boosting rounds |
20 |
Learning rate |
0.44 |
Maximum depth of a tree |
5 |
Maximum number of trees |
1000 |
Random seed |
0 |
Sample weight frequency |
Per tree level |
Tree growing policy |
Symmetric tree |
Maximum number of leaves |
31 |
LightGBM |
Number of decision trees |
20 |
Bagging fraction |
1 |
Number of threads in the physical core |
8 |
Maximum depth of a tree |
6 |
Number of boosting iterations |
100 |
Learning rate |
0.1 |
Maximum number of leaves on one tree |
Serial |
Bagging random seed |
3 |
Dropout rate |
0.1 |