Table 1.
Name | Loss function (L) | Regularizer (R) |
---|---|---|
AIC/BIC | ∥y−〈ω,x〉∥2 | ∥ω∥0 |
Lasso | ∥y−〈ω,x〉∥2 | ∥ω∥1 |
Elastic Net | ∥y−〈ω,x〉∥2 | + ∥ω∥1 |
Regularized Least Absolute | ||
Deviations Regression | ∥y−〈ω,x〉∥1 | ∥ω∥1 |
Classic SVM | max(0,1−y〈ω,x〉)a | |
ℓ 1-SVM | max(0,1−y〈ω,x〉)a | |
Logistic Regression | log(1+exp(−y〈ω,x〉)) |
*This is the so called Hinge loss
The ℓ 1- and ℓ 2-norm of a vector z=(z 1,…,z d)∈ℝ d are defined by and , respectively. The “ ℓ 0-norm” ∥z∥0, simply counts the number of non-zero entries of z