Table 3:
Traditional and deep learning classification methods.
Traditional Methods | ||||||
---|---|---|---|---|---|---|
Study | Approach | Database | Performance** | Reference | ||
Features | Classifier | Top1 Acc. | Top5 Acc. | |||
Chen, 2012 | SIFT, LBP, color and gabor | Multi-class Adaboost | Chen | 68.3% | 90.9% | [22] |
Beijbom et al., 2015 | SIFT, LBP, color, HOG and MR8 | SVM | 77.4% | 96.2% | [20] | |
Anthimopoulos et al., 2014 | SIFT and color | Bag of Words and SVM | Diabetes | 78.0% | - | [31] |
Bossard et al., 2014 | SURF and L*a*b color values | RFDC | Food-101 | 50.8% | - | [18] |
Hoashi et al., 2010 | Bag of features, color, gabor texture and HOG | MKL | Food85 | 62.5% | - | [25] |
Beijbom et al., 2015 | SIFT, LBP, Color, HOG and MR8 | SVM | Menu-Match | 51.2%* | [20] | |
Christodoulidis et al., 2015 | Color and LBP | SVM | Local dataset | 82.2% | - | [34] |
Pouladzadeh et al., 2014 | Color, texture, size and shape | SVM | 92.2% | - | [12] | |
Pouladzadeh et al., 2014 | Graph Cut, color, texture, size and shape | SVM | 95.0% | - | [12] | |
Kawano and Yanai, 2013 | Color and SURF | SVM | - | 81.6% | [27] | |
Farinella et al., 2014 | Bag of textons | SVM | PFID | 31.3% | - | [24] |
Yang et al., 2010 | Pairwise local features | SVM | 78.0% | - | [28] | |
He et al., 2014 | DCD, MDSIFT, SCD, SIFT | KNN | TADA | 64.5% | - | [29] |
Zhu et al., 2015 | Color, texture and SIFT | KNN | 70.0% | - | [10] | |
Matsuda et al., 2012 | SIFT, HOG, Gabor texture and color | MKL-SVM | UEC-Food-100 | 21.0% | 45.0% | [9] |
Liu et al., 2016 | Extended HOG and Color | Fisher Vector | 59.6% | 82.9% | [36] | |
Kawano and Yanai, 2014 | Color and HOG | Fisher Vector | 65.3% | - | [39] | |
Yanai and Kawano, 2015 | Color and HOG | Fisher Vector | 65.3% | 86.7% | [35] | |
Kawano and Yanai, 2014 | Fisher Vector, HOG and color | One x rest Linear classifier | UEC-Food-256 | 50.1% | 74.4% | [38] |
Yanai et al., 2015 | Color and HOG | Fisher Vector | 52.9% | 75.5% | [35] | |
Deep Leaning Methods | ||||||
Study | Approach | Dataset | Topi | Top5 | Reference | |
Anthimopoulos et al., 2014 | ANNnh | Diabetes | 75.0% | - | [31] | |
Bossard et al., 2014 | Food-101 | Food-101 | 56.4% | - | [18] | |
Yanai and Kawano, 2015 | DCNN-Food | 70.4% | - | [35] | ||
Liu et al., 2016 | DeepFood | 77.4% | 93.7% | [36] | ||
Meyers, 2015 | GoogleLeNet | 79.0% | - | [11] | ||
Hassannejad et al., 2016 | Inception v3 | 88.3% | 96.9% | [32] | ||
Meyers, 2015 | GoogleLeNet | Food201 segmented | 76.0% | - | [11] | |
Menu-Match | 81.4%* | - | ||||
Christodoulidis et al., 2015 | Patch-wise CNN | Own database | 84.90% | - | [34] | |
Pouladzadeh et al., 2016 | Graph cut+Deep Neural Network | 99.0% | - | [40] | ||
Kawano and Yanai, 2014 | OverFeat+Fisher Vector | UEC-Food-100 | 72.3% | 92.0% | [39] | |
Liu et al., 2016 | DeepFood | 76.3% | 94.6% | [36] | ||
Yanai and Kawano, 2015 | DCNN-Food | 78.8% | 95.2% | [35] | ||
Hassannejad et al., 2016 | Inception v3 | 81.5% | 97.3% | [32] | ||
Chen and Ngo, 2016 | Arch-D | 82.1% | 97.3% | [23] | ||
Liu et al., 2016 | DeepFood | UEC-Food-256 | 54.7% | 81.5% | [36] | |
Yanai and Kawano, 2015 | DCNN-Food | 67.6% | 89.0% | [35] | ||
Hassannejad et al., 2016 | Inception v3 | 76.2% | 92.6% | [32] | ||
Ciocca et al., 2016 | VGG | UNIMINB2016 | 78.3% | - | [8] | |
Chen and Ngo, 2016 | Arch-D | VIREO | 82.1% | 95.9% | [23] |
Represents the mean average precision
Top 1 and/or Top 5 indicate that the performance of the classification model was evaluated based on the first assigned class with the highest probability and/or the top 5 classes among the prediction for each given food item, respectively