Skip to main content
. 2025 May 2;15:15411. doi: 10.1038/s41598-025-96588-1

Table 10.

Comparison of TinyML models trained on Google Speech Commands Dataset.

Model Accuracy (%) Latency (ms) Power consumption Suitability for TinyML
Transformer-based Model43 92.0 80–150 High Not ideal due to computational cost
Hybrid CNN-RNN39 89.5 100–250 Moderate Suitable for noise-resilient applications
DNN (Deep Neural Network)36 99.0 50–100 High Requires optimization
CNN38 94.0 30–70 Moderate Highly suitable for real-time
RNN37 95.0 100–200 High Effective for sequential learning tasks
Decision Tree41 90.0 10–30 Low Best for ultra-low-power applications
SVM42 91.5 90–180 Moderate Balanced approach for edge deployment