Toward diverse tiny-model selection for microcontrollers
Publication Type
Journal Article
Publication Date
4-2025
Abstract
Enabling efficient and accurate deep neural network (DNN) inference on microcontrollers is challenging due to their constrained on-chip resources. Existing approaches mainly focus on compressing larger models, often compromising model accuracy as a trade-off. In this paper, we rethink the problem from the inverse perspective by directly constructing small/weak models, then enhancing their accuracy. Thus, we propose DiTMoS, a novel DNN training and inference framework featuring a selector-classifiers architecture, where the selector routes each input sample to the appropriate classifier for classification. DiTMoS is built on a key insight: a combination of weak models can exhibit high diversity and the union of them can significantly raise the upper bound of overall accuracy. To approach the upper bound, DiTMoS introduces three strategies including diverse training data splitting to enhance the classifiers’ diversity, adversarial selector-classifiers training to ensure synergistic interactions thereby maximizing their complementarity, and heterogeneous feature aggregation to improve the capacity of classifiers. We further design a network slicing technique to eliminate the extra memory consumption incurred by feature aggregation. We deploy DiTMoS on the Nucleo STM32F767ZI board and evaluate its performance across three time-series datasets for human activity recognition, keyword spotting, and emotion recognition tasks. The experimental results show that: (a) DiTMoS improves accuracy by up to 13.4% compared to the best baseline; (b) network slicing successfully eliminates the memory overhead introduced by feature aggregation, with only a minimal increase in latency.
Keywords
Accuracy, Training, Computational Modeling, Memory Management, Feature Extraction, Data Models, Sensors, Mobile Computing, Internet Of Things, Diversity Reception, Embedded Machine Learning, Model Diversity, Model Selection, Adversarial Training, Neural Network, Training Data, Deep Network, Deep Neural Network, Large Model, Input Samples, Emotion Recognition, Heterogeneous Characteristics, Classifier For Classification, Feature Aggregation, Memory Consumption, Data Split, Adversarial Training, Time Series Dataset, Efficient Inference, Human Activity Recognition, Network Slicing, Model Selection, Convolutional Neural Network, Classification Accuracy, Memory Usage, Microcontroller Unit, Convolutional Layers, Heterogeneous Architecture, Diverse Classification, Model Size, Concept Drift, Multiple Classes, Neural Architecture Search
Discipline
Artificial Intelligence and Robotics | Electrical and Computer Engineering
Research Areas
Intelligent Systems and Optimization
Publication
IEEE Transactions on Mobile Computing
Volume
24
Issue
9
First Page
1
Last Page
16
ISSN
1536-1233
Identifier
10.1109/TMC.2025.3561778
Publisher
Institute of Electrical and Electronics Engineers
Citation
MA, Xiao; HE, Shengfeng; QIAO, Hezhe; and MA, Dong.
Toward diverse tiny-model selection for microcontrollers. (2025). IEEE Transactions on Mobile Computing. 24, (9), 1-16.
Available at: https://ink.library.smu.edu.sg/sis_research/10522
Additional URL
https://doi.org/10.1109/TMC.2025.3561778