2022
DOI: 10.1145/3542819
|View full text |Cite
|
Sign up to set email alerts
|

Human Activity Recognition on Microcontrollers with Quantized and Adaptive Deep Neural Networks

Abstract: Human Activity Recognition (HAR) based on inertial data is an increasingly diffused task on embedded devices, from smartphones to ultra low-power sensors. Due to the high computational complexity of deep learning models, most embedded HAR systems are based on simple and not-so-accurate classic machine learning algorithms. This work bridges the gap between on-device HAR and deep learning, proposing a set of efficient one-dimensional Convolutional Neural Networks (CNNs) that can be deployed on general purpose mi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 46 publications
0
7
0
Order By: Relevance
“…Dynamic (or adaptive) inference techniques, including this work, are designed to overcome these limitations. They allow the deployment of a single model able to adapt its complexity at runtime, while keeping the memory overhead under control [24]- [26]. In practice, a dynamic model can be partially turned off when the external conditions require it, or when the processing input's difficulty allows it [22].…”
Section: Static and Dynamic ML Optimizationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Dynamic (or adaptive) inference techniques, including this work, are designed to overcome these limitations. They allow the deployment of a single model able to adapt its complexity at runtime, while keeping the memory overhead under control [24]- [26]. In practice, a dynamic model can be partially turned off when the external conditions require it, or when the processing input's difficulty allows it [22].…”
Section: Static and Dynamic ML Optimizationsmentioning
confidence: 99%
“…Literature works differ mainly in how they decompose the model. For instance, the authors of [24]- [26], [28] obtain a single sub-model by selectively deactivating a subset of the layers or channels of a network, or truncating the bit-width used to represent parameters. Other works extend the approach more than two sub-models [25], [29] or enhance the stopping criterion with class-aware thresholds [22].…”
Section: A Dynamic Inferencementioning
confidence: 99%
“… The encoding strategy used for the devised CNN architecture has reduced the search space of the architecture, which prevents the structure from expressing the diversity of the assigned tasks. [15] Human Activity Recognition A simple grid search algorithm has been explored how to optimize the models' hyper-parameters. To achieve a good trade-off between classification score and memory occupancy, a combination of sub-byte and mixed precision quantization was used.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, practitioners of deep learning encounter difficulty when it comes to manually building deep models and determining suitable configurations (e.g., model layers and operation types) through trial and error. Various steps are involved in feeding domain knowledge into DL, including Feature Engineering (FE) [13] , model generation [14] , and model deployment [15] , [16] . Because CNNs are based on layers, they allow the flexibility of adding or removing layers based on the training phase, which is then used in inference (classification) to classify the data.…”
Section: Introductionmentioning
confidence: 99%
“…Daghero et al applied Binary Neural Networks (BNNs) to HAR to decrease network complexity via an extreme form of quantization [6]; indeed, by using BNNs the precision of data format, both weights and layers input/output, is reduced to 1-bit precision. Authors propose a BNN inference library that targets RISC-V processors.Subsequently, authors extended their work [4,5] by proposing a set of efficient one-dimensional convolutional neural networks (1D CNNs) and testing optimization techniques such as sub-type and mixed-precision quantization. The aim was to find a good trade-off between accuracy and memory occupation.…”
Section: Related Workmentioning
confidence: 99%