Abstract
Wearable devices are becoming affordable in our society to provide services from simple fitness tracking to the detection of heartbeat disorders. In the case of elderly populations, these devices have great potential to enable proactive eldercare, which can increase the number of years of independent living. The wearables can capture healthcare data continuously. For meaningful insight, deep learning models are preferable to process this data for robust outcomes. One of the major challenges includes deploying these models on edge devices, such as smartphones or wearables. The bottleneck is a large number of parameters and compute-intensive operations. In this research, we propose a novel knowledge distillation scheme by introducing a self-revision concept. This scheme effectively reduces model size and transfers knowledge from a deep model to a distilled model by filling learning gaps during the training. To evaluate our distilled model, a publicly available dataset, ‘growing old together validation (GOTOV)’ is utilized, which is based on medical-grade standard wearables to monitor behavioral changes in the elderly. Our proposed model reduces the 0.7 million parameters to 1500, which enables edge intelligence. It achieves a 6% improvement in precision, a 9% increase in recall, and a 9% higher F1-score compared to the shallow model for recognizing elderly behavior.
Original language | English |
---|---|
Journal | IEEE Transactions on Artificial Intelligence |
Early online date | 09 Jan 2025 |
DOIs | |
Publication status | Early online date - 09 Jan 2025 |
Publications and Copyright Policy
This work is licensed under Queen’s Research Publications and Copyright Policy.Keywords
- Edge intelligence
- distilled model
- wearables
- eldercare