Abstract
This article presents a human fall detection system that addresses two critical challenges: privacy preservation and detection accuracy. We propose a comprehensive framework that integrates state-of-the-art machine learning models, multimodal data fusion, federated learning (FL), and Karush-Kuhn-Tucker (KKT)-based resource optimization. The system fuses data from wearable sensors and cameras using Gramian Angular Field (GAF) encoding to capture rich spatial-temporal features. To protect sensitive data, we adopt a privacy-preserving FL setup, where model training occurs locally on client devices without transferring raw data. A custom convolutional neural network (CNN) is designed to extract robust features from the fused multimodal inputs under FL constraints. To further improve efficiency, a KKT-based optimization strategy is employed to allocate computational tasks based on device capacity. Evaluated on the UP-Fall dataset, the proposed system achieves 91% accuracy, demonstrating its effectiveness in detecting human falls while ensuring data privacy and resource efficiency. This work contributes to safer, scalable, and real-world-applicable fall detection for elderly care.
| Original language | English |
|---|---|
| Pages (from-to) | 1087-1116 |
| Number of pages | 30 |
| Journal | CMES - Computer Modeling in Engineering and Sciences |
| Volume | 145 |
| Issue number | 1 |
| DOIs | |
| State | Published - 2025 |
Bibliographical note
Publisher Copyright:Copyright © 2025 The Authors. Published by Tech Science Press.
Keywords
- Multimodal approach
- fall detection
- federated learning
- privacy-preserving
- resource constraints
ASJC Scopus subject areas
- Software
- Modeling and Simulation
- Computer Science Applications