Author
Listed:
- Jungpil Shin
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
- Abu Saleh Musa Miah
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
- Rei Egawa
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
- Najmul Hassan
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
- Koki Hirooka
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
- Yoichi Tomioka
(School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan)
Abstract
Human fall detection is a significant healthcare concern, particularly among the elderly, due to its links to muscle weakness, cardiovascular issues, and locomotive syndrome. Accurate fall detection is crucial for timely intervention and injury prevention, which has led many researchers to work on developing effective detection systems. However, existing unimodal systems that rely solely on skeleton or sensor data face challenges such as poor robustness, computational inefficiency, and sensitivity to environmental conditions. While some multimodal approaches have been proposed, they often struggle to capture long-range dependencies effectively. In order to address these challenges, we propose a multimodal fall detection framework that integrates skeleton and sensor data. The system uses a Graph-based Spatial-Temporal Convolutional and Attention Neural Network (GSTCAN) to capture spatial and temporal relationships from skeleton and motion data information in stream-1, while a Bi-LSTM with Channel Attention (CA) processes sensor data in stream-2, extracting both spatial and temporal features. The GSTCAN model uses AlphaPose for skeleton extraction, calculates motion between consecutive frames, and applies a graph convolutional network (GCN) with a CA mechanism to focus on relevant features while suppressing noise. In parallel, the Bi-LSTM with CA processes inertial signals, with Bi-LSTM capturing long-range temporal dependencies and CA refining feature representations. The features from both branches are fused and passed through a fully connected layer for classification, providing a comprehensive understanding of human motion. The proposed system was evaluated on the Fall Up and UR Fall datasets, achieving a classification accuracy of 99.09% and 99.32%, respectively, surpassing existing methods. This robust and efficient system demonstrates strong potential for accurate fall detection and continuous healthcare monitoring.
Suggested Citation
Jungpil Shin & Abu Saleh Musa Miah & Rei Egawa & Najmul Hassan & Koki Hirooka & Yoichi Tomioka, 2025.
"Multimodal Fall Detection Using Spatial–Temporal Attention and Bi-LSTM-Based Feature Fusion,"
Future Internet, MDPI, vol. 17(4), pages 1-22, April.
Handle:
RePEc:gam:jftint:v:17:y:2025:i:4:p:173-:d:1635077
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:17:y:2025:i:4:p:173-:d:1635077. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.