TY - GEN
T1 - Video-Audio Multimodal Fall Detection Method
AU - Jamali, Mahtab
AU - Davidsson, Paul
AU - Khoshkangini, Reza
AU - Mihailescu, Radu-Casian
AU - Sexton, Elin
AU - Johannesson, Viktor
AU - Tillström, Jonas
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - Falls frequently present substantial safety hazards to those who are alone, particularly the elderly. Deploying a rapid and proficient method for detecting falls is a highly effective approach to tackle this concealed peril. The majority of existing fall detection methods rely on either visual data or wearable devices, both of which have drawbacks. This research presents a multimodal approach that integrates video and audio modalities to address the issue of fall detection systems and enhances the accuracy of fall detection in challenging environmental conditions. This multimodal approach, which leverages the benefits of attention mechanism in both video and audio streams, utilizes features from both modalities through feature-level fusion to detect falls in unfavorable conditions where visual systems alone are unable to do so. We assessed the performance of our multimodal fall detection model using Le2i and UP-Fall datasets. Additionally, we compared our findings with other fall detection methods. The outstanding results of our multimodal model indicate its superior performance compared to single fall detection models.
AB - Falls frequently present substantial safety hazards to those who are alone, particularly the elderly. Deploying a rapid and proficient method for detecting falls is a highly effective approach to tackle this concealed peril. The majority of existing fall detection methods rely on either visual data or wearable devices, both of which have drawbacks. This research presents a multimodal approach that integrates video and audio modalities to address the issue of fall detection systems and enhances the accuracy of fall detection in challenging environmental conditions. This multimodal approach, which leverages the benefits of attention mechanism in both video and audio streams, utilizes features from both modalities through feature-level fusion to detect falls in unfavorable conditions where visual systems alone are unable to do so. We assessed the performance of our multimodal fall detection model using Le2i and UP-Fall datasets. Additionally, we compared our findings with other fall detection methods. The outstanding results of our multimodal model indicate its superior performance compared to single fall detection models.
KW - Audio classification
KW - Fall detection
KW - Multimodal
KW - Video classification
UR - http://www.scopus.com/inward/record.url?scp=85210317498&partnerID=8YFLogxK
U2 - 10.1007/978-981-96-0125-7_6
DO - 10.1007/978-981-96-0125-7_6
M3 - Conference contribution
AN - SCOPUS:85210317498
SN - 978-981-96-0124-0
T3 - Lecture Notes in Computer Science
SP - 62
EP - 75
BT - PRICAI 2024: Trends in Artificial Intelligence
A2 - Hadfi, Rafik
A2 - Anthony, Patricia
A2 - Sharma, Alok
A2 - Ito, Takayuki
A2 - Bai, Quan
PB - Springer
T2 - 21st Pacific Rim International Conference on Artificial Intelligence, PRICAI 2024
Y2 - 18 November 2024 through 24 November 2024
ER -