dc.contributor.author | Shahabian Alashti, Mohamad Reza | |
dc.contributor.author | Bamorovat Abadi, Mohammad | |
dc.contributor.author | Holthaus, Patrick | |
dc.contributor.author | Menon, Catherine | |
dc.contributor.author | Amirabdollahian, Farshid | |
dc.date.accessioned | 2024-11-15T16:45:00Z | |
dc.date.available | 2024-11-15T16:45:00Z | |
dc.date.issued | 2024-10-23 | |
dc.identifier.citation | Shahabian Alashti , M R , Bamorovat Abadi , M , Holthaus , P , Menon , C & Amirabdollahian , F 2024 , Efficient Skeleton-based Human Activity Recognition in Ambient Assisted Living Scenarios with Multi-view CNN . in 2024 10th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics, BioRob 2024 . Proceedings of the IEEE RAS and EMBS International Conference on Biomedical Robotics and Biomechatronics , Institute of Electrical and Electronics Engineers (IEEE) , Heidelberg, Germany , pp. 979-984 , 2024 IEEE RAS EMBS 10th International Conference on Biomedical Robotics and Biomechatronics (BioRob) , Heidelberg , Germany , 1/09/24 . https://doi.org/10.1109/BioRob60516.2024.10719939 | |
dc.identifier.citation | conference | |
dc.identifier.isbn | 979-8-3503-8652-3 | |
dc.identifier.issn | 2155-1774 | |
dc.identifier.other | ORCID: /0000-0001-8450-9362/work/171844863 | |
dc.identifier.other | ORCID: /0000-0003-2072-5845/work/171845029 | |
dc.identifier.other | ORCID: /0000-0001-5037-9918/work/174228534 | |
dc.identifier.uri | http://hdl.handle.net/2299/28461 | |
dc.description | © 2024 IEEE. This is the accepted manuscript version of an article which has been published in final form at https://doi.org/10.1109/BioRob60516.2024.10719939 | |
dc.description.abstract | Human activity recognition (HAR) plays a critical role in diverse applications and domains, from assessments of ambient assistive living (AAL) settings and the development of smart environments to human-robot interaction (HRI) scenarios. However, using mobile robot cameras in such contexts has limitations like restricted field of view and possible noise. Therefore, employing additional fixed cameras can enhance the field of view and reduce susceptibility to noise. Never-theless, integrating additional camera perspectives increases complexity, a concern exacerbated by the number of real-time processes that robots should perform in the AAL scenario. This paper introduces our methodology that facilitates the combination of multiple views and compares different aspects of fusing information at low, medium and high levels. Their comparison is guided by parameters such as the number of training parameters, floating-point operations per second (FLOPs), training time, and accuracy. Our findings uncover a paradigm shift, challenging conventional beliefs by demonstrating that simplistic CNN models outperform their more complex counterparts using this innovation. Additionally, the pivotal role of pipeline and data combination emerges as a crucial factor in achieving better accuracy levels. In this study, integrating the additional view with the Robot-view resulted in an accuracy increase of up to 25 %. Ultimately, we have successfully attained a streamlined and efficient multi-view HAR pipeline, which will now be incorporated into AAL interaction scenarios. | en |
dc.format.extent | 6 | |
dc.format.extent | 538268 | |
dc.language.iso | eng | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | |
dc.relation.ispartof | 2024 10th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics, BioRob 2024 | |
dc.relation.ispartofseries | Proceedings of the IEEE RAS and EMBS International Conference on Biomedical Robotics and Biomechatronics | |
dc.title | Efficient Skeleton-based Human Activity Recognition in Ambient Assisted Living Scenarios with Multi-view CNN | en |
dc.contributor.institution | School of Physics, Engineering & Computer Science | |
dc.contributor.institution | Centre for AI and Robotics Research | |
dc.contributor.institution | Department of Computer Science | |
dc.contributor.institution | Centre for Future Societies Research | |
dc.contributor.institution | Robotics Research Group | |
dc.identifier.url | http://www.scopus.com/inward/record.url?scp=85208615120&partnerID=8YFLogxK | |
rioxxterms.versionofrecord | 10.1109/BioRob60516.2024.10719939 | |
rioxxterms.type | Other | |
herts.preservation.rarelyaccessed | true | |