Recognition of Activities of Daily Living from Topic Model

Ihianle, Isibor Kennedy and Naeem, Usman and Tawil, Abdel-Rahman H. (2016) Recognition of Activities of Daily Living from Topic Model. Procedia Computer Science, 98. pp. 24-31. ISSN 1877-0509

[img]
Preview
Text
Recognition of Activities of Daily Living from Topic Model.pdf - Published Version
Available under License Creative Commons Attribution Non-commercial No Derivatives.

Download (384kB)

Abstract

Research in ubiquitous and pervasive technologies have made it possible to recognise activities of daily living through non-intrusive sensors. The data captured from these sensors are required to be classified using various machine learning or knowledge driven techniques to infer and recognise activities. The process of discovering the activities and activity-object patterns from the sensors tagged to objects as they are used is critical to recognising the activities. In this paper, we propose a topic model process of discovering activities and activity-object patterns from the interactions of low level state-change sensors. We also develop a recognition and segmentation algorithm to recognise activities and recognise activity boundaries. Experimental results we present validates our framework and shows it is comparable to existing approaches.

Item Type: Article
Identification Number: https://doi.org/10.1016/j.procs.2016.09.007
Dates:
DateEvent
21 September 2016Published
Uncontrolled Keywords: Activity discoveryActivity recognitionPattern AnalysisProbabilistic Latent Semantic Analysis
Subjects: CAH11 - computing > CAH11-01 - computing > CAH11-01-01 - computer science
Divisions: Faculty of Computing, Engineering and the Built Environment
Faculty of Computing, Engineering and the Built Environment > School of Computing and Digital Technology
Depositing User: Oana-Andreea Dumitrascu
Date Deposited: 29 Jun 2017 13:18
Last Modified: 22 Mar 2023 12:01
URI: https://www.open-access.bcu.ac.uk/id/eprint/4754

Actions (login required)

View Item View Item

Research

In this section...