Unsupervised Human Activities Recognition

Human Activities Recognition (HAR) is an important component in assistive technologies; however, we have not seen wide adoption of HAR technologies in our homes. Two main hurdles to the wide adoption of HAR technologies in our homes are the expensive infrastructure requirement and the use of supervised learning in the HAR technologies. Many HAR research works have been carried out assuming an environment embedded with sensors. In addition, the majority of HAR technologies use supervised approaches, where there are labeled data to train the expert system. In reality, our natural living environment are not embedded with sensors. Labeled data are not available in our natural living environment. We are developing a framework for autonomous HAR suitable in our natural living environment, i.e. the sensor-less homes. The framework uses unsupervised learning approach to enable a robot, acting as a mobile sensor hub, to autonomously collect data and learn the different human activities without requiring manual (human) labeling of the data.

To develop a system of autonomous human activities recognition, we have proposed a pipeline with different processes in a broader perspective of human activity analysis.

The different stages apply different machine learning approaches including supervised and unsupervised. The learning and recognition stages have been extensively studied in HAR research community and have mainly applied supervised learning techniques. On the other hand, the discovery stage is a much less studied problem. The discovery stage attempts to differentiate or group different actions or activities. This resembles the ability of a child in knowing that one action is different or similar to another, despite not knowing what the actions are, i.e. without labels or without being told. If the discovery stage can successfully group different activities into their respective groups, these groups can be fed to the subsequent stage to learn a model for each group, i.e. each activity, with pseudo-labels. Subsequently, human-robot interaction can convert the pseudo-labels to linguistic labels depending on the language of the human interacting with the robot. This resembles young children interacting with adults or older children to label the models of the activities they have formed or learned through observation.

Our initial works are focused in solving the different problems in the discovery stage, i.e. human activities discovery.

We believe this is an important ability of any intelligent system to be able to self-learn. Recently, this type of machine learning has been referred as self-supervised learning. The concepts and techniques developed in the discovery stage will be applicable in other domains such as object recognition.


Application of Multi-Objective Evolutionary Computation in Human Activities Discovery

This is the PhD work of Parham. In this work, we explore the application of multi-objective evolutionary algorithms to solve the problem of human activities discovery (HAD). In particular, we improved the different stages in HAD by integrating various AI and computation techniques in the feature extraction and clustering stages.

People

  • Parham Hadikhani
  • Daphne Teck Ching Lai (Main Supervisor)
  • Ong Wee Hong

Data/Codes

Publications

Media


Investigation of Features and Unsupervised Learning Approach for Human Activities Discovery

This is the MSc work of Amran. In this work, we investigated the effectiveness of different features and clustering algorithms in human activities discovery.

People

Data/Codes

Publications

Media


Autonomous Learning and Recognition of Human Actions based on An Incremental Approach of Clustering

In this work, the initial concept of human activities discovery was developed. We proposed a framework or the concept of an autonomous or self-learning system that begin from observations of human activities to learning the models of the (new and unknown) activities discovered to recognition of activities. The discovery and learning of activities were done without requiring labels. The proposed framework can be generalized as a self-learning system for other types of recognition, e.g. object recognition.

People

  • Ong Wee Hong
  • Leon Palafox (Collaborator)
  • Takafumi Koseki (Supervisor)

Data/Codes

Publications

Media