Recognizing Human-Object Interactions in Videos

dc.contributor.advisorLi, Frederick
dc.contributor.authorAlmushyti, Muna Ibrahim M
dc.date.accessioned2023-09-13T08:21:56Z
dc.date.available2023-09-13T08:21:56Z
dc.date.issued2023-05-11
dc.description.abstractUnderstanding human actions that involve interacting with objects is very important due to the wide range of real-world applications, such as security surveillance and healthcare. In this thesis, three different approaches are presented for addressing the problem of human-object interactions (HOIs) recognition in videos. Firstly, we propose a hierarchical framework for analyzing human-object interactions in a video sequence. The framework comprises Long Short-Term Memory (LSTM) networks that capture human motion and temporal object information independently. These pieces of information are then combined through a bilinear layer and fed into a global deep LSTM to learn high-level information about HOIs. To concentrate on the key components of human and object temporal information, the proposed approach incorporates an attention mechanism into LSTMs. Secondly, we aim to achieve a holistic understanding of human-object interactions (HOIs) by exploiting both their local and global contexts through knowledge distillation. The local context graphs are used to learn the relationship between humans and objects at the frame level by capturing their co-occurrence at a specific time step. On the other hand, the global relation graph is constructed based on the video-level of human and object interactions, identifying their long-term relations throughout a video sequence. We investigate how knowledge from these context graphs can be distilled to their counterparts to improve HOI recognition. Lastly, we propose the Spatio-Temporal Interaction Transformer-based (STIT) network to reason about spatio-temporal changes of humans and objects. Specifically, the spatial transformers learn the local context of humans and objects at specific frame times. The temporal transformer then learns the relations at a higher level between spatial context representations at different time steps, capturing long-term dependencies across frames. We further investigate multiple hierarchy designs for learning human interactions. The effectiveness of each of the proposed methods mentioned above is evaluated using various video action datasets that include human-object interactions, such as Charades, CAD-120, and Something-Something V1.
dc.format.extent129
dc.identifier.urihttps://hdl.handle.net/20.500.14154/69154
dc.language.isoen
dc.publisherSaudi Digital Library
dc.subjectHuman-Object Interactions (HOIs)
dc.subjectKnowledge Distillation (KD)
dc.subjectGlobal Context
dc.subjectLocal Context
dc.subjectAttention
dc.subjectSpatio-Temporal
dc.subjectLong-Term Dependencies
dc.titleRecognizing Human-Object Interactions in Videos
dc.typeThesis
sdl.degree.departmentComputer Science
sdl.degree.disciplineComputer Vision, Human-Object Interactions
sdl.degree.grantorDurham University
sdl.degree.nameDoctor of Philosophy

Files

Copyright owned by the Saudi Digital Library (SDL) © 2025