Show simple item record

dc.contributor.advisorLacey, Gerard
dc.contributor.authorBruton, Seán
dc.date.accessioned2021-03-01T15:36:10Z
dc.date.available2021-03-01T15:36:10Z
dc.date.issued2021en
dc.date.submitted2021
dc.identifier.citationBruton, Seán, Recognising the fine-grained actions of a goal-directed activity from multi-modal images, Trinity College Dublin.School of Computer Science & Statistics, 2021en
dc.identifier.otherYen
dc.descriptionAPPROVEDen
dc.description.abstractThe ability to understand and respond to human activities can form the basis of many pervasive computing applications. Recognising the constituent actions of an activity can lead to a more detailed understanding of the activity and provide opportunities to develop applications for monitoring, training and assistance. We address the specific problem of recognising the fine-grained actions of a fixed-setting goal-directed activity from RGB-D videos. We design a novel convolutional neural network architecture, WeaveNet, for fine-grained action recognition from multiple image types. A spatio-temporal fusion method, Densely-Fused Action Images, is also presented for use in combination with WeaveNet. This combined architecture achieves an accuracy of 82.7\% at a mid-level granularity on a benchmark dataset, an improvement of 9\% over existing methods. We contribute a system for recording fine-grained actions involved in human-object interaction tasks, specifically including clinical skills. The system is novel due to its ability to record actions from multiple viewpoints using RGB-D cameras in a synchronised way. We present a dataset of clinical skill performances for the skill of venepuncture, including 60 performances, across 20 subjects, totalling over 15 hours of footage. The multi-modal, multi-camera characteristics of this dataset make it amenable to many fine-grained action recognition techniques. Together, the fine-grained action recognition technique, the system for recoding human-object interactions, and the dataset of clinical skill performances, make a significant contribution towards the development of next-generation pervasive computing applications.en
dc.language.isoenen
dc.publisherTrinity College Dublin. School of Computer Science & Statistics. Discipline of Computer Scienceen
dc.rightsYen
dc.subjectAction recognitionen
dc.subjectConvolutional neural networksen
dc.subjectMulti-modal imagesen
dc.titleRecognising the fine-grained actions of a goal-directed activity from multi-modal imagesen
dc.typeThesisen
dc.type.supercollectionthesis_dissertationsen
dc.type.supercollectionrefereed_publicationsen
dc.type.qualificationlevelDoctoralen
dc.identifier.peoplefinderurlhttps://tcdlocalportal.tcd.ie/pls/EnterApex/f?p=800:71:0::::P71_USERNAME:BRUTONSen
dc.identifier.rssinternalid224598en
dc.rights.ecaccessrightsopenAccess
dc.contributor.sponsorIrish Research Council (IRC)en
dc.identifier.urihttp://hdl.handle.net/2262/95437


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record