dc.contributor.author | Smolic, Aljosa | |
dc.date.accessioned | 2021-03-14T16:38:37Z | |
dc.date.available | 2021-03-14T16:38:37Z | |
dc.date.issued | 2020 | |
dc.date.submitted | 2020 | en |
dc.identifier.citation | Wang, Z., She, Q., Chalasani, T., Smolic, A., "CatNet: Class Incremental 3D ConvNets for Lifelong Egocentric Gesture Recognition," 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 2020, pp. 935-944 | en |
dc.identifier.other | Y | |
dc.description.abstract | Egocentric gestures are the most natural form of communication for humans to interact with wearable devices such as VR/AR helmets and glasses. A major issue in such scenarios for real-world applications is that may easily become necessary to add new gestures to the system e.g., a proper VR system should allow users to customize gestures incrementally. Traditional deep learning methods require storing all previous class samples in the system and training the model again from scratch by incorporating previous samples and new samples, which costs humongous memory and significantly increases computation over time. In this work, we demonstrate a lifelong 3D convolutional framework - c(C)la(a)ss increment(t)al net(Net)works (CatNet), which considers temporal information in videos and enables life-long learning for egocentric gesture video recognition by learning the feature representation of an exemplar set selected from previous class samples. Importantly, we propose a two-stream CatNet, which deploys RGB and depth modalities to train two separate networks. We evaluate Cat- Nets on a publicly available dataset - EgoGesture dataset, and show that CatNets can learn many classes incrementally over a long period of time. Results also demonstrate that the two-stream architecture achieves the best performance on both joint training and class incremental training compared to 3 other one-stream architectures. The codes and pre-trained models used in this work are provided at https://github.com/villawang/CatNet. | en |
dc.language.iso | en | en |
dc.relation.uri | https://v-sense.scss.tcd.ie/wp-content/uploads/2020/11/Wang_CatNet_Class_Incremental_3D_ConvNets_for_Lifelong_Egocentric_Gesture_Recognition_CVPRW_2020_paper.pdf | en |
dc.rights | Y | en |
dc.subject | Videos | en |
dc.subject | Task analysis | en |
dc.subject | Three-dimensional displays | en |
dc.subject | Computer architecture | en |
dc.subject | Training | en |
dc.subject | Spatiotemporal phenomena | en |
dc.subject | Gesture recognition | en |
dc.title | CatNet: Class Incremental 3D ConvNets for Lifelong Egocentric Gesture Recognition | en |
dc.title.alternative | Conference on Computer Vision and Pattern Recognition 2020 (CVPR 2020), 2020. | en |
dc.type | Conference Paper | en |
dc.type.supercollection | scholarly_publications | en |
dc.type.supercollection | refereed_publications | en |
dc.identifier.peoplefinderurl | http://people.tcd.ie/smolica | |
dc.identifier.rssinternalid | 225562 | |
dc.identifier.doi | 10.1109/CVPRW50498.2020.00123 | |
dc.rights.ecaccessrights | openAccess | |
dc.relation.doi | 10.1109/CVPRW50498.2020.00123 | en |
dc.relation.cites | Cites | en |
dc.relation.cites | Cites | en |
dc.subject.TCDTheme | Creative Technologies | en |
dc.subject.TCDTheme | Digital Engagement | en |
dc.subject.TCDTag | Data Analysis | en |
dc.subject.TCDTag | Information technology in education | en |
dc.subject.TCDTag | Multimedia & Creativity | en |
dc.identifier.rssuri | https://v-sense.scss.tcd.ie/wp-content/uploads/2020/11/Wang_CatNet_Class_Incremental_3D_ConvNets_for_Lifelong_Egocentric_Gesture_Recognition_CVPRW_2020_paper.pdf | |
dc.subject.darat_impairment | Other | en |
dc.status.accessible | N | en |
dc.contributor.sponsor | Science Foundation Ireland (SFI) | en |
dc.contributor.sponsorGrantNumber | 15/RP/2776 | en |
dc.identifier.uri | http://hdl.handle.net/2262/95670 | |