Show simple item record

dc.contributor.authorGraham, Yvetteen
dc.date.accessioned2021-03-28T15:36:39Z
dc.date.available2021-03-28T15:36:39Z
dc.date.issued2019en
dc.date.submitted2019en
dc.identifier.citationAlan Smeaton, Yvette Graham, Kevin McGuinness, Noel O'Connor, Se?n Quinn, Eric Arazo Sanchez, The Impact of Training Data Bias on Automatic Generation of Video Captions, Proceedings of the 25th International Conference on MultiMedia Modeling, 25th International Conference on MultiMedia Modeling, Thessaloniki, Greece, 2019, 178 - 190en
dc.identifier.otherYen
dc.descriptionPUBLISHEDen
dc.descriptionThessaloniki, Greeceen
dc.description.abstractA major issue in machine learning is availability of training data. While this historically referred to the availability of a sufficient volume of training data, recently this has shifted to the availability of sufficient unbiased training data. In this paper we focus on the effect of training data bias on an emerging multimedia application, the automatic captioning of short video clips. We use subsets of the same training data to generate different models for video captioning using the same machine learning technique and we evaluate the performances of different training data subsets using a well-known video caption benchmark, TRECVid. We train using the MSR-VTT video-caption pairs and we prune this to reduce and make the set of captions describing a video more homogeneously similar, or more diverse, or we prune randomly. We then assess the effectiveness of caption-generating trained with these variations using automatic metrics as well as direct assessment by human assessors. Our findings are preliminary and show that randomly pruning captions from the training data yields the worst performance and that pruning to make the data more homogeneous, or diverse, does improve performance slightly when compared to random. Our work points to the need for more training data, both more video clips but, more importantly, more captions for those videos.en
dc.format.extent178en
dc.format.extent190en
dc.language.isoenen
dc.rightsYen
dc.subjectVideo-to-languageen
dc.subjectVideo captioningen
dc.subjectVideo understandingen
dc.subjectSemantic similarityen
dc.titleThe Impact of Training Data Bias on Automatic Generation of Video Captionsen
dc.title.alternativeProceedings of the 25th International Conference on MultiMedia Modelingen
dc.title.alternative25th International Conference on MultiMedia Modelingen
dc.typeConference Paperen
dc.type.supercollectionscholarly_publicationsen
dc.type.supercollectionrefereed_publicationsen
dc.identifier.peoplefinderurlhttp://people.tcd.ie/ygrahamen
dc.identifier.rssinternalid226540en
dc.identifier.doihttp://dx.doi.org/10.1007%2F978-3-030-05710-7_15en
dc.rights.ecaccessrightsopenAccess
dc.subject.TCDThemeCreative Technologiesen
dc.subject.TCDTagNatural Language Processingen
dc.identifier.orcid_id0000-0001-6741-4855en
dc.subject.darat_thematicCommunicationen
dc.subject.darat_thematicGlobalizationen
dc.status.accessibleNen
dc.contributor.sponsorSFI stipenden
dc.contributor.sponsorGrantNumber12/RC/2289en
dc.identifier.urihttp://hdl.handle.net/2262/95914


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record