Show simple item record

dc.contributor.advisorBarrett, Edmonden
dc.contributor.authorGUERIN, DAVIDen
dc.date.accessioned2019-09-18T06:23:18Z
dc.date.available2019-09-18T06:23:18Z
dc.date.issued2019en
dc.date.submitted2019en
dc.identifier.citationGUERIN, DAVID, Computational Shedding in Stream Computing, Trinity College Dublin.School of Computer Science & Statistics, 2019en
dc.identifier.otherYen
dc.descriptionAPPROVEDen
dc.description.abstractStream Computing, a generic on-line data processing paradigm has emerged as the preferred approach in the processing of continuous data streams. Data streams suffer from a bursty characteristic where the data rate of the stream can spike temporarily up to orders of magnitude greater than normal expected levels. As such, producing timely application results is difficult as queues fill. The classic response to these temporary scenarios is to shed input data. However, Load Shedding (LS) impacts negatively on application output accuracy as relevant data is discarded before it is processed. Further, LS rates tend to be proportional to the input rate of the stream, as such high data rates can lead to high data loss during overload events. For many classes of applications, this can have a particularly negative impact on the quality of the output result, given that data is simply not processed before it is shed. This thesis presents a new approach, Computational Shedding (CS), to the problem of maintaining application result accuracy while attempting to forgo input data loss during transient busty data events. Rather shedding input data within the stream, we propose to adapt the application and shed tasks or subtasks of the executing application temporarily, to reduce message process costs in the stream. As such, this mechanism provides for an opportunity to temporarily increase processing capacity thereby forgoing the need for deliberate data loss during a bursty data event. We have evaluated this approach against traditional LS techniques in a number of ways, such as in terms of output application accuracy and application processing duration. In experimentation, we have found in applicable applications, subtasks can be discarded for a time. Remaining subtasks can continue to produce a valid imprecise result. CS was compared to LS alternatives which simply do not process any discarded data. We show that through the results of our evaluation, CS leads to more timely and accurate results when compared to LS alternatives.en
dc.publisherTrinity College Dublin. School of Computer Science & Statistics. Discipline of Computer Scienceen
dc.rightsYen
dc.subjectStream Computingen
dc.subjectComputational Sheddingen
dc.subjectLoad Sheddingen
dc.subjectData Stream Processingen
dc.subjectStream Computing Frameworken
dc.subjectBursty Data Streamsen
dc.subjectApproximate Resulten
dc.titleComputational Shedding in Stream Computingen
dc.typeThesisen
dc.type.supercollectionthesis_dissertationsen
dc.type.supercollectionrefereed_publicationsen
dc.type.qualificationlevelDoctoralen
dc.identifier.peoplefinderurlhttps://tcdlocalportal.tcd.ie/pls/EnterApex/f?p=800:71:0::::P71_USERNAME:DGUERINen
dc.identifier.rssinternalid206918en
dc.rights.ecaccessrightsopenAccess
dc.contributor.sponsorTCDen
dc.identifier.urihttp://hdl.handle.net/2262/89513


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record