Show simple item record

dc.contributor.authorShanker, Shreejithen
dc.date.accessioned2022-05-31T16:47:53Z
dc.date.available2022-05-31T16:47:53Z
dc.date.createdJuly 2022en
dc.date.issued2022en
dc.date.submitted2022en
dc.identifier.citationEashan Wadhwa, Shashwat Khandelwal, Shreejith Shanker, IMEC: A Memory-Efficient Convolution Algorithm For Quantised Neural Network Accelerators, 33rd IEEE International Conference on Application-specific Systems, Architectures and Processors, Gothenburg, Sweden, July 2022, IEEE, 2022en
dc.identifier.otherYen
dc.descriptionPUBLISHEDen
dc.descriptionGothenburg, Swedenen
dc.description.abstractQuantised convolution neural networks (QCNNs) on FPGAs have shown tremendous potential for deploying deep learning on resource constrained devices closer to the data source or in embedded applications. An essential building block of (Q)CNNs are the convolutional layers. FPGA implementations use modified versions of convolution kernels to reduce the resource overheads using variations of the sliding kernel algorithm. While these alleviate resource consumption to a certain degree, they still incur considerable (distributed) memory resources, requiring the use of larger FPGA devices with sufficient on-chip memory elements to implement deep QCNNs. In this paper, we present the Inverse Memory Efficient Convolution (IMEC) algorithm, a novel strategy to lower the memory consumption of convolutional layers in QCNNs. IMEC lowers the footprint of intermediate matrix buffers incurred within the convolutional layers and the multiply- accumulate (MAC) operators required at each layer through a series of data organisation and computational optimisations. We evaluate IMEC by integrating it into the BNN-PYNQ framework that can compile high-level QCNN representations to the FPGA bitstream. Our results show that IMEC can optimise memory footprint and the overall resource overhead of the convolutional layers by ∼33% and ∼20% (LUT and FF count) respectively, across multiple quantisation levels (1-bit to 8-bit), while maintaining identical inference accuracy as the state-of-the-art QCNN implementations.en
dc.language.isoenen
dc.publisherIEEEen
dc.rightsYen
dc.subjectResearch Subject Categories::TECHNOLOGYen
dc.subjectConvolution Neural Networksen
dc.subjectField Programmable Gate Arraysen
dc.subjectInference Algorithmsen
dc.titleIMEC: A Memory-Efficient Convolution Algorithm For Quantised Neural Network Acceleratorsen
dc.title.alternative33rd IEEE International Conference on Application-specific Systems, Architectures and Processorsen
dc.typeConference Paperen
dc.type.supercollectionscholarly_publicationsen
dc.type.supercollectionrefereed_publicationsen
dc.identifier.peoplefinderurlhttp://people.tcd.ie/shankersen
dc.identifier.rssinternalid243374en
dc.rights.ecaccessrightsopenAccess
dc.relation.sourceFINNen
dc.subject.TCDThemeMaking Irelanden
dc.subject.TCDThemeSmart & Sustainable Planeten
dc.subject.TCDThemeTelecommunicationsen
dc.subject.TCDTagARTIFICIAL NEURAL NETWORKSen
dc.subject.TCDTagQuantised Neural Networksen
dc.subject.TCDTagReconfigurable Computingen
dc.relation.sourceurihttps://xilinx.github.io/finn/en
dc.identifier.orcid_id0000-0002-9717-1804en
dc.status.accessibleNen
dc.identifier.urihttp://hdl.handle.net/2262/98718


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record