Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/171831
Full metadata record
DC FieldValueLanguage
dc.contributor.authorPan, Jianhongen_US
dc.contributor.authorYang, Siyuanen_US
dc.contributor.authorFoo, Lin Gengen_US
dc.contributor.authorKe, Qiuhongen_US
dc.contributor.authorRahmani, Hosseinen_US
dc.contributor.authorFan, Zhipengen_US
dc.contributor.authorLiu, Junen_US
dc.date.accessioned2023-11-09T04:26:58Z-
dc.date.available2023-11-09T04:26:58Z-
dc.date.issued2023-
dc.identifier.citationPan, J., Yang, S., Foo, L. G., Ke, Q., Rahmani, H., Fan, Z. & Liu, J. (2023). Progressive channel-shrinking network. IEEE Transactions On Multimedia. https://dx.doi.org/10.1109/TMM.2023.3291197en_US
dc.identifier.issn1520-9210en_US
dc.identifier.urihttps://hdl.handle.net/10356/171831-
dc.description.abstractCurrently, salience-based channel pruning makes continuous breakthroughs in network compression. In the realization, the salience mechanism is used as a metric of channel salience to guide pruning. Therefore, salience-based channel pruning can dynamically adjust the channel width at run-time, which provides a flexible pruning scheme. However, there are two problems emerging: a gating function is often needed to truncate the specific salience entries to zero, which destabilizes the forward propagation; dynamic architecture brings more cost for indexing in inference which bottlenecks the inference speed. In this paper, we propose a Progressive Channel-Shrinking (PCS) method to compress the selected salience entries at run-time instead of roughly approximating them to zero. We also propose a Running Shrinking Policy to provide a testing-static pruning scheme that can reduce the memory access cost for filter indexing. We evaluate our method on ImageNet and CIFAR10 datasets over two prevalent networks: ResNet and VGG, and demonstrate that our PCS outperforms all baselines and achieves state-of-the-art in terms of compression-performance tradeoff. Moreover, we observe a significant and practical acceleration of inference. The code will be released upon acceptance.en_US
dc.description.sponsorshipMinistry of Education (MOE)en_US
dc.description.sponsorshipNational Research Foundation (NRF)en_US
dc.language.isoenen_US
dc.relationT2EP20222-0035en_US
dc.relationAISG-100E-2020-065en_US
dc.relation.ispartofIEEE Transactions on Multimediaen_US
dc.rights© 2023 IEEE. All rights reserved.en_US
dc.subjectEngineering::Computer science and engineeringen_US
dc.titleProgressive channel-shrinking networken_US
dc.typeJournal Articleen
dc.contributor.schoolInterdisciplinary Graduate School (IGS)en_US
dc.identifier.doi10.1109/TMM.2023.3291197-
dc.identifier.scopus2-s2.0-85163437770-
dc.subject.keywordsProgressiveen_US
dc.subject.keywordsNetwork Shrinkingen_US
dc.description.acknowledgementThis work is supported by MOE AcRF Tier 2 (Proposal ID: T2EP20222-0035), National Research Foundation Singapore under its AI Singapore Programme (AISG-100E-2020-065), and SUTD SKI Project (SKI 2021 02 06). This work is also supported by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215.en_US
item.grantfulltextnone-
item.fulltextNo Fulltext-
Appears in Collections:IGS Journal Articles

Page view(s)

141
Updated on Sep 18, 2024

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.