Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/171831
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Pan, Jianhong | en_US |
dc.contributor.author | Yang, Siyuan | en_US |
dc.contributor.author | Foo, Lin Geng | en_US |
dc.contributor.author | Ke, Qiuhong | en_US |
dc.contributor.author | Rahmani, Hossein | en_US |
dc.contributor.author | Fan, Zhipeng | en_US |
dc.contributor.author | Liu, Jun | en_US |
dc.date.accessioned | 2023-11-09T04:26:58Z | - |
dc.date.available | 2023-11-09T04:26:58Z | - |
dc.date.issued | 2023 | - |
dc.identifier.citation | Pan, J., Yang, S., Foo, L. G., Ke, Q., Rahmani, H., Fan, Z. & Liu, J. (2023). Progressive channel-shrinking network. IEEE Transactions On Multimedia. https://dx.doi.org/10.1109/TMM.2023.3291197 | en_US |
dc.identifier.issn | 1520-9210 | en_US |
dc.identifier.uri | https://hdl.handle.net/10356/171831 | - |
dc.description.abstract | Currently, salience-based channel pruning makes continuous breakthroughs in network compression. In the realization, the salience mechanism is used as a metric of channel salience to guide pruning. Therefore, salience-based channel pruning can dynamically adjust the channel width at run-time, which provides a flexible pruning scheme. However, there are two problems emerging: a gating function is often needed to truncate the specific salience entries to zero, which destabilizes the forward propagation; dynamic architecture brings more cost for indexing in inference which bottlenecks the inference speed. In this paper, we propose a Progressive Channel-Shrinking (PCS) method to compress the selected salience entries at run-time instead of roughly approximating them to zero. We also propose a Running Shrinking Policy to provide a testing-static pruning scheme that can reduce the memory access cost for filter indexing. We evaluate our method on ImageNet and CIFAR10 datasets over two prevalent networks: ResNet and VGG, and demonstrate that our PCS outperforms all baselines and achieves state-of-the-art in terms of compression-performance tradeoff. Moreover, we observe a significant and practical acceleration of inference. The code will be released upon acceptance. | en_US |
dc.description.sponsorship | Ministry of Education (MOE) | en_US |
dc.description.sponsorship | National Research Foundation (NRF) | en_US |
dc.language.iso | en | en_US |
dc.relation | T2EP20222-0035 | en_US |
dc.relation | AISG-100E-2020-065 | en_US |
dc.relation.ispartof | IEEE Transactions on Multimedia | en_US |
dc.rights | © 2023 IEEE. All rights reserved. | en_US |
dc.subject | Engineering::Computer science and engineering | en_US |
dc.title | Progressive channel-shrinking network | en_US |
dc.type | Journal Article | en |
dc.contributor.school | Interdisciplinary Graduate School (IGS) | en_US |
dc.identifier.doi | 10.1109/TMM.2023.3291197 | - |
dc.identifier.scopus | 2-s2.0-85163437770 | - |
dc.subject.keywords | Progressive | en_US |
dc.subject.keywords | Network Shrinking | en_US |
dc.description.acknowledgement | This work is supported by MOE AcRF Tier 2 (Proposal ID: T2EP20222-0035), National Research Foundation Singapore under its AI Singapore Programme (AISG-100E-2020-065), and SUTD SKI Project (SKI 2021 02 06). This work is also supported by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215. | en_US |
item.grantfulltext | none | - |
item.fulltext | No Fulltext | - |
Appears in Collections: | IGS Journal Articles |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.