Tianyu Zhang tianyu.zhang2@durham.ac.uk
PGR Student Doctor of Philosophy
FMDConv: Fast multi-attention dynamic convolution via speed-accuracy trade-off
Zhang, Tianyu; Wan, Fan; Duan, Haoran; Tong, Kevin W.; Deng, Jingjing; Long, Yang
Authors
Fan Wan fan.wan@durham.ac.uk
PGR Student Doctor of Philosophy
Haoran Duan haoran.duan@durham.ac.uk
PGR Student Doctor of Philosophy
Kevin W. Tong
Dr Jingjing Deng jingjing.deng@durham.ac.uk
Assistant Professor
Dr Yang Long yang.long@durham.ac.uk
Associate Professor
Abstract
Spatial convolution is fundamental in constructing deep Convolutional Neural Networks (CNNs) for visual recognition. While dynamic convolution enhances model accuracy by adaptively combining static kernels, it incurs significant computational overhead, limiting its deployment in resource-constrained environments such as federated edge computing. To address this, we propose Fast Multi-Attention Dynamic Convolution (FMDConv), which integrates input attention, temperature-degraded kernel attention, and output attention to optimize the speed-accuracy trade-off. FMDConv achieves a better balance between accuracy and efficiency by selectively enhancing feature extraction with lower complexity. Furthermore, we introduce two novel quantitative metrics, the Inverse Efficiency Score and Rate-Correct Score, to systematically evaluate this trade-off. Extensive experiments on CIFAR-10, CIFAR-100, and ImageNet demonstrate that FMDConv reduces the computational cost by up to 49.8% on ResNet-18 and 42.2% on ResNet-50 compared to prior multi-attention dynamic convolution methods while maintaining competitive accuracy. These advantages make FMDConv highly suitable for real-world, resource-constrained applications.
Citation
Zhang, T., Wan, F., Duan, H., Tong, K. W., Deng, J., & Long, Y. (2025). FMDConv: Fast multi-attention dynamic convolution via speed-accuracy trade-off. Knowledge-Based Systems, 317, Article 113393. https://doi.org/10.1016/j.knosys.2025.113393
Journal Article Type | Article |
---|---|
Acceptance Date | Mar 18, 2025 |
Online Publication Date | Apr 7, 2025 |
Publication Date | 2025-05 |
Deposit Date | May 14, 2025 |
Publicly Available Date | May 14, 2025 |
Journal | Knowledge-Based Systems |
Print ISSN | 0950-7051 |
Electronic ISSN | 1872-7409 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 317 |
Article Number | 113393 |
DOI | https://doi.org/10.1016/j.knosys.2025.113393 |
Public URL | https://durham-repository.worktribe.com/output/3789890 |
Files
Published Journal Article
(2.3 Mb)
PDF
Publisher Licence URL
http://creativecommons.org/licenses/by/4.0/
Copyright Statement
/© 2025 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
You might also like
DS-Depth: Dynamic and Static Depth Estimation via a Fusion Cost Volume
(2023)
Journal Article
SID-NERF: Few-Shot Nerf Based on Scene Information Distribution
(2024)
Presentation / Conference Contribution
CTNeRF: Cross-time Transformer for dynamic neural radiance field from monocular video
(2024)
Journal Article
Wearable-based behaviour interpolation for semi-supervised human activity recognition
(2024)
Journal Article
Dynamic Unary Convolution in Transformers
(2023)
Journal Article