TY - GEN
T1 - Channel-Wise Pruning via Learn Gates&BN for Object Detection Tasks
AU - Chen, Min Xiang
AU - Chen, Po Han
AU - Tsai, Chia Chi
N1 - Publisher Copyright:
© 2024, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
PY - 2024
Y1 - 2024
N2 - Network pruning is an important research area aimed at addressing the high computational costs of deep neural networks. Previous studies [1] have indicated that it is not necessary to follow the conventional pruning process of training a large, redundant network, but rather, more diverse and higher-performing potential models can be directly pruned from randomly initialized weights. However, our experimental results on the MS COCO 2017 dataset [2] demonstrate that this approach is not applicable when compressing object detection models. We believe this outcome is associated with the complex network structures involved in object detection, making it relatively challenging to explore pruned architectures from random weights. To address this issue, we improved the existing Learn gates method and incorporated Batch normalization [3] to jointly learn channel importance. This enhances the learning capability of channel importance in a shorter time frame and facilitates the exploration of suitable pruned network structures within pre-trained weights. When applying our network pruning method to object detection models YOLOv3 [4] and YOLOv4 [5], our approach achieves higher accuracy with only a brief period of network structure learning.
AB - Network pruning is an important research area aimed at addressing the high computational costs of deep neural networks. Previous studies [1] have indicated that it is not necessary to follow the conventional pruning process of training a large, redundant network, but rather, more diverse and higher-performing potential models can be directly pruned from randomly initialized weights. However, our experimental results on the MS COCO 2017 dataset [2] demonstrate that this approach is not applicable when compressing object detection models. We believe this outcome is associated with the complex network structures involved in object detection, making it relatively challenging to explore pruned architectures from random weights. To address this issue, we improved the existing Learn gates method and incorporated Batch normalization [3] to jointly learn channel importance. This enhances the learning capability of channel importance in a shorter time frame and facilitates the exploration of suitable pruned network structures within pre-trained weights. When applying our network pruning method to object detection models YOLOv3 [4] and YOLOv4 [5], our approach achieves higher accuracy with only a brief period of network structure learning.
UR - http://www.scopus.com/inward/record.url?scp=85184285130&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85184285130&partnerID=8YFLogxK
U2 - 10.1007/978-981-99-9412-0_25
DO - 10.1007/978-981-99-9412-0_25
M3 - Conference contribution
AN - SCOPUS:85184285130
SN - 9789819994113
T3 - Lecture Notes in Electrical Engineering
SP - 238
EP - 249
BT - Genetic and Evolutionary Computing - Proceedings of the Fifteenth International Conference on Genetic and Evolutionary Computing, 2023
A2 - Pan, Jeng-Shyang
A2 - Pan, Zhigeng
A2 - Hu, Pei
A2 - Lin, Jerry Chun-Wei
PB - Springer Science and Business Media Deutschland GmbH
T2 - 15th International Conference on Genetic and Evolutionary Computing, ICGEC 2023
Y2 - 6 October 2023 through 8 October 2023
ER -