|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM332903885 |
003 |
DE-627 |
005 |
20231225220532.0 |
007 |
cr uuu---uuuuu |
008 |
231225s2021 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TIP.2021.3125263
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1109.xml
|
035 |
|
|
|a (DE-627)NLM332903885
|
035 |
|
|
|a (NLM)34752395
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Han, Yizeng
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Spatially Adaptive Feature Refinement for Efficient Inference
|
264 |
|
1 |
|c 2021
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Completed 10.12.2021
|
500 |
|
|
|a Date Revised 14.12.2021
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status MEDLINE
|
520 |
|
|
|a Spatial redundancy commonly exists in the learned representations of convolutional neural networks (CNNs), leading to unnecessary computation on high-resolution features. In this paper, we propose a novel Spatially Adaptive feature Refinement (SAR) approach to reduce such superfluous computation. It performs efficient inference by adaptively fusing information from two branches: one conducts standard convolution on input features at a lower spatial resolution, and the other one selectively refines a set of regions at the original resolution. The two branches complement each other in feature learning, and both of them evoke much less computation than standard convolution. SAR is a flexible method that can be conveniently plugged into existing CNNs to establish models with reduced spatial redundancy. Experiments on CIFAR and ImageNet classification, COCO object detection and PASCAL VOC semantic segmentation tasks validate that the proposed SAR can consistently improve the network performance and efficiency. Notably, our results show that SAR only refines less than 40% of the regions in the feature representations of a ResNet for 97% of the samples in the validation set of ImageNet to achieve comparable accuracy with the original model, revealing the high computational redundancy in the spatial dimension of CNNs
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Huang, Gao
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Song, Shiji
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Yang, Le
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Zhang, Yitian
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Jiang, Haojun
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|d 1992
|g 30(2021) vom: 09., Seite 9345-9358
|w (DE-627)NLM09821456X
|x 1941-0042
|7 nnns
|
773 |
1 |
8 |
|g volume:30
|g year:2021
|g day:09
|g pages:9345-9358
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TIP.2021.3125263
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 30
|j 2021
|b 09
|h 9345-9358
|