|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM347306217 |
003 |
DE-627 |
005 |
20231226033522.0 |
007 |
cr uuu---uuuuu |
008 |
231226s2023 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TPAMI.2022.3212615
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1157.xml
|
035 |
|
|
|a (DE-627)NLM347306217
|
035 |
|
|
|a (NLM)36215372
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Lin, Mingbao
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a SiMaN
|b Sign-to-Magnitude Network Binarization
|
264 |
|
1 |
|c 2023
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Completed 10.04.2023
|
500 |
|
|
|a Date Revised 10.04.2023
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a Binary neural networks (BNNs) have attracted broad research interest due to their efficient storage and computational ability. Nevertheless, a significant challenge of BNNs lies in handling discrete constraints while ensuring bit entropy maximization, which typically makes their weight optimization very difficult. Existing methods relax the learning using the sign function, which simply encodes positive weights into +1s, and -1s otherwise. Alternatively, we formulate an angle alignment objective to constrain the weight binarization to {0,+1} to solve the challenge. In this article, we show that our weight binarization provides an analytical solution by encoding high-magnitude weights into +1s, and 0s otherwise. Therefore, a high-quality discrete solution is established in a computationally efficient manner without the sign function. We prove that the learned weights of binarized networks roughly follow a Laplacian distribution that does not allow entropy maximization, and further demonstrate that it can be effectively solved by simply removing the l2 regularization during network training. Our method, dubbed sign-to-magnitude network binarization (SiMaN), is evaluated on CIFAR-10 and ImageNet, demonstrating its superiority over the sign-based state-of-the-arts. Our source code, experimental settings, training logs and binary models are available at https://github.com/lmbxmu/SiMaN
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Ji, Rongrong
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Xu, Zihan
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Zhang, Baochang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Chao, Fei
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Lin, Chia-Wen
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Shao, Ling
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on pattern analysis and machine intelligence
|d 1979
|g 45(2023), 5 vom: 10. Mai, Seite 6277-6288
|w (DE-627)NLM098212257
|x 1939-3539
|7 nnns
|
773 |
1 |
8 |
|g volume:45
|g year:2023
|g number:5
|g day:10
|g month:05
|g pages:6277-6288
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TPAMI.2022.3212615
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 45
|j 2023
|e 5
|b 10
|c 05
|h 6277-6288
|