|
|
|
|
LEADER |
01000caa a22002652 4500 |
001 |
NLM374424284 |
003 |
DE-627 |
005 |
20240710232445.0 |
007 |
cr uuu---uuuuu |
008 |
240703s2024 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TIP.2024.3411452
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1466.xml
|
035 |
|
|
|a (DE-627)NLM374424284
|
035 |
|
|
|a (NLM)38954579
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Wang, Shuo
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Feature Mixture on Pre-Trained Model for Few-Shot Learning
|
264 |
|
1 |
|c 2024
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Revised 10.07.2024
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a Few-shot learning (FSL) aims at recognizing a novel object under limited training samples. A robust feature extractor (backbone) can significantly improve the recognition performance of the FSL model. However, training an effective backbone is a challenging issue since 1) designing and validating structures of backbones are time-consuming and expensive processes, and 2) a backbone trained on the known (base) categories is more inclined to focus on the textures of the objects it learns, which is hard to describe the novel samples. To solve these problems, we propose a feature mixture operation on the pre-trained (fixed) features: 1) We replace a part of the values of the feature map from a novel category with the content of other feature maps to increase the generalizability and diversity of training samples, which avoids retraining a complex backbone with high computational costs. 2) We use the similarities between the features to constrain the mixture operation, which helps the classifier focus on the representations of the novel object where these representations are hidden in the features from the pre-trained backbone with biased training. Experimental studies on five benchmark datasets in both inductive and transductive settings demonstrate the effectiveness of our feature mixture (FM). Specifically, compared with the baseline on the Mini-ImageNet dataset, it achieves 3.8% and 4.2% accuracy improvements for 1 and 5 training samples, respectively. Additionally, the proposed mixture operation can be used to improve other existing FSL methods based on backbone training
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Lu, Jinda
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Xu, Haiyang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Hao, Yanbin
|e verfasserin
|4 aut
|
700 |
1 |
|
|a He, Xiangnan
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|d 1992
|g 33(2024) vom: 01., Seite 4104-4115
|w (DE-627)NLM09821456X
|x 1941-0042
|7 nnns
|
773 |
1 |
8 |
|g volume:33
|g year:2024
|g day:01
|g pages:4104-4115
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TIP.2024.3411452
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 33
|j 2024
|b 01
|h 4104-4115
|