|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM359510914 |
003 |
DE-627 |
005 |
20231226081030.0 |
007 |
cr uuu---uuuuu |
008 |
231226s2023 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TPAMI.2023.3295772
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1198.xml
|
035 |
|
|
|a (DE-627)NLM359510914
|
035 |
|
|
|a (NLM)37450360
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Gupta, Akshita
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Generative Multi-Label Zero-Shot Learning
|
264 |
|
1 |
|c 2023
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Revised 07.11.2023
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a Multi-label zero-shot learning strives to classify images into multiple unseen categories for which no data is available during training. The test samples can additionally contain seen categories in the generalized variant. Existing approaches rely on learning either shared or label-specific attention from the seen classes. Nevertheless, computing reliable attention maps for unseen classes during inference in a multi-label setting is still a challenge. In contrast, state-of-the-art single-label generative adversarial network (GAN) based approaches learn to directly synthesize the class-specific visual features from the corresponding class attribute embeddings. However, synthesizing multi-label features from GANs is still unexplored in the context of zero-shot setting. When multiple objects occur jointly in a single image, a critical question is how to effectively fuse multi-class information. In this work, we introduce different fusion approaches at the attribute-level, feature-level and cross-level (across attribute and feature-levels) for synthesizing multi-label features from their corresponding multi-label class embeddings. To the best of our knowledge, our work is the first to tackle the problem of multi-label feature synthesis in the (generalized) zero-shot setting. Our cross-level fusion-based generative approach outperforms the state-of-the-art on three zero-shot benchmarks: NUS-WIDE, Open Images and MS COCO. Furthermore, we show the generalization capabilities of our fusion approach in the zero-shot detection task on MS COCO, achieving favorable performance against existing methods
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Narayan, Sanath
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Khan, Salman
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Khan, Fahad Shahbaz
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Shao, Ling
|e verfasserin
|4 aut
|
700 |
1 |
|
|a van de Weijer, Joost
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on pattern analysis and machine intelligence
|d 1979
|g 45(2023), 12 vom: 14. Dez., Seite 14611-14624
|w (DE-627)NLM098212257
|x 1939-3539
|7 nnns
|
773 |
1 |
8 |
|g volume:45
|g year:2023
|g number:12
|g day:14
|g month:12
|g pages:14611-14624
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TPAMI.2023.3295772
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 45
|j 2023
|e 12
|b 14
|c 12
|h 14611-14624
|