|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM358124921 |
003 |
DE-627 |
005 |
20231226074059.0 |
007 |
cr uuu---uuuuu |
008 |
231226s2023 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TIP.2023.3285426
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1193.xml
|
035 |
|
|
|a (DE-627)NLM358124921
|
035 |
|
|
|a (NLM)37310815
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Niu, Kai
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Improving Inconspicuous Attributes Modeling for Person Search by Language
|
264 |
|
1 |
|c 2023
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Completed 20.06.2023
|
500 |
|
|
|a Date Revised 20.06.2023
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a Person search by language aims to retrieve the interested pedestrian images based on natural language sentences. Although great efforts have been made to address the cross-modal heterogeneity, most of the current solutions suffer from only capturing salient attributes while ignoring inconspicuous ones, being weak in distinguishing very similar pedestrians. In this work, we propose the Adaptive Salient Attribute Mask Network (ASAMN) to adaptively mask the salient attributes for cross-modal alignments, and therefore induce the model to simultaneously focus on inconspicuous attributes. Specifically, we consider the uni-modal and cross-modal relations for masking salient attributes in the Uni-modal Salient Attribute Mask (USAM) and Cross-modal Salient Attribute Mask (CSAM) modules, respectively. Then the Attribute Modeling Balance (AMB) module is presented to randomly select a proportion of masked features for cross-modal alignments, ensuring the balance of modeling capacity of both salient attributes and inconspicuous ones. Extensive experiments and analyses have been carried out to validate the effectiveness and generalization capacity of our proposed ASAMN method, and we have obtained the state-of-the-art retrieval performance on the widely-used CUHK-PEDES and ICFG-PEDES benchmarks
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Huang, Tao
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Huang, Linjiang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Wang, Liang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Zhang, Yanning
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|d 1992
|g 32(2023) vom: 01., Seite 3429-3441
|w (DE-627)NLM09821456X
|x 1941-0042
|7 nnns
|
773 |
1 |
8 |
|g volume:32
|g year:2023
|g day:01
|g pages:3429-3441
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TIP.2023.3285426
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 32
|j 2023
|b 01
|h 3429-3441
|