|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM328737607 |
003 |
DE-627 |
005 |
20231225203559.0 |
007 |
cr uuu---uuuuu |
008 |
231225s2021 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TIP.2021.3099409
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1095.xml
|
035 |
|
|
|a (DE-627)NLM328737607
|
035 |
|
|
|a (NLM)34329164
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Deng, Jiajun
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a MINet
|b Meta-Learning Instance Identifiers for Video Object Detection
|
264 |
|
1 |
|c 2021
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Revised 05.08.2021
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a Recent advances in video object detection have characterized the exploration of temporal coherence across frames to enhance object detector. Nevertheless, previous solutions either rely on additional inputs (e.g., optical flow) to guide feature aggregation, or complex post-processing to associate bounding boxes. In this paper, we introduce a simple but effective design that learns instance identifiers for instance association in a meta-learning paradigm, which requires no auxiliary inputs or post-processing. Specifically, we present Meta-Learnt Instance Identifier Networks (namely MINet) that novelly meta-learns instance identifiers to recognize identical instances across frames in a single forward-pass, leading to the robust online linking of instances. Technically, depending on the detection results of previous frames, we teach MINet to learn the weights of an instance identifier on the fly, which can be well applied to up-coming frames. Such meta-learning paradigm enables instance identifiers to be flexibly adapted to novel frames at inference. Furthermore, MINet writes/updates the detection results of previous instances into memory and reads from memory when performing inference to encourage temporal consistency for video object detection. Our MINet is appealing in the sense that it is pluggable to any object detection model. Extensive experiments on ImageNet VID dataset demonstrate the superiority of MINet. More remarkably, by integrating MINet into Faster R-CNN, we achieve 80.2% mAP on ImageNet VID dataset
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Pan, Yingwei
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Yao, Ting
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Zhou, Wengang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Li, Houqiang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Mei, Tao
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|d 1992
|g 30(2021) vom: 15., Seite 6879-6891
|w (DE-627)NLM09821456X
|x 1941-0042
|7 nnns
|
773 |
1 |
8 |
|g volume:30
|g year:2021
|g day:15
|g pages:6879-6891
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TIP.2021.3099409
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 30
|j 2021
|b 15
|h 6879-6891
|