|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM314594337 |
003 |
DE-627 |
005 |
20231225153050.0 |
007 |
cr uuu---uuuuu |
008 |
231225s2023 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TPAMI.2020.3021756
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1048.xml
|
035 |
|
|
|a (DE-627)NLM314594337
|
035 |
|
|
|a (NLM)32886607
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Li, Shijie
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a MS-TCN++
|b Multi-Stage Temporal Convolutional Network for Action Segmentation
|
264 |
|
1 |
|c 2023
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Completed 07.05.2023
|
500 |
|
|
|a Date Revised 07.05.2023
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a With the success of deep learning in classifying short trimmed videos, more attention has been focused on temporally segmenting and classifying activities in long untrimmed videos. State-of-the-art approaches for action segmentation utilize several layers of temporal convolution and temporal pooling. Despite the capabilities of these approaches in capturing temporal dependencies, their predictions suffer from over-segmentation errors. In this paper, we propose a multi-stage architecture for the temporal action segmentation task that overcomes the limitations of the previous approaches. The first stage generates an initial prediction that is refined by the next ones. In each stage we stack several layers of dilated temporal convolutions covering a large receptive field with few parameters. While this architecture already performs well, lower layers still suffer from a small receptive field. To address this limitation, we propose a dual dilated layer that combines both large and small receptive fields. We further decouple the design of the first stage from the refining stages to address the different requirements of these stages. Extensive evaluation shows the effectiveness of the proposed model in capturing long-range dependencies and recognizing action segments. Our models achieve state-of-the-art results on three datasets: 50Salads, Georgia Tech Egocentric Activities (GTEA), and the Breakfast dataset
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Farha, Yazan Abu
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Liu, Yun
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Cheng, Ming-Ming
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Gall, Juergen
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on pattern analysis and machine intelligence
|d 1979
|g 45(2023), 6 vom: 04. Juni, Seite 6647-6658
|w (DE-627)NLM098212257
|x 1939-3539
|7 nnns
|
773 |
1 |
8 |
|g volume:45
|g year:2023
|g number:6
|g day:04
|g month:06
|g pages:6647-6658
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TPAMI.2020.3021756
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 45
|j 2023
|e 6
|b 04
|c 06
|h 6647-6658
|