|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM273587951 |
003 |
DE-627 |
005 |
20231225001550.0 |
007 |
cr uuu---uuuuu |
008 |
231225s2017 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TIP.2017.2722238
|2 doi
|
028 |
5 |
2 |
|a pubmed24n0911.xml
|
035 |
|
|
|a (DE-627)NLM273587951
|
035 |
|
|
|a (NLM)28682255
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Le Meur, Olivier
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Visual Attention Saccadic Models Learn to Emulate Gaze Patterns From Childhood to Adulthood
|
264 |
|
1 |
|c 2017
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Completed 30.07.2018
|
500 |
|
|
|a Date Revised 30.07.2018
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status PubMed-not-MEDLINE
|
520 |
|
|
|a How people look at visual information reveals fundamental information about themselves, their interests and their state of mind. While previous visual attention models output static 2D saliency maps, saccadic models aim to predict not only where observers look at but also how they move their eyes to explore the scene. In this paper, we demonstrate that saccadic models are a flexible framework that can be tailored to emulate observer's viewing tendencies. More specifically, we use fixation data from 101 observers split into five age groups (adults, 8-10 y.o., 6-8 y.o., 4-6 y.o., and 2 y.o.) to train our saccadic model for different stages of the development of human visual system. We show that the joint distribution of saccade amplitude and orientation is a visual signature specific to each age group, and can be used to generate age-dependent scan paths. Our age-dependent saccadic model does not only output human-like, age-specific visual scan paths, but also significantly outperforms other state-of-the-art saliency models. We demonstrate that the computational modeling of visual attention, through the use of saccadic model, can be efficiently adapted to emulate the gaze behavior of a specific group of observers
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Coutrot, Antoine
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Liu, Zhi
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Rama, Pia
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Le Roch, Adrien
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Helo, Andrea
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|d 1992
|g 26(2017), 10 vom: 06. Okt., Seite 4777-4789
|w (DE-627)NLM09821456X
|x 1941-0042
|7 nnns
|
773 |
1 |
8 |
|g volume:26
|g year:2017
|g number:10
|g day:06
|g month:10
|g pages:4777-4789
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TIP.2017.2722238
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d 26
|j 2017
|e 10
|b 06
|c 10
|h 4777-4789
|