Publikationsserver der Universitätsbibliothek Marburg

Titel:Measuring gaze and pupil in the real world: object-based attention,3D eye tracking and applications
Autor:Stoll, Josef
Weitere Beteiligte: Einhäuser-Treyer, Wolfgang (Prof. Dr.)
Veröffentlicht:2015
URI:https://archiv.ub.uni-marburg.de/diss/z2015/0231
URN: urn:nbn:de:hebis:04-z2015-02310
DOI: https://doi.org/10.17192/z2015.0231
DDC: Naturwissenschaften
Titel (trans.):Blick- und Pupillen-Messung in natürlicher Umgebung: Objekt-basierte Aufmerksamkeit, 3D Augenbewegungsmessung und Anwendungen
Publikationsdatum:2015-05-28
Lizenz:https://rightsstatements.org/vocab/InC-NC/1.0/

Dokument

Schlagwörter:
Gaze allocations on objects, Mobile eye-tracking, Locked-in-Syndrom, overt attention, Blicksetzung auf Objekten, Mobile Augenbewegungsmessung, Pupillometrie, eye-tracking, pupillometry, Pupillometry, Locked-in Syndrome

Summary:
This dissertation contains studies on visual attention, as measured by gaze orientation, and the use of mobile eye-tracking and pupillometry in applications. It combines the development of methods for mobile eye-tracking (studies II and III) with experimental studies on gaze guidance and pupillary responses in patients (studies IV and VI) and healthy observers (studies I and V). Object based attention / Study I What is the main factor of fixation guidance in natural scenes? Low-level features or objects? We developed a fixation-predicting model, which regards preferred viewing locations (PVL) per object and combines these distributions over the entirety of existing objects in the scene. Object-based fixation predictions for natural scene viewing perform at par with the best early salience model, that are based on low-level features. However, when stimuli are manipulated so that low-level features and objects are dissociated, the greater prediction power of saliency models diminishes. Thus, we dare to claim, that highly developed saliency models implicitly obtain object-hood and that fixation selection is mainly influenced by objects and much less by low-level features. Consequently, attention guidance in natural scenes is object-based. 3D tracking / Study II The second study focussed on improving calibration procedures for eye-in-head positions with a mobile eye-tracker.We used a mobile eye-tracker prototype, the EyeSeeCam with a high video-oculography (VOG) sampling rate and the technical gadget to follow the users gaze direction instantaneously with a rotatable camera. For a better accuracy in eye-positioning, we explored a refinement in the implementation of the eye-in-head calibration that yields a measure for fixation distance, which led to a mobile eye-tracker 3D calibration. Additionally, by developing the analytical mechanics for parametrically reorienting the gaze-centred camera, the 3D calibration could be applied to reliably record gaze-centred videos. Such videos are suitable as stimuli for investigating gaze-behaviour during object manipulation or object recognition in real worlds point-of-view (PoV) perspective. In fact, the 3D calibration produces a higher accuracy in positioning the gaze-centred camera over the whole 3D visual range. Study III, eye-tracking methods With a further development on the EyeSeeCam we achieved to record gaze-in-world data, by superposing eye-in-head and head-in-world coordinates. This novel approach uses a combination of few absolute head-positions extracted manually from the PoV video and of relative head-shifts integrated over angular velocities and translational accelerations, both given by an inertia measurement unit (IMU) synchronized to the VOG data. Gaze-in-world data consist of room-referenced gaze directions and their origins within the environment. They easily allow to assign fixation targets by using a 3D model of the measuring environment – a strong rationalisation regarding fixation analysis. Applications Study III Daylight is an important perceptual factor for visual comfort, but can also create discomfort glare situations during office work, so we developed to measure its behavioural influences. We achieve to compare luminance distributions and fixations in a real-world setting, by also recording indoor luminance variations time-resolved using luminance maps of a scenery spanning over a 3pi sr. Luminance evaluations in the workplace environment yield a well controlled categorisation of different lighting conditions and a localisation as well as a brightness measure of glare sources.We used common tasks like reading, typing on a computer, a phone call and thinking about a subject. The 3D model gives the possibility to test for gaze distribution shifts in the presence of glare patches and for variations between lighting conditions. Here, a low contrast lighting condition with no sun inside and a high contrast lighting condition with direct sunlight inside were compared. When the participants are not engaged in any visually focused task and the presence of the task support is minimal, the dominant view directions are inclined towards the view outside the window under the low contrast lighting conditions, but this tendency is less apparent and sways more towards the inside of the room under the high contrast lighting condition. This result implicates an avoidance of glare sources in gaze behaviour. In a second more extensive series of experiments, the participants’ subjective assessments of the lighting conditions will be included. Thus, the influence of glare can be analysed in more detail and tested whether visual discomfort judgements are correlated in differences in gaze-behaviour. Study IV The advanced eye-tracker calibration found application in several following projects and included in this dissertation is an investigation with patients suffering either from idiopathic Parkinson’s disease or from progressive supranuclear palsy (PSP) syndrome. PSP’s key symptom is the decreased ability to carry out vertical saccades and thus the main diagnostic feature for differentiating between the two forms of Parkinson’s syndrome. By measuring ocular movements during a rapid (< 20s) procedure with a standardized fixation protocol, we could successfully differentiate pre-diagnosed patients between idiopathic Parkinson’s disease and PSP, thus between PSP patients and HCs too. In PSP patients, the EyeSeeCam detected prominent impairment of both saccade velocity and amplitude. Furthermore, we show the benefits of a mobile eye-tracking device for application in clinical practice. Study V Decision-making is one of the basic cognitive processes of human behaviours and thus, also evokes a pupil dilation. Since this dilation reflects a marker for the temporal occurrence of the decision, we wondered whether individuals can read decisions from another’s pupil and thus become a mentalist. For this purpose, a modified version of the rock-paper-scissors childhood game was played with 3 prototypical opponents, while their eyes were video taped. These videos served as stimuli for further persons, who competed in rock-paper-scissors. Our results show, that reading decisions from a competitor’s pupil can be achieved and players can raise their winning probability significantly above chance. This ability does not require training but the instruction, that the time of maximum pupil dilation was indicative of the opponent’s choice. Therefore we conclude, that people could use the pupil to detect cognitive decisions in another individual, if they get explicit knowledge of the pupil’s utility. Study VI For patients with severe motor disabilities, a robust mean of communication is a crucial factor for well-being. Locked-in-Syndrome (LiS) patients suffer from quadriplegia and lack the ability of articulating their voice, though their consciousness is fully intact. While classic and incomplete LiS allows at least voluntary vertical eye movements or blinks to be used for communication, total LiS patients are not able to perform such movements. What remains, are involuntarily evoked muscle reactions, like it is the case with the pupillary response. The pupil dilation reflects enhanced cognitive or emotional processing, which we successfully observed in LiS patients. Furthermore, we created a communication system based on yes-no questions combined with the task of solving arithmetic problems during matching answer intervals, that yet invokes the most solid pupil dilation usable on a trial-by-trial basis for decoding yes or no as answers. Applied to HCs and patients with various severe motor disabilities, we provide the proof of principle that pupil responses allow communication for all tested HCs and 4/7 typical LiS patients. Résumé Together, the methods established within this thesis are promising advances in measuring visual attention allocation with 3D eye-tracking in real world and in the use of pupillometry as on-line measurement of cognitive processes. The two most outstanding findings are the possibility to communicate with complete LiS patients and further a conclusive evidence that objects are the primary unit of fixation selection in natural scenes.

Bibliographie / References

  1. S5. Brainard, D.H. (1997). The Psychophysics Toolbox. Spat. Vis. 10, 433-436.
  2. Michael Land. Looking and acting: vision and eye movements in natural behaviour. Oxford University Press, 2009.
  3. Michael I Posner and Jin Fan. Attention as an organ system. Topics in integrative neuroscience, pages 31–61, 2008.
  4. S6. Pelli, D.G. (1997). The VideoToolbox software for visual psychophysics: Transforming numbers into movies. Spat. Vis. 10, 437-442.
  5. Benjamin W Tatler, Mary M Hayhoe, Michael F Land, and Dana H Ballard. Eye guidance in natural vision: Reinterpreting salience. Journal of vision, 11(5):5, 2011.
  6. John M. Henderson. Human gaze control during real-world scene perception. Trends in cognitive sciences, 7(11):498–504, November 2003.
  7. Benjamin T. Vincent, Roland Baddeley, Alessia Correani, Tom Troscianko, and Ute Leonards. Do we look at lights? using mixture modelling to distinguish between low-and high-level factors in natural image viewing. Visual Cognition, 17(6-7):856– 879, 2009.
  8. Benjamin W Tatler and Benjamin T Vincent. Systematic tendencies in scene view- ing. Journal of Eye Movement Research, 2(2):1–18, 2008.
  9. Tom Foulsham and Alan Kingstone. Goal-driven and bottom-up gaze in an active real-world search task. In Proceedings of the Symposium on Eye Tracking Research and Applications, ETRA '12, pages 189–192, New York, NY, USA, 2012. ACM.
  10. Roger HS Carpenter. Movements of the eyes (2nd rev. Pion Limited, 1988.
  11. Edward Awh, Artem V. Belopolsky, and Jan Theeuwes. Top-down versus bottom- up attentional control: a failed theoretical dichotomy. Trends in cognitive sciences, 16(8):437–443, August 2012.
  12. Alla Ignashchenkova, Peter W Dicke, Thomas Haarmeier, and Peter Thier. Neuron- specific contribution of the superior colliculus to overt and covert shifts of attention. Nature neuroscience, 7(1):56–64, 2004.
  13. Maria Stamelou, Rohan de Silva, Oscar Arias-Carrión, Evangelia Boura, Matthias Höllerhage, Wolfgang H Oertel, Ulrich Müller, and Günter U Höglinger. Rational therapeutic approaches to progressive supranuclear palsy. Brain, 133(6):1578–1590, 2010.
  14. Robert M Steinman, Zygmunt Pizlo, Tatiana I Forofonova, and Julie Epelboim. One fixates accurately in order to see clearly not because one sees clearly. Spatial vision, 16(3):225–241, 2003.
  15. Antje Nuthmann and John M. Henderson. Object-based attentional selection in sce- ne viewing. Journal of vision, 10(8), 2010.
  16. ER Samuels and E Szabadi. Functional neuroanatomy of the noradrenergic locus coeruleus: its roles in the regulation of arousal and autonomic function part ii: phy- siological and pharmacological manipulations and pathological alterations of locus coeruleus activity in humans. Current neuropharmacology, 6(3):254, 2008.
  17. Josef Stoll, Stefan Kohlbecher, Svenja Marx, Erich Schneider, and Wolfgang Ein- häuser. Mobile three dimensional gaze tracking. Studies in health technology and informatics, 163:616–622, 2010.
  18. Heiner Deubel and Werner X Schneider. Saccade target selection and object recogni- tion: Evidence for a common attentional mechanism. Vision research, 36(12):1827– 1837, 1996.
  19. Denis Oberkampf, Daniel F DeMenthon, and Larry S Davis. Iterative pose esti- mation using coplanar points. In Computer Vision and Pattern Recognition, 1993. Proceedings CVPR'93., 1993 IEEE Computer Society Conference on, pages 626– 627. IEEE, 1993.
  20. Thomas Dera, Guido Boning, Stanislavs Bardins, and Erich Schneider. Low-latency video tracking of horizontal, vertical, and torsional eye movements as a basis for 3dof realtime motion control of a head-mounted camera. In Systems, Man and Cybernetics, 2006. SMC'06. IEEE International Conference on, volume 6, pages 5191–5196. IEEE, 2006.
  21. Xiaohui Shen and Ying Wu. A unified approach to salient object detection via low rank matrix recovery. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pages 853–860. IEEE, 2012.
  22. Zhengyou Zhang. A flexible new technique for camera calibration. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 22(11):1330–1334, 2000.
  23. D Guitton and M Volle. Gaze control in humans: eye-head coordination during orienting movements to targets within and beyond the oculomotor range. Journal of neurophysiology, 58(3):427–459, 1987.
  24. Michael F Land. Vision, eye movements, and natural behavior. Visual neuroscience, 26(01):51–62, 2009.
  25. G. Bauer, F. Gerstenbrand, and E. Rumpl. Varieties of the locked-in syndrome. Journal of Neurology, 221(2):77–91, 1979.
  26. Jackson Beatty and Daniel Kahneman. Pupillary changes in two memory tasks. Psychonomic Science, 5(10):371–372, 1966.
  27. Anne M. Treisman and Garry Gelade. A feature-integration theory of attention. Cognitive Psychology, 12(1):97–136, January 1980.
  28. Summary -Zusammenfassung Dissertation PHD STUDIES 2008-2015, AG Neurophysik, FB Physik TITLE Measuring gaze and pupil in the real world: object-based attention, 3D tracking and applications SUBJECTS Visual attention, models, eye-tracking, pupillary response Conference talk 2011 J. Stoll, S. Kohlbecher, et al.: Mobile three dimensional gaze tracking. Medicine Meets Virtual Reality, Long Beach, CA, USA Conference abstracts 2013 J. Stoll, M. Sarey Khanie, et al.: Real-world tasks with full control over the visual scene: combining mobile gaze tracking and 4pi light-field measurements. Talk, 17th European Conference on Eye Movements, Lund, SE.
  29. T Wertheim. Über die indirekte sehschärfe. Zeitschrift für Psychologie & Physiolo- gie der Sinnesorgane, 7:172–187, 1894.
  30. Stoll, M. Sarey Khanie, et al.: Combining wearable eye-tracking with 4π light- field measurements: towards controlling all bottom-up and top-down factors dri- ving overt attention during real-world tasks. Poster, 10th Göttingen Meeting of the German Neuroscience Society, Göttingen, DE.
  31. Jean-Yves Bouguet. Complete camera calibration toolbox for matlab, 2004.
  32. Berufsgenossenschaft der Feinmechanik. Elektrotechnik: Expositionsgrenzwerte für künstliche optische strahlung. Berufsgenossenschaftliche Informationen für Sicher- heit und Gesundheit bei der Arbeit, 2004.
  33. B Winn, D Whitaker, D B Elliott, and N J Phillips. Factors affecting light-adapted pupil size in normal human subjects. Investigative Ophthalmology & Visual Science, 35(3):1132–7, 1994.
  34. P.R. Boyce. Human Factors in Lighting, Third Edition. Taylor & Francis, 2014.
  35. Bernard Marius 't Hart and Wolfgang Einhäuser. Mind the step: complementary effects of an implicit task on eye and head movements in real-life gaze allocation. Experimental brain research, 223(2):233–249, 2012. [39] Bernard Marius 't Hart, Johannes Vockeroth, Frank Schumann, Klaus Bartl, Erich Schneider, Peter Koenig, and Wolfgang Einhäuser. Gaze allocation in natural stimu- li: Comparing free exploration to head-fixed viewing conditions. Visual Cognition, 17(6-7):1132–1158, 2009.
  36. Ali Samii, John G Nutt, and Bruce R Ransom. Parkinson's disease. The Lancet, 363(9423):1783–1793, 2015/01/30 2004.
  37. Julie Brisson, Marc Mainville, Dominique Mailloux, Christelle Beaulieu, Josette Serres, and Sylvain Sirois. Pupil diameter measurement errors as a function of gaze direction in corneal reflection eyetrackers. Behavior research methods, 45(4):1322– 1331, 2013. [12] Claus Bundesen. A theory of visual attention. Psychological Review, 97(4):523– 547, October 1990.
  38. Vidhya Navalpakkam and Laurent Itti. Search goal tunes visual features optimally. Neuron, 53(4):605–617, 2007.
  39. S2. Giacino, J., Kalmar, K., and Whyte, J. (2004). The JFK Coma Recovery Scale-Revised: measurement characteristics and diagnostic utility. Arch. Phys. Med. Rehabil. 85, 2020-2029.
  40. Gustav Osterberg. Topography of the layer of rods and cones in the human retina. Nyt Nordisk Forlag, 1935.
  41. Does the FOUR score correctly diagnose the vegetative and minimally conscious states? Ann. Neurol. 60, 744-745.
  42. S4. Schneider, E., Villgrattner, T., Vockeroth, J., Bartl, K., Kohlbecher, S., Bardins, S., Ulbrich, H., and Brandt, T. (2009). EyeSeeCam: an eye movement-driven head camera for the examination of natural visual exploration. Ann. N. Y. Acad. Sci. 1164, 461-467.
  43. Jackson Beatty. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychological bulletin, 91(2):276, 1982.
  44. Neil A Harrison, C Ellie Wilson, and Hugo D Critchley. Processing of observed pupil size modulates perception of sadness and predicts empathy. Emotion, 7(4):724, 2007.
  45. Antonio Torralba, Aude Oliva, Monica S Castelhano, and John M Henderson. Con- textual guidance of eye movements and attention in real-world scenes: the role of global features in object search. Psychological review, 113(4):766, 2006.
  46. David Marr and Herbert Keith Nishihara. Representation and recognition of the spatial organization of three-dimensional shapes. Proceedings of the Royal Society of London B: Biological Sciences, 200(1140):269–294, 1978.
  47. Michael F Land. Eye movements and the control of actions in everyday life. Pro- gress in retinal and eye research, 25(3):296–324, 2006.
  48. Gary Aston-Jones and Jonathan D Cohen. An integrative theory of locus coeruleus- norepinephrine function: adaptive gain and optimal performance. Annu. Rev. Neu- rosci., 28:403–450, 2005.
  49. pROC: an open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics 12, 77.
  50. Laurent Itti, Christof Koch, and Ernst Niebuhr. A model of saliency-based visu- al attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11):1254–1259, November 1998.
  51. Jack B Kuipers. Quaternions and rotation sequences, volume 66. Princeton univer- sity press Princeton, 1999.
  52. Paul Viola and Michael Jones. Robust Real-time Object Detection. International Journal of Computer Vision -to appear, # 2002.
  53. Adam Herout, Jiri Havel, Lukas Polok, Michal Hradis, Pavel Zemcik, Radovan Josth, and Roman Juranek. Low-Level image features for Real-Time object detecti- on. INTECH Open Access Publisher, 2010.
  54. Brian J White and Douglas P Munoz. Separate visual signals for saccade initiation during target selection in the primate superior colliculus. The Journal of Neuros- cience, 31(5):1570–1578, 2011.
  55. Paola Binda, Maria Pereverzeva, and Scott O. Murray. Attention to bright surfaces enhances the pupillary light reflex. The Journal of Neuroscience, 33(5):2199–2204, 2013.
  56. Wolfgang Einhäuser, Merrielle Spain, and Pietro Perona. Objects predict fixations better than early saliency. Journal of Vision, 8(14):18, 2008.
  57. Wolfgang Einhäuser, Ueli Rutishauser, and Christof Koch. Task-demands can im- mediately reverse the effects of sensory-driven saliency in complex visual stimuli. Journal of Vision, 8(2):2, 2008.
  58. Alex D Hwang, Emily C Higgins, and Marc Pomplun. A model of top-down atten- tional control during visual search in complex scenes. Journal of Vision, 9(5):25, 2009.
  59. S7. DeLong, E.R., DeLong, D.M., and Clarke-Pearson, D.L. (1988). Comparing the Areas under Two or More Correlated Receiver Operating Characteristic Curves: A Nonparametric Approach.
  60. Michael F Land. Predictable eye-head coordination during driving. Nature, 1992.
  61. Susan J Sara. The locus coeruleus and noradrenergic modulation of cognition. Na- ture reviews neuroscience, 10(3):211–223, 2009.
  62. G. Westheimer. Relative localization in the human fovea: radial–tangential ani- sotropy. Proceedings of the Royal Society of London B: Biological Sciences, 268(1471):995–999, 2001.
  63. Antje Nuthmann and Wolfgang Einhäuser. A new approach to modeling the in- fluence of image features on fixation selection in scenes. Annals of the New York Academy of Sciences, In press, 2015.
  64. Giacomo Rizzolatti, Lucia Riggio, Isabella Dascola, and Carlo Umiltá. Reorien- ting attention across the horizontal and vertical meridians: evidence in favor of a premotor theory of attention. Neuropsychologia, 25(1):31–40, 1987.
  65. Vidhya Navalpakkam and Laurent Itti. Modeling the influence of task on attention. Vision research, 45(2):205–231, 2005.
  66. James S Maxwell and Clifton M Schor. The coordination of binocular eye move- ments: vertical and torsional alignment. Vision research, 46(21):3537–3548, 2006.
  67. Alex D Hwang, Hsueh-Cheng Wang, and Marc Pomplun. Semantic guidance of eye movements in real-world scenes. Vision research, 51(10):1192–1205, 2011.
  68. Marisa Carrasco. Visual attention: The past 25 years. Vision Research, 51(13):1484 – 1525, 2011. Vision Research 50th Anniversary Issue: Part 2.
  69. Helder Araújo, Rodrigo L Carceroni, and Christopher M Brown. A fully projective formulation to improve the accuracy of lowe's pose-estimation algorithm. Computer Vision and Image Understanding, 70(2):227–238, 1998.
  70. Eckhard H Hess and James M Polt. Pupil size in relation to mental activity during simple problem-solving. Science, 143(3611):1190–1192, 1964.
  71. Wolfgang Einhäuser, Frank Schumann, Stanislavs Bardins, Klaus Bartl, Guido Bö- ning, Erich Schneider, and Peter König. Human eye-head co-ordination in natural exploration. Network: Computation in Neural Systems, 18(3):267–297, 2007.
  72. Wolfgang Einhäuser, Christof Koch, and Olivia Carter. Pupil dilation betrays the timing of decisions. Frontiers in human neuroscience, 4:18, 2010.
  73. Bernard Marius 't Hart, Hannah Schmidt, Ingo Klein-Harmeyer, Christine Roth, and Wolfgang Einhäuser. The role of low-level features for rapid object detection and guidance of gaze in natural scenes. Journal of Vision, 12(9):807–807, 2012.


* Das Dokument ist im Internet frei zugänglich - Hinweise zu den Nutzungsrechten