eye position
- North America > United States > California > Santa Barbara County > Santa Barbara (0.14)
- North America > United States > Oregon > Lane County > Eugene (0.14)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.05)
- Asia > India (0.04)
- North America > United States > California > Santa Barbara County > Santa Barbara (0.14)
- North America > United States > Oregon > Lane County > Eugene (0.14)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.05)
- Asia > India (0.04)
Real-Time Sleepiness Detection for Driver State Monitoring System
Ghimire, Deepak, Jeong, Sunghwan, Yoon, Sunhong, Park, Sanghyun, Choi, Juhwan
Driver face monitoring system can detect driver fatigue, which is an important factor in a large number of accidents, using computer vision techniques. In this paper we present a real-time technique for driver eye state detection. At first face is detected and the eyes are searched inside face region for tracking. A normalized cross correlation based online dynamic template matching technique with combination of Kalman filter tracking is proposed to track the detected eye positions in the subsequent image frames. Support vector machine with histogram of orientation gradient features is used for classification of state of the eyes as open or closed. If the eye(s) state is detected as closed for a specified amount of time the driver is considered to be sleeping and an alarm will be generated.
- Asia > India (0.05)
- North America > United States > California > San Diego County > San Diego (0.05)
- North America > United States > Washington > King County > Seattle (0.04)
- (5 more...)
Spatial Representations in the Parietal Cortex May Use Basis Functions
The parietal cortex is thought to represent the egocentric posi(cid:173) tions of objects in particular coordinate systems. We propose an alternative approach to spatial perception of objects in the pari(cid:173) etal cortex from the perspective of sensorimotor transformations. The responses of single parietal neurons can be modeled as a gaus(cid:173) sian function of retinal position multiplied by a sigmoid function of eye position, which form a set of basis functions. We show here how these basis functions can be used to generate receptive fields in either retinotopic or head-centered coordinates by simple linear transformations. This raises the possibility that the parietal cortex does not attempt to compute the positions of objects in a partic(cid:173) ular frame of reference but instead computes a general purpose representation of the retinal location and eye position from which any transformation can be synthesized by direct projection.
Modeling Human Eye Movements with Neural Networks in a Maze-Solving Task
Li, Jason, Watters, Nicholas, Yingting, null, Wang, null, Sohn, Hansem, Jazayeri, Mehrdad
From smoothly pursuing moving objects to rapidly shifting gazes during visual search, humans employ a wide variety of eye movement strategies in different contexts. While eye movements provide a rich window into mental processes, building generative models of eye movements is notoriously difficult, and to date the computational objectives guiding eye movements remain largely a mystery. In this work, we tackled these problems in the context of a canonical spatial planning task, maze-solving. We collected eye movement data from human subjects and built deep generative models of eye movements using a novel differentiable architecture for gaze fixations and gaze shifts. We found that human eye movements are best predicted by a model that is optimized not to perform the task as efficiently as possible but instead to run an internal simulation of an object traversing the maze. This not only provides a generative model of eye movements in this task but also suggests a computational theory for how humans solve the task, namely that humans use mental simulation.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- Europe > Netherlands > Drenthe > Assen (0.04)
Eye Tracking Has The Potential To Completely Change AR And VR
Note: Moor Insights & Strategy Analyst Anshel Sag made significant contributions to this blog. AR and VR are still in the growth phase of development with billions of R&D and content money being poured in annually. A lot of that money is being invested in technologies to improve the experience as the current experience is, well, rough. There is a multitude of technologies that can help the immersive technology industry to move forward with AR and VR. Some of those technologies have to do with the display or wireless communications to make the experience better than it is today.
An Object-Based Bayesian Framework for Top-Down Visual Attention
Borji, Ali (University of Southern California) | Sihite, Dicky N. (University of Southern California) | Itti, Laurent (University of Southern California)
We introduce a new task-independent framework to model top-down overt visual attention based on graph-ical models for probabilistic inference and reasoning. We describe a Dynamic Bayesian Network (DBN) that infers probability distributions over attended objects and spatial locations directly from observed data. Probabilistic inference in our model is performed over object-related functions which are fed from manual annotations of objects in video scenes or by state-of-the-art object detection models. Evaluating over ∼3 hours (appx. 315,000 eye fixations and 12,600 saccades) of observers playing 3 video games (time-scheduling, driving, and flight combat), we show that our approach is significantly more predictive of eye fixations compared to: 1) simpler classifier-based models also developed here that map a signature of a scene (multi-modal information from gist, bottom-up saliency, physical actions, and events) to eye positions, 2) 14 state-of-the-art bottom-up saliency models, and 3) brute-force algorithms such as mean eye position. Our results show that the proposed model is more effective in employing and reasoning over spatio-temporal visual data.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Europe > Spain (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Government (0.93)
- Leisure & Entertainment > Games > Computer Games (0.34)
Congruence between model and human attention reveals unique signatures of critical visual events
Current computational models of bottom-up and top-down components of attention are predictive of eye movements across a range of stimuli and of simple, fixed visual tasks (such as visual search for a target among distractors). However, to date there exists no computational framework which can reliably mimic human gaze behavior in more complex environments and tasks, such as driving a vehicle through traffic. Here, we develop a hybrid computational/behavioral framework, combining simple models for bottom-up salience and top-down relevance, and looking for changes in the predictive power of these components at different critical event times during 4.7 hours (500,000 video frames) of observers playing car racing and flight combat video games. This approach is motivated by our observation that the predictive strengths of the salience and relevance models exhibit reliable temporal signatures during critical event windows in the task sequence--for example, when the game player directly engages an enemy plane in a flight combat game, the predictive strength of the salience model increases significantly, while that of the relevance model decreases significantly. Our new framework combines these temporal signatures to implement several event detectors. Critically, we find that an event detector based on fused behavioral and stimulus information (in the form of the model's predictive strength) is much stronger than detectors based on behavioral information alone (eye position) or image information alone (model prediction maps). This approach to event detection, based on eye tracking combined with computational models applied to the visual input, may have useful applications as a less-invasive alternative to other event detection approaches based on neural signatures derived from EEG or fMRI recordings.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
Congruence between model and human attention reveals unique signatures of critical visual events
Current computational models of bottom-up and top-down components of attention are predictive of eye movements across a range of stimuli and of simple, fixed visual tasks (such as visual search for a target among distractors). However, to date there exists no computational framework which can reliably mimic human gaze behavior in more complex environments and tasks, such as driving a vehicle through traffic. Here, we develop a hybrid computational/behavioral framework, combining simple models for bottom-up salience and top-down relevance, and looking for changes in the predictive power of these components at different critical event times during 4.7 hours (500,000 video frames) of observers playing car racing and flight combat video games. This approach is motivated by our observation that the predictive strengths of the salience and relevance models exhibit reliable temporal signatures during critical event windows in the task sequence--for example, when the game player directly engages an enemy plane in a flight combat game, the predictive strength of the salience model increases significantly, while that of the relevance model decreases significantly. Our new framework combines these temporal signatures to implement several event detectors. Critically, we find that an event detector based on fused behavioral and stimulus information (in the form of the model's predictive strength) is much stronger than detectors based on behavioral information alone (eye position) or image information alone (model prediction maps). This approach to event detection, based on eye tracking combined with computational models applied to the visual input, may have useful applications as a less-invasive alternative to other event detection approaches based on neural signatures derived from EEG or fMRI recordings.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
Congruence between model and human attention reveals unique signatures of critical visual events
Current computational models of bottom-up and top-down components of attention arepredictive of eye movements across a range of stimuli and of simple, fixed visual tasks (such as visual search for a target among distractors). However, todate there exists no computational framework which can reliably mimic human gaze behavior in more complex environments and tasks, such as driving a vehicle through traffic. Here, we develop a hybrid computational/behavioral framework, combining simple models for bottom-up salience and top-down relevance, andlooking for changes in the predictive power of these components at different critical event times during 4.7 hours (500,000 video frames) of observers playing car racing and flight combat video games. This approach is motivated by our observation that the predictive strengths of the salience and relevance models exhibitreliable temporal signatures during critical event windows in the task sequence--for example, when the game player directly engages an enemy plane in a flight combat game, the predictive strength of the salience model increases significantly, while that of the relevance model decreases significantly. Our new framework combines these temporal signatures to implement several event detectors. Critically,we find that an event detector based on fused behavioral and stimulus information (in the form of the model's predictive strength) is much stronger than detectors based on behavioral information alone (eye position) or image information alone(model prediction maps). This approach to event detection, based on eye tracking combined with computational models applied to the visual input, may have useful applications as a less-invasive alternative to other event detection approaches based on neural signatures derived from EEG or fMRI recordings.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)