In collaboration with GAIDG Lab external collaborators, York University PI Petros Faloutsos, PhD Student Melissa Kremer, and Master’s student Peter Caruana, and Rutgers University PI Mubbasir Kapadia, we develop a parametric pseudo-saliency map approach for speeding up vision-based cognition in crowd agent modelling PSM: Parametric Saliency Maps for Autonomous Pedestrians.
Abstract
Modeling visual attention is an important aspect of simulating realistic virtual humans. This work proposes a parametric model and method for generating real-time saliency maps from the perspective of virtual agents which approximate those of vision-based saliency approaches. The model aggregates a saliency score from user-defined parameters for objects and characters in an agent’s view and uses that to output a 2D saliency map which can be modulated by an attention field to incorporate 3D information as well as a character’s state of attentiveness. The aggregate and parameterized structure of the method allows the user to model a range of diverse agents. The user may also expand the model with additional layers and parameters. The proposed method can be combined with normative and pathological models of the human visual field and gaze controllers, such as the recently proposed model of egocentric distractions for casual pedestrians that we use in our results.