Title
PSM: Parametric Saliency Maps for Autonomous Pedestrians
Abstract
ABSTRACTModeling visual attention is an important aspect of simulating realistic virtual humans. This work proposes a parametric model and method for generating real-time saliency maps from the perspective of virtual agents which approximate those of vision-based saliency approaches. The model aggregates a saliency score from user-defined parameters for objects and characters in an agent’s view and uses that to output a 2D saliency map which can be modulated by an attention field to incorporate 3D information as well as a character’s state of attentiveness. The aggregate and parameterized structure of the method allows the user to model a range of diverse agents. The user may also expand the model with additional layers and parameters. The proposed method can be combined with normative and pathological models of the human visual field and gaze controllers, such as the recently proposed model of egocentric distractions for casual pedestrians that we use in our results.
Year
DOI
Venue
2021
10.1145/3487983.3488299
MIG
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
5
Name
Order
Citations
PageRank
Melissa Kremer100.68
Peter Caruana200.34
M. Brandon Haworth3217.74
Mubbasir Kapadia454658.07
Petros Faloutsos53082422.32