Abstract | ||
---|---|---|
Eye movements provide insight into what parts of an image a viewer finds most salient, interesting, or relevant to the task at hand. Unfortunately, eye tracking data, a commonly-used proxy for attention, is cumbersome to collect. Here we explore an alternative: a comprehensive web-based toolbox for crowdsourcing visual attention. We draw from four main classes of attention-capturing methodologies in the literature. ZoomMaps is a novel zoom-based interface that captures viewing on a mobile phone. CodeCharts is a self-reporting methodology that records points of interest at precise viewing durations. ImportAnnots is an "annotation" tool for selecting important image regions, and cursor-based BubbleView lets viewers click to deblur a small area. We compare these methodologies using a common analysis framework in order to develop appropriate use cases for each interface. This toolbox and our analyses provide a blueprint for how to gather attention data at scale without an eye tracker.
|
Year | DOI | Venue |
---|---|---|
2020 | 10.1145/3313831.3376799 | CHI '20: CHI Conference on Human Factors in Computing Systems
Honolulu
HI
USA
April, 2020 |
DocType | ISBN | Citations |
Conference | 978-1-4503-6708-0 | 2 |
PageRank | References | Authors |
0.37 | 0 | 8 |
Name | Order | Citations | PageRank |
---|---|---|---|
Newman Anelise | 1 | 2 | 0.70 |
Barry A. McNamara | 2 | 2 | 0.70 |
Fosco Camilo | 3 | 2 | 0.70 |
Zhang Yun Bin | 4 | 2 | 0.37 |
Sukhum Pat | 5 | 2 | 0.70 |
Matthew Tancik | 6 | 14 | 4.75 |
Namwook Kim | 7 | 179 | 12.31 |
Zoya Gavrilov | 8 | 287 | 16.20 |