TY - JOUR
T1 - The Perceptual Proxies of Visual Comparison
AU - Jardine, Nicole
AU - Ondov, Brian D.
AU - Elmqvist, Niklas
AU - Franconeri, Steven
N1 - Funding Information:
Nicole Jardine was supported by the U.S. National Science Foundation grant number DRL-1661264 while affiliated with Northwestern University. Brian Ondov was supported by the Intramural Research Program of the National Human Genome Research Institute, a part of the U.S. National Institutes of Health. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the respective funding agencies.
Publisher Copyright:
© 2019 IEEE.
PY - 2020/1
Y1 - 2020/1
N2 - Perceptual tasks in visualizations often involve comparisons. Of two sets of values depicted in two charts, which set had values that were the highest overall? Which had the widest range? Prior empirical work found that the performance on different visual comparison tasks (e.g., 'biggest delta', 'biggest correlation') varied widely across different combinations of marks and spatial arrangements. In this paper, we expand upon these combinations in an empirical evaluation of two new comparison tasks: the 'biggest mean' and 'biggest range' between two sets of values. We used a staircase procedure to titrate the difficulty of the data comparison to assess which arrangements produced the most precise comparisons for each task. We find visual comparisons of biggest mean and biggest range are supported by some chart arrangements more than others, and that this pattern is substantially different from the pattern for other tasks. To synthesize these dissonant findings, we argue that we must understand which features of a visualization are actually used by the human visual system to solve a given task. We call these perceptual proxies. For example, when comparing the means of two bar charts, the visual system might use a 'Mean length' proxy that isolates the actual lengths of the bars and then constructs a true average across these lengths. Alternatively, it might use a 'Hull Area' proxy that perceives an implied hull bounded by the bars of each chart and then compares the areas of these hulls. We propose a series of potential proxies across different tasks, marks, and spatial arrangements. Simple models of these proxies can be empirically evaluated for their explanatory power by matching their performance to human performance across these marks, arrangements, and tasks. We use this process to highlight candidates for perceptual proxies that might scale more broadly to explain performance in visual comparison.
AB - Perceptual tasks in visualizations often involve comparisons. Of two sets of values depicted in two charts, which set had values that were the highest overall? Which had the widest range? Prior empirical work found that the performance on different visual comparison tasks (e.g., 'biggest delta', 'biggest correlation') varied widely across different combinations of marks and spatial arrangements. In this paper, we expand upon these combinations in an empirical evaluation of two new comparison tasks: the 'biggest mean' and 'biggest range' between two sets of values. We used a staircase procedure to titrate the difficulty of the data comparison to assess which arrangements produced the most precise comparisons for each task. We find visual comparisons of biggest mean and biggest range are supported by some chart arrangements more than others, and that this pattern is substantially different from the pattern for other tasks. To synthesize these dissonant findings, we argue that we must understand which features of a visualization are actually used by the human visual system to solve a given task. We call these perceptual proxies. For example, when comparing the means of two bar charts, the visual system might use a 'Mean length' proxy that isolates the actual lengths of the bars and then constructs a true average across these lengths. Alternatively, it might use a 'Hull Area' proxy that perceives an implied hull bounded by the bars of each chart and then compares the areas of these hulls. We propose a series of potential proxies across different tasks, marks, and spatial arrangements. Simple models of these proxies can be empirically evaluated for their explanatory power by matching their performance to human performance across these marks, arrangements, and tasks. We use this process to highlight candidates for perceptual proxies that might scale more broadly to explain performance in visual comparison.
KW - Graphical perception
KW - crowdsourced evaluation
KW - visual comparison
KW - visual perception
UR - http://www.scopus.com/inward/record.url?scp=85075600582&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85075600582&partnerID=8YFLogxK
U2 - 10.1109/TVCG.2019.2934786
DO - 10.1109/TVCG.2019.2934786
M3 - Article
C2 - 31443016
AN - SCOPUS:85075600582
SN - 1077-2626
VL - 26
SP - 1012
EP - 1021
JO - IEEE Transactions on Visualization and Computer Graphics
JF - IEEE Transactions on Visualization and Computer Graphics
IS - 1
M1 - 8807320
ER -