Skip to content

IEEE VIS 2017: Perception, Evaluation, Vision Science

Questions of how visualization is read and understood are a common topic in research, and this year was no exception. This part of my IEEE VIS 2017 report covers papers dealing with perceptual issues, the match between visualization and tasks, statistical reasoning, as well as a panel on vision science.


Data Visualization Saliency Model: A Tool for Evaluating Abstract Data Visualizations by Laura E. Matzen, Michael J. Haass, Kristin M. Divis, Zhiyuan Wang, and Andrew T. Wilson tries to predict what people will look at in a visualization. Saliency models are usually built for photographs of natural scenes, and it’s not clear if and how they apply to visualization. The authors of this paper adapted an existing model to take into consideration factors like text, color, and the different kinds of spatial scales you see in a visualization. Matlab code for their model is available, though under a somewhat odd license (especially considering that this is U.S. government work).

Priming and Anchoring Effects in Visualization by André Calero Valdez, Martina Ziefle, and Michael Sedlmair looks at how priming might affect people’s understanding of visualization. Priming is when what you see first influences how you perceive what comes next. In a series of studies, the authors showed that these effects also exist in the particular case of scatterplots that show separable or non-separable clusters. Seeing one kind of plot first changes the likelihood of you judging a subsequent plot as the same or another type.

Evaluating Interactive Graphical Encodings for Data Visualization by Bahador Saket, Arjun Srinivasan, Eric D. Ragan, and Alex Endert extends the common studies of visual encodings to include the effect of interaction. How precisely can use scale a bar, bubble, etc. to fit a particular value? They find that the two are closely linked, which is not surprising, but it’s useful to know. You can read their Medium posting with more explanation, look at their Materials and source code, or read their paper.


Scatterplots: Tasks, Data, and Designs by Alper Sarikaya and Michael Gleicher looks at the design space of scatterplots and a number of variations on them (like hexagonal bins and splatterplots), and what kinds of tasks they are each useful for.

Considerations for Visualizing Comparison by Michael Gleicher asks for questions about comparisons: what comparison is being made, what makes it hard, what can be done to address the challenges, and what design choices are available. The result is a design process for building visualizations that support comparison.

Bridging From Goals to Tasks with Design Study Analysis Reports by Heidi Lam, Melanie Tory, and Tamara Munzner asks how you can go from a stated question to a visualization task. That translation is not nearly as obvious as you’d assume, but they break it down and show different ways of getting from one to the other. Their collection of materials is also available.


Imagining Replications: Graphical Prediction & Discrete Visualizations Improve Recall & Estimation of Effect Uncertainty by Jessica Hullman, Matthew Kay, Yea-Seul Kim, and Samana Shrestha went over my head a bit, being the first presentation Friday morning. They look at statistical reasoning, in particular in replication studies, which is generally pretty difficult. Using a number of variations like discretized probability distributions, they were able to help them make better guesses what replications with different outcomes meant for the original studies.

Evaluating Cartogram Effectiveness by Sabrina Nusrat, Muhammad Jawaherul Alam, and Stephen Kobourov looked at a number of variations on cartograms and the kinds of tasks that are sometimes difficult to do with them, like telling which elements are adjacent. It would have helped to have something else in there as a reference, IMHO, like a bar chart or non-cartogram bubbles.

Panel: Vision Science Meets Visualization

This panel by Christine Nothelfer, Zoya Bylinskii, Madison Elliott, Cindy Xiong, Danielle Albers Szafir, Ronald Rensink, Todd Horowitz, Steven Franconeri, Karen Schloss, and Ruth Rosenholtz covered a lot of ground, much more than I can write up here (plus I also missed almost half of it). It was really interesting to get an overview of current topics in vision research, and also to get some myths busted about things like peripheral vision (which is much better than we often assume).

Two things I want to point to in particular are a review paper, Four Types of Ensemble Coding in Data Visualizations, which gives a good introduction to ensemble coding, and the InfoVis x Vision Science website. You can sign up there to get news about relevant research and get invited to a Slack channel to discuss research related to vision science and visualization.

See also:

Posted by Robert Kosara on October 11, 2017. Filed under conference, IEEEVIS.