Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 32
Filter
1.
IEEE Trans Vis Comput Graph ; 30(1): 584-594, 2024 Jan.
Article in English | MEDLINE | ID: mdl-38096099

ABSTRACT

This study presents insights from interviews with nineteen Knowledge Graph (KG) practitioners who work in both enterprise and academic settings on a wide variety of use cases. Through this study, we identify critical challenges experienced by KG practitioners when creating, exploring, and analyzing KGs that could be alleviated through visualization design. Our findings reveal three major personas among KG practitioners - KG Builders, Analysts, and Consumers - each of whom have their own distinct expertise and needs. We discover that KG Builders would benefit from schema enforcers, while KG Analysts need customizable query builders that provide interim query results. For KG Consumers, we identify a lack of efficacy for node-link diagrams, and the need for tailored domain-specific visualizations to promote KG adoption and comprehension. Lastly, we find that implementing KGs effectively in practice requires both technical and social solutions that are not addressed with current tools, technologies, and collaborative workflows. From the analysis of our interviews, we distill several visualization research directions to improve KG usability, including knowledge cards that balance digestibility and discoverability, timeline views to track temporal changes, interfaces that support organic discovery, and semantic explanations for AI and machine learning predictions.

2.
Article in English | MEDLINE | ID: mdl-37983146

ABSTRACT

Data integration is often performed to consolidate information from multiple disparate data sources during visual data analysis. However, integration operations are usually separate from visual analytics operations such as encode and filter in both interface design and empirical research. We conducted a preliminary user study to investigate whether and how data integration should be incorporated directly into the visual analytics process. We used two interface alternatives featuring contrasting approaches to the data preparation and analysis workflow: manual file-based ex-situ integration as a separate step from visual analytics operations; and automatic UI-based in-situ integration merged with visual analytics operations. Participants were asked to complete specific and free-form tasks with each interface, browsing for patterns, generating insights, and summarizing relationships between attributes distributed across multiple files. Analyzing participants' interactions and feedback, we found both task completion time and total interactions to be similar across interfaces and tasks, as well as unique integration strategies between interfaces and emergent behaviors related to satisficing and cognitive bias. Participants' time spent and interactions emergent strategies revealed that in-situ integration enabled users to spend more time on analysis tasks compared with ex-situ integration. Participants' integration strategies and analytical behaviors revealed differences in interface usage for generating and tracking hypotheses and insights , yet their emergent behaviors suggested that in-situ integration could negatively affect the ability to generate and track hypotheses and insights. With these results, we synthesized preliminary guidelines for designing future visual analytics interfaces that can support integrating attributes throughout an active analysis process.

3.
IEEE Comput Graph Appl ; 43(4): 111-120, 2023.
Article in English | MEDLINE | ID: mdl-37432777

ABSTRACT

Visualization researchers and visualization professionals seek appropriate abstractions of visualization requirements that permit considering visualization solutions independently from specific problems. Abstractions can help us design, analyze, organize, and evaluate the things we create. The literature has many task structures (taxonomies, typologies, etc.), design spaces, and related "frameworks" that provide abstractions of the problems a visualization is meant to address. In this Visualization Viewpoints article, we introduce a different one, a problem space that complements existing frameworks by focusing on the needs that a visualization is meant to solve. We believe it provides a valuable conceptual tool for designing and discussing visualizations.

4.
Appetite ; 188: 106610, 2023 09 01.
Article in English | MEDLINE | ID: mdl-37269883

ABSTRACT

Food purchase choices, one of the main determinants of food consumption, is highly influenced by food environments. Given the surge in online grocery shopping because of the COVID-19 pandemic, interventions in digital environments present more than ever an opportunity to improve the nutritional quality of food purchase choices. One such opportunity can be found in gamification. Participants (n = 1228) shopped for 12 items from a shopping list on a simulated online grocery platform. We randomized them into four groups in a 2 × 2 factorial design: presence vs. absence of gamification, and high vs. low budget. Participants in the gamification groups saw foods with 1 (least nutritious) to 5 (most nutritious) crown icons and a scoreboard with a tally of the number of crowns the participant collected. We estimated ordinary least squares and Poisson regression models to test the impact of the gamification and budget on the nutritional quality of the shopping basket. In the absence of gamification and low budget, participants collected 30.78 (95% CI [30.27; 31.29]) crowns. In the gamification and low budget condition, participants increased the nutritional quality of their shopping basket by collecting more crowns (B = 4.15, 95% CI [3.55; 4.75], p < 0.001). The budget amount ($50 vs. $30) did not alter the final shopping basket (B = 0.45, 95% CI [-0.02; 1.18], p = 0.057), nor moderated the gamification effect. Gamification increased the nutritional quality of the final shopping baskets and nine of 12 shopping list items in this hypothetical experiment. Gamifying nutrition labels may be an effective strategy to improve the nutritional quality of food choices in online grocery stores, but further research is needed.


Subject(s)
COVID-19 , Food Preferences , Humans , Consumer Behavior , Gamification , Nutritional Status , Pandemics
5.
Article in English | MEDLINE | ID: mdl-37030764

ABSTRACT

Presenting a predictive model's performance is a communication bottleneck that threatens collaborations between data scientists and subject matter experts. Accuracy and error metrics alone fail to tell the whole story of a model - its risks, strengths, and limitations - making it difficult for subject matter experts to feel confident in their decision to use a model. As a result, models may fail in unexpected ways or go entirely unused, as subject matter experts disregard poorly presented models in favor of familiar, yet arguably substandard methods. In this paper, we describe an iterative study conducted with both subject matter experts and data scientists to understand the gaps in communication between these two groups. We find that, while the two groups share common goals of understanding the data and predictions of the model, friction can stem from unfamiliar terms, metrics, and visualizations - limiting the transfer of knowledge to SMEs and discouraging clarifying questions being asked during presentations. Based on our findings, we derive a set of communication guidelines that use visualization as a common medium for communicating the strengths and weaknesses of a model. We provide a demonstration of our guidelines in a regression modeling scenario and elicit feedback on their use from subject matter experts. From our demonstration, subject matter experts were more comfortable discussing a model's performance, more aware of the trade-offs for the presented model, and better equipped to assess the model's risks - ultimately informing and contextualizing the model's use beyond text and numbers.

6.
Article in English | MEDLINE | ID: mdl-37022242

ABSTRACT

Multiple-view (MV) visualizations have become ubiquitous for visual communication and exploratory data visualization. However, most existing MV visualizations are designed for the desktop, which can be unsuitable for the continuously evolving displays of varying screen sizes. In this paper, we present a two-stage adaptation framework that supports the automated retargeting and semi-automated tailoring of a desktop MV visualization for rendering on devices with displays of varying sizes. First, we cast layout retargeting as an optimization problem and propose a simulated annealing technique that can automatically preserve the layout of multiple views. Second, we enable fine-tuning for the visual appearance of each view, using a rule-based auto configuration method complemented with an interactive interface for chart-oriented encoding adjustment. To demonstrate the feasibility and expressivity of our proposed approach, we present a gallery of MV visualizations that have been adapted from the desktop to small displays. We also report the result of a user study comparing visualizations generated using our approach with those by existing methods. The outcome indicates that the participants generally prefer visualizations generated using our approach and find them to be easier to use.

7.
IEEE Trans Vis Comput Graph ; 29(2): 1559-1572, 2023 Feb.
Article in English | MEDLINE | ID: mdl-34748493

ABSTRACT

Projection techniques are often used to visualize high-dimensional data, allowing users to better understand the overall structure of multi-dimensional spaces on a 2D screen. Although many such methods exist, comparably little work has been done on generalizable methods of inverse-projection - the process of mapping the projected points, or more generally, the projection space back to the original high-dimensional space. In this article we present NNInv, a deep learning technique with the ability to approximate the inverse of any projection or mapping. NNInv learns to reconstruct high-dimensional data from any arbitrary point on a 2D projection space, giving users the ability to interact with the learned high-dimensional representation in a visual analytics system. We provide an analysis of the parameter space of NNInv, and offer guidance in selecting these parameters. We extend validation of the effectiveness of NNInv through a series of quantitative and qualitative analyses. We then demonstrate the method's utility by applying it to three visualization tasks: interactive instance interpolation, classifier agreement, and gradient visualization.

8.
IEEE Trans Vis Comput Graph ; 28(9): 3093-3112, 2022 09.
Article in English | MEDLINE | ID: mdl-33434132

ABSTRACT

Progressive visualization is fast becoming a technique in the visualization community to help users interact with large amounts of data. With progressive visualization, users can examine intermediate results of complex or long running computations, without waiting for the computation to complete. While this has shown to be beneficial to users, recent research has identified potential risks. For example, users may misjudge the uncertainty in the intermediate results and draw incorrect conclusions or see patterns that are not present in the final results. In this article, we conduct a comprehensive set of studies to quantify the advantages and limitations of progressive visualization. Based on a recent report by Micallef et al., we examine four types of cognitive biases that can occur with progressive visualization: uncertainty bias, illusion bias, control bias, and anchoring bias. The results of the studies suggest a cautious but promising use of progressive visualization - while there can be significant savings in task completion time, accuracy can be negatively affected in certain conditions. These findings confirm earlier reports of the benefits and drawbacks of progressive visualization and that continued research into mitigating the effects of cognitive biases is necessary.


Subject(s)
Cognition , Computer Graphics , Bias , Uncertainty
9.
IEEE Trans Vis Comput Graph ; 28(1): 737-746, 2022 01.
Article in English | MEDLINE | ID: mdl-34587039

ABSTRACT

Interactive visualization design and research have primarily focused on local data and synchronous events. However, for more complex use cases-e.g., remote database access and streaming data sources-developers must grapple with distributed data and asynchronous events. Currently, constructing these use cases is difficult and time-consuming; developers are forced to operationally program low-level details like asynchronous database querying and reactive event handling. This approach is in stark contrast to modern methods for browser-based interactive visualization, which feature high-level declarative specifications. In response, we present DIEL, a declarative framework that supports asynchronous events over distributed data. As in many declarative languages, DIEL developers specify only what data they want, rather than procedural steps for how to assemble it. Uniquely, DIEL models asynchronous events (e.g., user interactions, server responses) as streams of data that are captured in event logs. To specify the state of a visualization at any time, developers write declarative queries over the data and event logs; DIEL compiles and optimizes a corresponding dataflow graph, and automatically generates necessary low-level distributed systems details. We demonstrate DIEL'S performance and expressivity through example interactive visualizations that make diverse use of remote data and asynchronous events. We further evaluate DIEL'S usability using the Cognitive Dimensions of Notations framework, revealing wins such as ease of change, and compromises such as premature commitments.

10.
J Chem Inf Model ; 61(11): 5524-5534, 2021 11 22.
Article in English | MEDLINE | ID: mdl-34752100

ABSTRACT

Photoswitches are molecules that undergo a reversible, structural isomerization after exposure to certain wavelengths of light. The dynamic control offered by molecular photoswitches is favorable for materials chemistry, photopharmacology, and catalysis applications. Ideal photoswitches absorb visible light and have long-lived metastable isomers. We used high-throughput virtual screening to predict the absorption maxima (λmax) of the E-isomer and half-life (t1/2) of the Z-isomer. However, computing the photophysical and kinetic stabilities with density functional theory of each entry of a virtual molecular library containing thousands or millions of molecules is prohibitively time-consuming. We applied active search, a machine-learning technique, to intelligently search a chemical search space of 255 991 photoswitches based on 29 known azoarenes and their derivatives. We iteratively trained the active search algorithm on whether a candidate absorbed visible light (λmax > 450 nm). Active search was found to triple the discovery rate compared to random search. Further, we projected 1962 photoswitches to 2D using the Uniform Manifold Approximation and Projection algorithm and found that λmax depends on the core, which is tunable by substituents. We then incorporated a second stage of screening to predict the stabilities of the Z-isomers for the top candidates of each core. We identified four ideal photoswitches that concurrently satisfy the following criteria: λmax > 450 nm and t1/2 > 2 h.These candidates had λmax and t1/2 range from 465 to 531 nm and hours to days, respectively.


Subject(s)
Light , Catalysis , Half-Life , Isomerism
11.
IEEE Trans Vis Comput Graph ; 27(2): 1514-1524, 2021 Feb.
Article in English | MEDLINE | ID: mdl-33048683

ABSTRACT

Multiple-view visualization (MV) is a layout design technique often employed to help users see a large number of data attributes and values in a single cohesive representation. Because of its generalizability, the MV design has been widely adopted by the visualization community to help users examine and interact with large, complex, and high-dimensional data. However, although ubiquitous, there has been little work to categorize and analyze MVs in order to better understand its design space. As a result, there has been little to no guideline in how to use the MV design effectively. In this paper, we present an in-depth study of how MVs are designed in practice. We focus on two fundamental measures of multiple-view patterns: composition, which quantifies what view types and how many are there; and configuration, which characterizes spatial arrangement of view layouts in the display space. We build a new dataset containing 360 images of MVs collected from IEEE VIS, EuroVis, and PacificVis publications 2011 to 2019, and make fine-grained annotations of view types and layouts for these visualization images. From this data we conduct composition and configuration analyses using quantitative metrics of term frequency and layout topology. We identify common practices around MVs, including relationship of view types, popular view layouts, and correlation between view types and layouts. We combine the findings into a MV recommendation system, providing interactive tools to explore the design space, and support example-based design.

12.
IEEE Trans Vis Comput Graph ; 27(2): 401-411, 2021 Feb.
Article in English | MEDLINE | ID: mdl-33048700

ABSTRACT

Static scatterplots often suffer from the overdraw problem on big datasets where object overlap causes undesirable visual clutter. The use of zooming in scatterplots can help alleviate this problem. With multiple zoom levels, more screen real estate is available, allowing objects to be placed in a less crowded way. We call this type of visualization scalable scatterplot visualizations, or SSV for short. Despite the potential of SSVs, existing systems and toolkits fall short in supporting the authoring of SSVs due to three limitations. First, many systems have limited scalability, assuming that data fits in the memory of one computer. Second, too much developer work, e.g., using custom code to generate mark layouts or render objects, is required. Third, many systems focus on only a small subset of the SSV design space (e.g. supporting a specific type of visual marks). To address these limitations, we have developed Kyrix-S, a system for easy authoring of SSVs at scale. Kyrix-S derives a declarative grammar that enables specification of a variety of SSVs in a few tens of lines of code, based on an existing survey of scatterplot tasks and designs. The declarative grammar is supported by a distributed layout algorithm which automatically places visual marks onto zoom levels. We store data in a multi-node database and use multi-node spatial indexes to achieve interactive browsing of large SSVs. Extensive experiments show that 1) Kyrix-S enables interactive browsing of SSVs of billions of objects, with response times under 500ms and 2) Kyrix-S achieves 4X-9X reduction in specification compared to a state-of-the-art authoring system.

13.
IEEE Trans Vis Comput Graph ; 27(2): 1731-1741, 2021 02.
Article in English | MEDLINE | ID: mdl-33048737

ABSTRACT

Most visual analytics systems assume that all foraging for data happens before the analytics process; once analysis begins, the set of data attributes considered is fixed. Such separation of data construction from analysis precludes iteration that can enable foraging informed by the needs that arise in-situ during the analysis. The separation of the foraging loop from the data analysis tasks can limit the pace and scope of analysis. In this paper, we present CAVA, a system that integrates data curation and data augmentation with the traditional data exploration and analysis tasks, enabling information foraging in-situ during analysis. Identifying attributes to add to the dataset is difficult because it requires human knowledge to determine which available attributes will be helpful for the ensuing analytical tasks. CAVA crawls knowledge graphs to provide users with a a broad set of attributes drawn from external data to choose from. Users can then specify complex operations on knowledge graphs to construct additional attributes. CAVA shows how visual analytics can help users forage for attributes by letting users visually explore the set of available data, and by serving as an interface for query construction. It also provides visualizations of the knowledge graph itself to help users understand complex joins such as multi-hop aggregations. We assess the ability of our system to enable users to perform complex data combinations without programming in a user study over two datasets. We then demonstrate the generalizability of CAVA through two additional usage scenarios. The results of the evaluation confirm that CAVA is effective in helping the user perform data foraging that leads to improved analysis outcomes, and offer evidence in support of integrating data augmentation as a part of the visual analytics pipeline.

14.
Public Health Nutr ; 23(15): 2717-2727, 2020 10.
Article in English | MEDLINE | ID: mdl-32713393

ABSTRACT

OBJECTIVE: To describe characteristics of self-identified popular diet followers and compare mean BMI across these diets, stratified by time following diet. DESIGN: Cross-sectional, web-based survey administered in 2015. SETTING: Non-localised, international survey. PARTICIPANTS: Self-selected followers of popular diets (n 9019) were recruited to the survey via social media and email announcements by diet community leaders, categorised into eight major diet groups. RESULTS: General linear models were used to compare mean BMI among (1) short-term (<1 year) and long-term (≥1 year) followers within diet groups and (2) those identifying as 'try to eat healthy' (TTEH) to all other diet groups, stratified by time following the specific diet. Participants were 82 % female, 93 % White and 96 % non-Hispanic. Geometric mean BMI was lower (P < 0·05 for all) among longer-term followers (≥1 year) of whole food, plant-based (WFPB), vegan, whole food and low-carb diets compared with shorter-term followers. Among those following their diet for 1-5 years (n 4067), geometric mean BMI (kg/m2) were lower (P < 0·05 for all) for all groups compared with TTEH (26·4 kg/m2): WFPB (23·2 kg/m2), vegan (23·5 kg/m2), Paleo (24·6 kg/m2), vegetarian (25·0 kg/m2), whole food (24·6 kg/m2), Weston A. Price (23·5 kg/m2) and low-carb (24·7 kg/m2). CONCLUSION: Our findings suggest that BMI is lower among individuals who made active decisions to adhere to a specific diet, particularly more plant-based diets and/or diets limiting highly processed foods, compared with those who simply TTEH. BMI is also lower among individuals who follow intentional eating plans for longer time periods.


Subject(s)
Body Weight , Diabetes Mellitus, Type 2 , Diet , Adolescent , Adult , Body Mass Index , Cross-Sectional Studies , Feasibility Studies , Female , Humans , Male , Middle Aged , Pilot Projects , Self Report , Young Adult
15.
IEEE Comput Graph Appl ; 40(3): 73-82, 2020.
Article in English | MEDLINE | ID: mdl-32356729

ABSTRACT

Interactive data exploration and analysis is an inherently personal process. One's background, experience, interests, cognitive style, personality, and other sociotechnical factors often shape such a process, as well as the provenance of exploring, analyzing, and interpreting data. This Viewpoint posits both what personal information and how such personal information could be taken into account to design more effective visual analytic systems, a valuable and under-explored direction.

16.
IEEE Trans Vis Comput Graph ; 26(1): 1246-1255, 2020 01.
Article in English | MEDLINE | ID: mdl-31442990

ABSTRACT

Latency in a visualization system is widely believed to affect user behavior in measurable ways, such as requiring the user to wait for the visualization system to respond, leading to interruption of the analytic flow. While this effect is frequently observed and widely accepted, precisely how latency affects different analysis scenarios is less well understood. In this paper, we examine the role of latency in the context of visual search, an essential task in data foraging and exploration using visualization. We conduct a series of studies on Amazon Mechanical Turk and find that under certain conditions, latency is a statistically significant predictor of visual search behavior, which is consistent with previous studies. However, our results also suggest that task type, task complexity, and other factors can modulate the effect of latency, in some cases rendering latency statistically insignificant in predicting user behavior. This suggests a more nuanced view of the role of latency than previously reported. Building on these results and the findings of prior studies, we propose design guidelines for measuring and interpreting the effects of latency when evaluating performance on visual search tasks.

17.
IEEE Trans Vis Comput Graph ; 26(1): 863-873, 2020 01.
Article in English | MEDLINE | ID: mdl-31502978

ABSTRACT

The performance of deep learning models is dependent on the precise configuration of many layers and parameters. However, there are currently few systematic guidelines for how to configure a successful model. This means model builders often have to experiment with different configurations by manually programming different architectures (which is tedious and time consuming) or rely on purely automated approaches to generate and train the architectures (which is expensive). In this paper, we present Rapid Exploration of Model Architectures and Parameters, or REMAP, a visual analytics tool that allows a model builder to discover a deep learning model quickly via exploration and rapid experimentation of neural network architectures. In REMAP, the user explores the large and complex parameter space for neural network architectures using a combination of global inspection and local experimentation. Through a visual overview of a set of models, the user identifies interesting clusters of architectures. Based on their findings, the user can run ablation and variation experiments to identify the effects of adding, removing, or replacing layers in a given architecture and generate new models accordingly. They can also handcraft new models using a simple graphical interface. As a result, a model builder can build deep learning models quickly, efficiently, and without manual programming. We inform the design of REMAP through a design study with four deep learning model builders. Through a use case, we demonstrate that REMAP allows users to discover performant neural network architectures efficiently using visual exploration and user-defined semi-automated searches through the model space.

18.
IEEE Comput Graph Appl ; 39(5): 20-32, 2019.
Article in English | MEDLINE | ID: mdl-31199255

ABSTRACT

Interactive model steering helps people incrementally build machine learning models that are tailored to their domain and task. Existing visual analytic tools allow people to steer a single model (e.g., assignment attribute weights used by a dimension reduction model). However, the choice of model is critical in such situations. What if the model chosen is suboptimal for the task, dataset, or question being asked? What if instead of parameterizing and steering this model, a different model provides a better fit? This paper presents a technique to allow users to inspect and steer multiple machine learning models. The technique steers and samples models from a broader set of learning algorithms and model types. We incorporate this technique into a visual analytic prototype, BEAMES, that allows users to perform regression tasks via multimodel steering. This paper demonstrates the effectiveness of BEAMES via a use case, and discusses broader implications for multimodel steering.

19.
IEEE Trans Vis Comput Graph ; 25(3): 1474-1488, 2019 03.
Article in English | MEDLINE | ID: mdl-29993809

ABSTRACT

Recent visualization research efforts have incorporated experimental techniques and perceptual models from the vision science community. Perceptual laws such as Weber's law, for example, have been used to model the perception of correlation in scatterplots. While this thread of research has progressively refined the modeling of the perception of correlation in scatterplots, it remains unclear as to why such perception can be modeled using relatively simple functions, e.g., linear and log-linear. In this paper, we investigate a longstanding hypothesis that people use visual features in a chart as a proxy for statistical measures like correlation. For a given scatterplot, we extract 49 candidate visual features and evaluate which best align with existing models and participant judgments. The results support the hypothesis that people attend to a small number of visual features when discriminating correlation in scatterplots. We discuss how this result may account for prior conflicting findings, and how visual features provide a baseline for future model-based approaches in visualization evaluation and design.


Subject(s)
Computer Graphics , Judgment/physiology , Psychophysics/methods , Visual Perception/physiology , Female , Humans , Male , Models, Statistical
20.
Article in English | MEDLINE | ID: mdl-30137006

ABSTRACT

When inspecting information visualizations under time critical settings, such as emergency response or monitoring the heart rate in a surgery room, the user only has a small amount of time to view the visualization "at a glance". In these settings, it is important to provide a quantitative measure of the visualization to understand whether or not the visualization is too "complex" to accurately judge at a glance. This paper proposes Pixel Approximate Entropy (PAE), which adapts the approximate entropy statistical measure commonly used to quantify regularity and unpredictability in time-series data, as a measure of visual complexity for line charts. We show that PAE is correlated with user-perceived chart complexity, and that increased chart PAE correlates with reduced judgement accuracy. 'We also find that the correlation between PAE values and participants' judgment increases when the user has less time to examine the line charts.

SELECTION OF CITATIONS
SEARCH DETAIL
...