Information

This full-day tutorial will be held during IEEE VisWeek 2012 on Monday, October 15, 2012 from 08:30am to 05:55pm in the Grand Ballroom A of the Sheraton Seattle. We thank all participating speakers and are looking forward to your attendance.

Content

Organizers

Schedule

Session Speaker Description Materials Time
Session 1: Introductory session Alex T. Pang Historical review of uncertainty visualization including a perspective on the whole field Slides 55 + 10 min
Session 2: Uncertainty Modeling Hans-Christian Hege Overview on methods for modeling uncertainties in data; statistical modeling of spatio-temporal data Slides 55 + 10 min
Session 3: Statistical Visualization Kristin Potter Statistical representations of uncertainty and techniques for visualizing these statistics Slides 55 + 10 min
Lunch
Session 4: Structural uncertainty Rüdiger Westermann, Tobias Pfaffelmoser Emphasis on the relevance and use of correlation as an indicator for the variability of structures in scalar fields Slides 55 + 10 min
Session 5: Parameter space analysis Torsten Möller Abstraction of tasks and data in the realm of parameter exploration Slides
Videos
55 + 10 min
Session 6: Closing session Stefan Bruckner, Christoph Heinzl, Eduard Gröller Similarity-based approaches for the visual exploration and analysis of parameter spaces, industrial applications and outlook of this domain Slides, Part 1 (Bruckner)
Slides, Part 2 (Heinzl)
Video, Part 2 (Heinzl)
Slides, Part 3 (Göller)
60 + 10 min

Abstract

Within the past decades visualization advanced to a powerful means of exploring and analyzing data. Recent developments in both hard- and software contributed to previously unthinkable evaluations and visualizations of data with strongly increasing sizes and levels of complexity.

Providing just insight into available data of a problem seems not to be sufficient anymore: Uncertainty and parameter space analyses in visualization are becoming more prevalent and may be found in astronomic, (bio)-medical, industrial, and engineering applications. The major goal is to find out, at which stage of the pipeline - from data acquisition to the final rendering of the output image - how much uncertainty is introduced and consequently how the desired result (e.g., a dimensional measurement feature) is affected. Therefore effective methods and techniques are required by domain specialists, which help to understand how data is generated, how reliable is the generated data, and where and why data is uncertain.

Furthermore, as the problems to investigate are becoming increasingly complex, also finding suitable algorithms providing the desired solution tends to be more difficult. Additional questions may arise, e.g., how does a slight parameter change modify the result, how stable is a parameter, in which range is a parameter stable or which parameter set is optimal for a specific problem. Metaphorically speaking, an algorithm for solving a problem may be seen as finding a path through some rugged terrain (the core problem) ranging from the high grounds of theory to the haunted swamps of heuristics. There are many different paths through this terrain with different levels of comfort, length, and stability. Finding all possible paths corresponds in our case to doing an analysis of all possible parameters of a problem solving algorithm, which yields a typically multi-dimensional parameter space. This parameter space allows for an analysis of the quality and stability of a specific parameter set. In many cases of conventional visualization approaches the issues of uncertainty and parameter space analyses are neglected. For a long time, uncertainty - if visualized at all - used to be depicted as blurred data. But in most cases the uncertainty in the base data is not considered at all and just the quantities of interest are calculated. And even to calculate these quantities of interest, too often an empirically found parameter set is used to parameterize the underlying algorithms without exploring its sensitivity to changes and without exploring the whole parameter space to find the global or a local optimum.

This tutorial aims to open minds and to look at our data and the parameter sets of our algorithms with a healthy skepticism. In the tutorial we combine uncertainty visualization and parameter space analyses which we believe is essential for the acceptance and applicability of future algorithms and techniques. The tutorial provides six sessions starting with an overview of uncertainty visualization including a historical perspective, uncertainty modeling and statistical visualization. The second part of the tutorial will be dedicated to structural uncertainty, parameter space analysis, industrial applications of uncertainty visualization and an outlook in this domain.

Additional Files and Images

Additional images and videos

Teaser: Uncertainty visualization

Additional files

Tutorial overview: Speaker biographies and abstract

Weblinks

No further information available.

BibTeX

@WorkshopTalk{VisWeek-Tutorial-2012-Uncertainty,
  title =      "IEEE VisWeek 2012 Tutorial on Uncertainty and Parameter
               Space Analysis in Visualization",
  author =     "Christoph Heinzl and Stefan Bruckner and Meister Eduard
               Gr{"o}ller and Alex Pang and Hans-Christian Hege and Kristin
               Potter and R{"u}diger Westermann and Tobias Pfaffelmoser and
               Torsten M{"o}ller",
  year =       "2012",
  abstract =   "Within the past decades visualization advanced to a powerful
               means of exploring and analyzing data. Recent developments
               in both hard- and software contributed to previously
               unthinkable evaluations and visualizations of data with
               strongly increasing sizes and levels of complexity. 
               Providing just insight into available data of a problem
               seems not to be sufficient anymore: Uncertainty and
               parameter space analyses in visualization are becoming more
               prevalent and may be found in astronomic, (bio)-medical,
               industrial, and engineering applications. The major goal is
               to find out, at which stage of the pipeline - from data
               acquisition to the final rendering of the output image - how
               much uncertainty is introduced and consequently how the
               desired result (e.g., a dimensional measurement feature) is
               affected. Therefore effective methods and techniques are
               required by domain specialists, which help to understand how
               data is generated, how reliable is the generated data, and
               where and why data is uncertain.  Furthermore, as the
               problems to investigate are becoming increasingly complex,
               also finding suitable algorithms providing the desired
               solution tends to be more difficult. Additional questions
               may arise, e.g., how does a slight parameter change modify
               the result, how stable is a parameter, in which range is a
               parameter stable or which parameter set is optimal for a
               specific problem. Metaphorically speaking, an algorithm for
               solving a problem may be seen as finding a path through some
               rugged terrain (the core problem) ranging from the high
               grounds of theory to the haunted swamps of heuristics. There
               are many different paths through this terrain with different
               levels of comfort, length, and stability. Finding all
               possible paths corresponds in our case to doing an analysis
               of all possible parameters of a problem solving algorithm,
               which yields a typically multi-dimensional parameter space.
               This parameter space allows for an analysis of the quality
               and stability of a specific parameter set. In many cases of
               conventional visualization approaches the issues of
               uncertainty and parameter space analyses are neglected. For
               a long time, uncertainty - if visualized at all - used to be
               depicted as blurred data. But in most cases the uncertainty
               in the base data is not considered at all and just the
               quantities of interest are calculated. And even to calculate
               these quantities of interest, too often an empirically found
               parameter set is used to parameterize the underlying
               algorithms without exploring its sensitivity to changes and
               without exploring the whole parameter space to find the
               global or a local optimum.  This tutorial aims to open minds
               and to look at our data and the parameter sets of our
               algorithms with a healthy skepticism. In the tutorial we
               combine uncertainty visualization and parameter space
               analyses which we believe is essential for the acceptance
               and applicability of future algorithms and techniques. The
               tutorial provides six sessions starting with an overview of
               uncertainty visualization including a historical
               perspective, uncertainty modeling and statistical
               visualization. The second part of the tutorial will be
               dedicated to structural uncertainty, parameter space
               analysis, industrial applications of uncertainty
               visualization and an outlook in this domain. ",
  month =      oct,
  location =   "Seattle, WA, USA",
  keywords =   "uncertainty visualization, parameter space analysis",
  URL =        "https://www.cg.tuwien.ac.at/research/publications/2012/VisWeek-Tutorial-2012-Uncertainty/",
}