Distribution-based Exploration and Visualization of Large-scale Vector and Multivariate Fields


Book Description

Due to the ever increasing of computing power in the last few decades, the size of scientific data produced by various scientific simulations has been growing rapidly. As a result, effective techniques to visualize and explore those large-scale scientific data are becoming more and more important in understanding the data. However, for data at such a large scale, effective analysis and visualization is a non-trivial task due to several reasons. First, it is often time consuming and memory intensive to perform visualization and analysis directly on the original data. Second, as the data become large and complex, visualization usually suffers from visual cluttering and occlusion, which makes it difficult for users to understand the data. In order to address the aforementioned challenges, in this dissertation, a distribution-based query-driven framework to visualize and analyze large-scale scientific data is proposed. We propose to use statistical distributions to summarize large-scale data sets. The summarized data is then used to substitute the original data to support efficient and interactive query-driven visualization which is often free of occlusion. In this dissertation, the proposed framework is applied to flow fields and multivariate scalar fields. We first demonstrate the application of the proposed framework to flow fields. For a flow field, the statistical data summarization is computed from geometries such as streamlines and stream surfaces computed from the flow field. Stream surfaces and streamlines are two popular methods for visualizing flow fields. When the data size is large, distributed memory parallelism usually is needed. In this dissertation, a new scalable algorithm is proposed to compute stream surfaces from large-scale flow fields efficiently on distributed memory machines. After we obtain a large number of computed streamlines or stream surfaces, a direct visualization of all the densely computed geometries is seldom useful due to visual cluttering and occlusion. To solve the visual cluttering problem, a distribution-based query-driven framework to explore those densely computed streamlines is presented. Then, the proposed framework is applied to multivariate scalar fields. When dealing with multivariate data, in order to understand the data, it is often useful to show the regions of interest based on user specified criteria. In the presence of large-scale multivariate data, efficient techniques to summarize the data and answer users’ queries are needed. In this dissertation, we first propose to use multivariate histograms to summarize the data and demonstrate how effective query-driven visualization can be achieved based on those multivariate histograms. However, storing multivariate histograms in the form of multi-dimensional arrays is very expensive. To enable efficient visualization and exploration of multivariate data sets, we present a compact structure to store multivariate histograms to reduce their huge space cost while supporting different kinds of histogram query operations efficiently. We also present an interactive system to assist users to effectively design multivariate transfer functions. Multiple regions of interest could be highlighted through multivariate volume rendering based on the user specified multivariate transfer function.




In Situ Visualization for Computational Science


Book Description

This book provides an overview of the emerging field of in situ visualization, i.e. visualizing simulation data as it is generated. In situ visualization is a processing paradigm in response to recent trends in the development of high-performance computers. It has great promise in its ability to access increased temporal resolution and leverage extensive computational power. However, the paradigm also is widely viewed as limiting when it comes to exploration-oriented use cases. Furthermore, it will require visualization systems to become increasingly complex and constrained in usage. As research efforts on in situ visualization are growing, the state of the art and best practices are rapidly maturing. Specifically, this book contains chapters that reflect state-of-the-art research results and best practices in the area of in situ visualization. Our target audience are researchers and practitioners from the areas of mathematics computational science, high-performance computing, and computer science that work on or with in situ techniques, or desire to do so in future.




Scientific Visualization


Book Description

Based on the seminar that took place in Dagstuhl, Germany in June 2011, this contributed volume studies the four important topics within the scientific visualization field: uncertainty visualization, multifield visualization, biomedical visualization and scalable visualization. • Uncertainty visualization deals with uncertain data from simulations or sampled data, uncertainty due to the mathematical processes operating on the data, and uncertainty in the visual representation, • Multifield visualization addresses the need to depict multiple data at individual locations and the combination of multiple datasets, • Biomedical is a vast field with select subtopics addressed from scanning methodologies to structural applications to biological applications, • Scalability in scientific visualization is critical as data grows and computational devices range from hand-held mobile devices to exascale computational platforms. Scientific Visualization will be useful to practitioners of scientific visualization, students interested in both overview and advanced topics, and those interested in knowing more about the visualization process.




Visualization of Large Scale Volumetric Datasets


Book Description

In this thesis, we address the problem of large-scale data visualization from two aspects, dimensionality and resolution. We introduce a novel data structure called Differential Time- Histogram Table (DTHT) for visualization of time-varying (4D) scalar data. The proposed data structure takes advantage of the coherence in time-varying datasets and allows efficient updates of data necessary for rendering during data exploration and visualization while guaranteeing that the scalar field visualized is within a given error tolerance of the scalar field sampled. To address the high-resolution datasets, we propose a hierarchical data structure and introduce a novel hybrid framework to improve the quality of multi-resolution visualization. For more accurate rendering at coarser levels of detail, we reduce aliasing artifacts by approximating data distribution with a Gaussian basis at each level of detail and we reduce blurring by using transparent isosurfaces to capture high-frequency features usually missed in coarse resolution renderings.




Mathematical Foundations of Scientific Visualization, Computer Graphics, and Massive Data Exploration


Book Description

The goal of visualization is the accurate, interactive, and intuitive presentation of data. Complex numerical simulations, high-resolution imaging devices and incre- ingly common environment-embedded sensors are the primary generators of m- sive data sets. Being able to derive scienti?c insight from data increasingly depends on having mathematical and perceptual models to provide the necessary foundation for effective data analysis and comprehension. The peer-reviewed state-of-the-art research papers included in this book focus on continuous data models, such as is common in medical imaging or computational modeling. From the viewpoint of a visualization scientist, we typically collaborate with an application scientist or engineer who needs to visually explore or study an object which is given by a set of sample points, which originally may or may not have been connected by a mesh. At some point, one generally employs low-order piecewise polynomial approximationsof an object, using one or several dependent functions. In order to have an understanding of a higher-dimensional geometrical “object” or function, ef?cient algorithms supporting real-time analysis and manipulation (- tation, zooming) are needed. Often, the data represents 3D or even time-varying 3D phenomena (such as medical data), and the access to different layers (slices) and structures (the underlying topology) comprising such data is needed.




In Situ Techniques for Visualizing Large-scale Scientific Simulations


Book Description

The advancement of supercomputing technology enables scientific simulations at extreme scale and thus scientists' capability to study more complex problems at greater accuracy, leading to important scientific discoveries and breakthroughs. To validate and evaluate a simulation, the conventional approach is to output the simulation data to a massive storage device for post hoc analysis. One problem with this approach is that both the I/O capability and storage capacity are lagging behind a simulation's ability to generate data. As a result, scientists can only output a small fraction of the data, usually by sacrificing temporal resolution. Otherwise, a majority of the supercomputing time would be devoted to I/O. As supercomputing moves from the current petascale (1015) towards exascale (1018), the gap between the computing power and I/O capability is expected to become even wider, and therefore more data is expected to be discarded, which defeats the original purpose of performing the simulation at extreme scale. It has become clear that some simulated phenomena can only be captured at high precision and fine detail during the simulation before any data is discarded. This essentially suggests in situ processing. In situ processing is a promising concept, but there are still certain limitations to the current generation of in situ methods. Computational fluid dynamics (CFD) is an application domain that particularly benefits from in situ techniques. Common in situ tasks during CFD simulations include computing flow statistics, direct rendering of the flow field, and extracting flow features of interest. Performing these tasks in situ often requires the scientists to know a priori the specific aspects of the simulated phenomena to be captured or extracted. Often, such knowledge is not available before analyzing the data generated from simulations. Furthermore, many in situ methods pose a limitation; that is, the resulting geometric or imagery representations of the flow feature prohibit exploration. Such limitation greatly reduces the possibility of subsequent scientific analyses. This dissertation aims to enhance the explorability of in situ visualization in order to alleviate the requirement of a priori knowledge, and also evaluate the usability of in situ approach using two state-of-the-art combustion simulation codes. The basis of my approach is to allow in situ visualization parameters to be specified in a way to broaden the extent of post hoc exploration and analysis. Each of my techniques balances the explorability of the outputs against the costs to generate it. To support scalar field visualization, I derive a design embedding multiple layers of depth information into the in situ outputs, which not only enables post hoc isosurfaces visualization, but also supports 3D feature extraction and tracking without needing the original simulation data. To support vector field visualization, I derive a design based on an image-based, point-sprite cloud representation of flow movement to provide pathtubes exploration in multiple dimensions. To support post hoc statistical analysis, I generate regional probability distribution functions in situ, with which I conduct a long-term user-centered design study to iteratively enhance the usability of the solution. This dissertation introduces several new approaches to in situ visualization and analysis and successfully demonstrates their potential values to the simulation scientists.




Scalable Interactive Visualization


Book Description

This book is a printed edition of the Special Issue "Scalable Interactive Visualization" that was published in Informatics




Visualization of Time-Oriented Data


Book Description

Time is an exceptional dimension that is common to many application domains such as medicine, engineering, business, or science. Due to the distinct characteristics of time, appropriate visual and analytical methods are required to explore and analyze them. This book starts with an introduction to visualization and historical examples of visual representations. At its core, the book presents and discusses a systematic view of the visualization of time-oriented data along three key questions: what is being visualized (data), why something is visualized (user tasks), and how it is presented (visual representation). To support visual exploration, interaction techniques and analytical methods are required that are discussed in separate chapters. A large part of this book is devoted to a structured survey of 101 different visualization techniques as a reference for scientists conducting related research as well as for practitioners seeking information on how their time-oriented data can best be visualized.




Topological Methods in Data Analysis and Visualization IV


Book Description

This book presents contributions on topics ranging from novel applications of topological analysis for particular problems, through studies of the effectiveness of modern topological methods, algorithmic improvements on existing methods, and parallel computation of topological structures, all the way to mathematical topologies not previously applied to data analysis. Topological methods are broadly recognized as valuable tools for analyzing the ever-increasing flood of data generated by simulation or acquisition. This is particularly the case in scientific visualization, where the data sets have long since surpassed the ability of the human mind to absorb every single byte of data. The biannual TopoInVis workshop has supported researchers in this area for a decade, and continues to serve as a vital forum for the presentation and discussion of novel results in applications in the area, creating a platform to disseminate knowledge about such implementations throughout and beyond the community. The present volume, resulting from the 2015 TopoInVis workshop held in Annweiler, Germany, will appeal to researchers in the fields of scientific visualization and mathematics, domain scientists with an interest in advanced visualization methods, and developers of visualization software systems.




An Introduction to Applied Multivariate Analysis with R


Book Description

The majority of data sets collected by researchers in all disciplines are multivariate, meaning that several measurements, observations, or recordings are taken on each of the units in the data set. These units might be human subjects, archaeological artifacts, countries, or a vast variety of other things. In a few cases, it may be sensible to isolate each variable and study it separately, but in most instances all the variables need to be examined simultaneously in order to fully grasp the structure and key features of the data. For this purpose, one or another method of multivariate analysis might be helpful, and it is with such methods that this book is largely concerned. Multivariate analysis includes methods both for describing and exploring such data and for making formal inferences about them. The aim of all the techniques is, in general sense, to display or extract the signal in the data in the presence of noise and to find out what the data show us in the midst of their apparent chaos. An Introduction to Applied Multivariate Analysis with R explores the correct application of these methods so as to extract as much information as possible from the data at hand, particularly as some type of graphical representation, via the R software. Throughout the book, the authors give many examples of R code used to apply the multivariate techniques to multivariate data.