Performance Analysis of Parallel Applications for HPC


Book Description

This book presents a hybrid static-dynamic approach for efficient performance analysis of parallel applications on HPC systems. Performance analysis is essential to finding performance bottlenecks and understanding the performance behaviors of parallel applications on HPC systems. However, current performance analysis techniques usually incur significant overhead. Our book introduces a series of approaches for lightweight performance analysis. We combine static and dynamic analysis to reduce the overhead of performance analysis. Based on this hybrid static-dynamic approach, we then propose several innovative techniques for various performance analysis scenarios, including communication analysis, memory analysis, noise analysis, computation analysis, and scalability analysis. Through these specific performance analysis techniques, we convey to readers the idea of using static analysis to support dynamic analysis. To gain the most from the book, readers should have a basic grasp of parallel computing, computer architecture, and compilation techniques.




Sustained Simulation Performance 2019 and 2020


Book Description

This book presents the state of the art in High Performance Computing on modern supercomputer architectures. It addresses trends in hardware and software development in general. The contributions cover a broad range of topics, from performance evaluations in context with power efficiency to Computational Fluid Dynamics and High Performance Data Analytics. In addition, they explore new topics like the use of High Performance Computers in the field of Artificial Intelligence and Machine Learning. All contributions are based on selected papers presented at the 30th Workshop on Sustained Simulation Performance (WSSP) held at the High Performance Computing Center, University of Stuttgart, Germany in October 2019 and on the papers for the planned Workshop on Sustained Simulation Performance in March 2020, which could not take place due to the Covid-19 pandemic.




Supercomputing


Book Description

This book constitutes the refereed proceedings of the 8th Russian Supercomputing Days on Supercomputing, RuSCDays 2022, which took place in Moscow, Russia, in September 2022. The 49 full papers and 1 short paper presented in this volume were carefully reviewed and selected from 94 submissions. The papers are organized in the following topical sections: Supercomputer Simulation; HPC, BigData, AI: Architectures, Technologies, Tools; Distributed and Cloud Computing.




Automated Optimization Methods for Scientific Workflows in e-Science Infrastructures


Book Description

Scientific workflows have emerged as a key technology that assists scientists with the design, management, execution, sharing and reuse of in silico experiments. Workflow management systems simplify the management of scientific workflows by providing graphical interfaces for their development, monitoring and analysis. Nowadays, e-Science combines such workflow management systems with large-scale data and computing resources into complex research infrastructures. For instance, e-Science allows the conveyance of best practice research in collaborations by providing workflow repositories, which facilitate the sharing and reuse of scientific workflows. However, scientists are still faced with different limitations while reusing workflows. One of the most common challenges they meet is the need to select appropriate applications and their individual execution parameters. If scientists do not want to rely on default or experience-based parameters, the best-effort option is to test different workflow set-ups using either trial and error approaches or parameter sweeps. Both methods may be inefficient or time consuming respectively, especially when tuning a large number of parameters. Therefore, scientists require an effective and efficient mechanism that automatically tests different workflow set-ups in an intelligent way and will help them to improve their scientific results. This thesis addresses the limitation described above by defining and implementing an approach for the optimization of scientific workflows. In the course of this work, scientists’ needs are investigated and requirements are formulated resulting in an appropriate optimization concept. In a following step, this concept is prototypically implemented by extending a workflow management system with an optimization framework, including general mechanisms required to conduct workflow optimization. As optimization is an ongoing research topic, different algorithms are provided by pluggable extensions (plugins) that can be loosely coupled with the framework, resulting in a generic and quickly extendable system. In this thesis, an exemplary plugin is introduced which applies a Genetic Algorithm for parameter optimization. In order to accelerate and therefore make workflow optimization feasible at all, e-Science infrastructures are utilized for the parallel execution of scientific workflows. This is empowered by additional extensions enabling the execution of applications and workflows on distributed computing resources. The actual implementation and therewith the general approach of workflow optimization is experimentally verified by four use cases in the life science domain. All workflows were significantly improved, which demonstrates the advantage of the proposed workflow optimization. Finally, a new collaboration-based approach is introduced that harnesses optimization provenance to make optimization faster and more robust in the future.







Euro-Par 2005 Parallel Processing


Book Description

Euro-Par 2005 was the eleventh conference in the Euro-Par series. It was organized by the Centre for Informatics and Information Technology (CITI) and the Department of Informatics of the Faculty of Science and Technology of Universidade Nova de Lisboa, at the Campus of Monte de Caparica.




Applied Parallel Computing. New Paradigms for HPC in Industry and Academia


Book Description

The papers in this volume were presented at PARA 2000, the Fifth International Workshop on Applied Parallel Computing. PARA 2000 was held in Bergen, Norway, June 18-21, 2000. The workshop was organized by Parallab and the Department of Informatics at the University of Bergen. The general theme for PARA 2000 was New paradigms for HPC in industry and academia focusing on: { High-performance computing applications in academia and industry, { The use of Java in high-performance computing, { Grid and Meta computing, { Directions in high-performance computing and networking, { Education in Computational Science. The workshop included 9 invited presentations and 39 contributed pres- tations. The PARA 2000 meeting began with a one-day tutorial on OpenMP programming led by Timothy Mattson. This was followed by a three-day wor- hop. The rst three PARA workshops were held at the Technical University of Denmark (DTU), Lyngby (1994, 1995, and 1996). Following PARA’96, an - ternational steering committee for the PARA meetings was appointed and the committee decided that a workshop should take place every second year in one of the Nordic countries. The 1998 workshop was held at Ume a University, Sweden. One important aim of these workshops is to strengthen the ties between HPC centers, academia, and industry in the Nordic countries as well as worldwide. The University of Bergen organized the 2000 workshop and the next workshop in the year 2002 will take place at the Helsinki University of Technology, Espoo, Finland.







Parallel Processing for Scientific Computing


Book Description

Scientific computing has often been called the third approach to scientific discovery, emerging as a peer to experimentation and theory. Historically, the synergy between experimentation and theory has been well understood: experiments give insight into possible theories, theories inspire experiments, experiments reinforce or invalidate theories, and so on. As scientific computing has evolved to produce results that meet or exceed the quality of experimental and theoretical results, it has become indispensable.Parallel processing has been an enabling technology in scientific computing for more than 20 years. This book is the first in-depth discussion of parallel computing in 10 years; it reflects the mix of topics that mathematicians, computer scientists, and computational scientists focus on to make parallel processing effective for scientific problems. Presently, the impact of parallel processing on scientific computing varies greatly across disciplines, but it plays a vital role in most problem domains and is absolutely essential in many of them. Parallel Processing for Scientific Computing is divided into four parts: The first concerns performance modeling, analysis, and optimization; the second focuses on parallel algorithms and software for an array of problems common to many modeling and simulation applications; the third emphasizes tools and environments that can ease and enhance the process of application development; and the fourth provides a sampling of applications that require parallel computing for scaling to solve larger and realistic models that can advance science and engineering. This edited volume serves as an up-to-date reference for researchers and application developers on the state of the art in scientific computing. It also serves as an excellent overview and introduction, especially for graduate and senior-level undergraduate students interested in computational modeling and simulation and related computer science and applied mathematics aspects.Contents List of Figures; List of Tables; Preface; Chapter 1: Frontiers of Scientific Computing: An Overview; Part I: Performance Modeling, Analysis and Optimization. Chapter 2: Performance Analysis: From Art to Science; Chapter 3: Approaches to Architecture-Aware Parallel Scientific Computation; Chapter 4: Achieving High Performance on the BlueGene/L Supercomputer; Chapter 5: Performance Evaluation and Modeling of Ultra-Scale Systems; Part II: Parallel Algorithms and Enabling Technologies. Chapter 6: Partitioning and Load Balancing; Chapter 7: Combinatorial Parallel and Scientific Computing; Chapter 8: Parallel Adaptive Mesh Refinement; Chapter 9: Parallel Sparse Solvers, Preconditioners, and Their Applications; Chapter 10: A Survey of Parallelization Techniques for Multigrid Solvers; Chapter 11: Fault Tolerance in Large-Scale Scientific Computing; Part III: Tools and Frameworks for Parallel Applications. Chapter 12: Parallel Tools and Environments: A Survey; Chapter 13: Parallel Linear Algebra Software; Chapter 14: High-Performance Component Software Systems; Chapter 15: Integrating Component-Based Scientific Computing Software; Part IV: Applications of Parallel Computing. Chapter 16: Parallel Algorithms for PDE-Constrained Optimization; Chapter 17: Massively Parallel Mixed-Integer Programming; Chapter 18: Parallel Methods and Software for Multicomponent Simulations; Chapter 19: Parallel Computational Biology; Chapter 20: Opportunities and Challenges for Parallel Computing in Science and Engineering; Index.