Data Management on New Hardware


Book Description

This book contains selected papers from the 7th International Workshop on Accelerating Analytics and Data Management Systems Using Modern Processor and Storage Architectures, ADMS 2016, and the 4th International Workshop on In-Memory Data Management and Analytics, IMDM 2016, held in New Dehli, India, in September 2016. The joint Workshops were co-located with VLDB 2016. The 9 papers presented were carefully reviewed and selected from 18 submissions. They investigate opportunities in accelerating analytics/data management systems and workloads (including traditional OLTP, data warehousing/OLAP, ETL streaming/real-time, business analytics, and XML/RDF processing) running memory-only environments, using processors (e.g. commodity and specialized multi-core, GPUs and FPGAs, storage systems (e.g. storage-class memories like SSDs and phase-change memory), and hybrid programming models like CUDA, OpenCL, and Open ACC. The papers also explore the interplay between overall system design, core algorithms, query optimization strategies, programming approaches, performance modeling and evaluation, from the perspective of data management applications.




Proceedings of the Future Technologies Conference (FTC) 2019


Book Description

This book presents state-of-the-art intelligent methods and techniques for solving real-world problems and offers a vision of future research. Featuring 143 papers from the 4th Future Technologies Conference, held in San Francisco, USA, in 2019, it covers a wide range of important topics, including, but not limited to, computing, electronics, artificial intelligence, robotics, security and communications and their applications to the real world. As such, it is an interesting, exciting and inspiring read.




Transaction Processing on Modern Hardware


Book Description

The last decade has brought groundbreaking developments in transaction processing. This resurgence of an otherwise mature research area has spurred from the diminishing cost per GB of DRAM that allows many transaction processing workloads to be entirely memory-resident. This shift demanded a pause to fundamentally rethink the architecture of database systems. The data storage lexicon has now expanded beyond spinning disks and RAID levels to include the cache hierarchy, memory consistency models, cache coherence and write invalidation costs, NUMA regions, and coherence domains. New memory technologies promise fast non-volatile storage and expose unchartered trade-offs for transactional durability, such as exploiting byte-addressable hot and cold storage through persistent programming that promotes simpler recovery protocols. In the meantime, the plateauing single-threaded processor performance has brought massive concurrency within a single node, first in the form of multi-core, and now with many-core and heterogeneous processors. The exciting possibility to reshape the storage, transaction, logging, and recovery layers of next-generation systems on emerging hardware have prompted the database research community to vigorously debate the trade-offs between specialized kernels that narrowly focus on transaction processing performance vs. designs that permit transactionally consistent data accesses from decision support and analytical workloads. In this book, we aim to classify and distill the new body of work on transaction processing that has surfaced in the last decade to navigate researchers and practitioners through this intricate research subject.




New Trends in Databases and Information Systems


Book Description

Database and information systems technologies have been rapidly evolving in several directions over the past years. New types and kinds of data, new types of applications and information systems to support them raise diverse challenges to be addressed. The so-called big data challenge, streaming data management and processing, social networks and other complex data analysis, including semantic reasoning into information systems supporting for instance trading, negotiations, and bidding mechanisms are just some of the emerging research topics. This volume contains papers contributed by six workshops: ADBIS Workshop on GPUs in Databases (GID 2012), Mining Complex and Stream Data (MCSD'12), International Workshop on Ontologies meet Advanced Information Systems (OAIS'2012), Second Workshop on Modeling Multi-commodity Trade: Data models and processing (MMT'12), 1st ADBIS Workshop on Social Data Processing (SDP'12), 1st ADBIS Workshop on Social and Algorithmic Issues in Business Support (SAIBS), and the Ph.D. Consortium associated with the ADBIS 2012 conference that report on the recent developments and an ongoing research in the aforementioned areas.




Big Data


Book Description

Although there are already some books published on Big Data, most of them only cover basic concepts and society impacts and ignore the internal implementation details-making them unsuitable to R&D people. To fill such a need, Big Data: Storage, Sharing, and Security examines Big Data management from an R&D perspective. It covers the 3S desi




Euro-Par 2008 Parallel Processing


Book Description

This book constitutes the refereed proceedings of the 14th International Conference on Parallel Computing, Euro-Par 2008, held in Las Palmas de Gran Canaria, Spain, in August 2008. The 86 revised papers presented were carefully reviewed and selected from 264 submissions. The papers are organized in topical sections on support tools and environments; performance prediction and evaluation; scheduling and load balancing; high performance architectures and compilers; parallel and distributed databases; grid and cluster computing; peer-to-peer computing; distributed systems and algorithms; parallel and distributed programming; parallel numerical algorithms; distributed and high-performance multimedia; theory and algorithms for parallel computation; and high performance networks.




Industrial Engineering: Concepts, Methodologies, Tools, and Applications


Book Description

Industrial engineering affects all levels of society, with innovations in manufacturing and other forms of engineering oftentimes spawning cultural or educational shifts along with new technologies. Industrial Engineering: Concepts, Methodologies, Tools, and Applications serves as a vital compendium of research, detailing the latest research, theories, and case studies on industrial engineering. Bringing together contributions from authors around the world, this three-volume collection represents the most sophisticated research and developments from the field of industrial engineering and will prove a valuable resource for researchers, academics, and practitioners alike.




Storage Systems


Book Description

Storage Systems: Organization, Performance, Coding, Reliability and Their Data Processing was motivated by the 1988 Redundant Array of Inexpensive/Independent Disks proposal to replace large form factor mainframe disks with an array of commodity disks. Disk loads are balanced by striping data into strips—with one strip per disk— and storage reliability is enhanced via replication or erasure coding, which at best dedicates k strips per stripe to tolerate k disk failures. Flash memories have resulted in a paradigm shift with Solid State Drives (SSDs) replacing Hard Disk Drives (HDDs) for high performance applications. RAID and Flash have resulted in the emergence of new storage companies, namely EMC, NetApp, SanDisk, and Purestorage, and a multibillion-dollar storage market. Key new conferences and publications are reviewed in this book.The goal of the book is to expose students, researchers, and IT professionals to the more important developments in storage systems, while covering the evolution of storage technologies, traditional and novel databases, and novel sources of data. We describe several prototypes: FAWN at CMU, RAMCloud at Stanford, and Lightstore at MIT; Oracle's Exadata, AWS' Aurora, Alibaba's PolarDB, Fungible Data Center; and author's paper designs for cloud storage, namely heterogeneous disk arrays and hierarchical RAID. - Surveys storage technologies and lists sources of data: measurements, text, audio, images, and video - Familiarizes with paradigms to improve performance: caching, prefetching, log-structured file systems, and merge-trees (LSMs) - Describes RAID organizations and analyzes their performance and reliability - Conserves storage via data compression, deduplication, compaction, and secures data via encryption - Specifies implications of storage technologies on performance and power consumption - Exemplifies database parallelism for big data, analytics, deep learning via multicore CPUs, GPUs, FPGAs, and ASICs, e.g., Google's Tensor Processing Units




Network and Parallel Computing


Book Description

This book constitutes the refereed proceedings of the IFIP International Conference on Network and Parallel Computing, NPC 2007. It covers network applications: cluster and grid computing, peer-to-peer computing; network technologies: network algorithms, network reliability and dependability; network and parallel architectures: multicore design issues, performance modeling and evaluation; and parallel and distributed software: data mining, parallel programming tools and compilers.




Supervised Descriptive Pattern Mining


Book Description

This book provides a general and comprehensible overview of supervised descriptive pattern mining, considering classic algorithms and those based on heuristics. It provides some formal definitions and a general idea about patterns, pattern mining, the usefulness of patterns in the knowledge discovery process, as well as a brief summary on the tasks related to supervised descriptive pattern mining. It also includes a detailed description on the tasks usually grouped under the term supervised descriptive pattern mining: subgroups discovery, contrast sets and emerging patterns. Additionally, this book includes two tasks, class association rules and exceptional models, that are also considered within this field. A major feature of this book is that it provides a general overview (formal definitions and algorithms) of all the tasks included under the term supervised descriptive pattern mining. It considers the analysis of different algorithms either based on heuristics or based on exhaustive search methodologies for any of these tasks. This book also illustrates how important these techniques are in different fields, a set of real-world applications are described. Last but not least, some related tasks are also considered and analyzed. The final aim of this book is to provide a general review of the supervised descriptive pattern mining field, describing its tasks, its algorithms, its applications, and related tasks (those that share some common features). This book targets developers, engineers and computer scientists aiming to apply classic and heuristic-based algorithms to solve different kinds of pattern mining problems and apply them to real issues. Students and researchers working in this field, can use this comprehensive book (which includes its methods and tools) as a secondary textbook.