Usage-Driven Database Design


Book Description

Design great databases—from logical data modeling through physical schema definition. You will learn a framework that finally cracks the problem of merging data and process models into a meaningful and unified design that accounts for how data is actually used in production systems. Key to the framework is a method for taking the logical data model that is a static look at the definition of the data, and merging that static look with the process models describing how the data will be used in actual practice once a given system is implemented. The approach solves the disconnect between the static definition of data in the logical data model and the dynamic flow of the data in the logical process models. The design framework in this book can be used to create operational databases for transaction processing systems, or for data warehouses in support of decision support systems. The information manager can be a flat file, Oracle Database, IMS, NoSQL, Cassandra, Hadoop, or any other DBMS. Usage-Driven Database Design emphasizes practical aspects of design, and speaks to what works, what doesn’t work, and what to avoid at all costs. Included in the book are lessons learned by the author over his 30+ years in the corporate trenches. Everything in the book is grounded on good theory, yet demonstrates a professional and pragmatic approach to design that can come only from decades of experience. Presents an end-to-end framework from logical data modeling through physical schema definition. Includes lessons learned, techniques, and tricks that can turn a database disaster into a success. Applies to all types of database management systems, including NoSQL such as Cassandra and Hadoop, and mainstream SQL databases such as Oracle and SQL Server What You'll Learn Create logical data models that accurately reflect the real world of the user Create usage scenarios reflecting how applications will use a new database Merge static data models with dynamic process models to create resilient yet flexible database designs Support application requirements by creating responsive database schemas in any database architecture Cope with big data and unstructured data for transaction processing and decision support systems Recognize when relational approaches won’t work, and when to turn toward NoSQL solutions such as Cassandra or Hadoop Who This Book Is For System developers, including business analysts, database designers, database administrators, and application designers and developers who must design or interact with database systems




The Data Model Resource Book, Volume 1


Book Description

A quick and reliable way to build proven databases for core business functions Industry experts raved about The Data Model Resource Book when it was first published in March 1997 because it provided a simple, cost-effective way to design databases for core business functions. Len Silverston has now revised and updated the hugely successful 1st Edition, while adding a companion volume to take care of more specific requirements of different businesses. This updated volume provides a common set of data models for specific core functions shared by most businesses like human resources management, accounting, and project management. These models are standardized and are easily replicated by developers looking for ways to make corporate database development more efficient and cost effective. This guide is the perfect complement to The Data Model Resource CD-ROM, which is sold separately and provides the powerful design templates discussed in the book in a ready-to-use electronic format. A free demonstration CD-ROM is available with each copy of the print book to allow you to try before you buy the full CD-ROM.




Data Model Patterns: A Metadata Map


Book Description

Data Model Patterns: A Metadata Map not only presents a conceptual model of a metadata repository but also demonstrates a true enterprise data model of the information technology industry itself. It provides a step-by-step description of the model and is organized so that different readers can benefit from different parts. It offers a view of the world being addressed by all the techniques, methods, and tools of the information processing industry (for example, object-oriented design, CASE, business process re-engineering, etc.) and presents several concepts that need to be addressed by such tools. This book is pertinent, with companies and government agencies realizing that the data they use represent a significant corporate resource recognize the need to integrate data that has traditionally only been available from disparate sources. An important component of this integration is management of the "metadata" that describe, catalogue, and provide access to the various forms of underlying business data. The "metadata repository" is essential to keep track of the various physical components of these systems and their semantics. The book is ideal for data management professionals, data modeling and design professionals, and data warehouse and database repository designers. - A comprehensive work based on the Zachman Framework for information architecture—encompassing the Business Owner's, Architect's, and Designer's views, for all columns (data, activities, locations, people, timing, and motivation) - Provides a step-by-step description of model and is organized so that different readers can benefit from different parts - Provides a view of the world being addressed by all the techniques, methods and tools of the information processing industry (for example, object-oriented design, CASE, business process re-engineering, etc.) - Presents many concepts that are not currently being addressed by such tools — and should be




Developing High Quality Data Models


Book Description

Developing High Quality Data Models provides an introduction to the key principles of data modeling. It explains the purpose of data models in both developing an Enterprise Architecture and in supporting Information Quality; common problems in data model development; and how to develop high quality data models, in particular conceptual, integration, and enterprise data models. The book is organized into four parts. Part 1 provides an overview of data models and data modeling including the basics of data model notation; types and uses of data models; and the place of data models in enterprise architecture. Part 2 introduces some general principles for data models, including principles for developing ontologically based data models; and applications of the principles for attributes, relationship types, and entity types. Part 3 presents an ontological framework for developing consistent data models. Part 4 provides the full data model that has been in development throughout the book. The model was created using Jotne EPM Technologys EDMVisualExpress data modeling tool. This book was designed for all types of modelers: from those who understand data modeling basics but are just starting to learn about data modeling in practice, through to experienced data modelers seeking to expand their knowledge and skills and solve some of the more challenging problems of data modeling. - Uses a number of common data model patterns to explain how to develop data models over a wide scope in a way that is consistent and of high quality - Offers generic data model templates that are reusable in many applications and are fundamental for developing more specific templates - Develops ideas for creating consistent approaches to high quality data models




Conceptual Data Modeling and Database Design: A Fully Algorithmic Approach, Volume 1


Book Description

This new book aims to provide both beginners and experts with a completely algorithmic approach to data analysis and conceptual modeling, database design, implementation, and tuning, starting from vague and incomplete customer requests and ending with IBM DB/2, Oracle, MySQL, MS SQL Server, or Access based software applications. A rich panoply of s




PHealth 2015


Book Description

Smart mobile systems, smart textiles, smart implants and sensor controlled medical devices are among the recent developments which have become important enablers for telemedicine and next-generation health services. Social media and gamification have added yet another dimension to Personalized Health (pHealth). This book presents the proceedings of pHealth 2015, the 12th International Conference on Wearable Micro and Nano Technologies for Personalized Health, held in Västerås, Sweden, in June 2015. The conference addressed mobile technologies, knowledge-driven applications and computer-assisted decision support, as well as apps designed to support the elderly and those with chronic conditions in their daily lives. The 23 conference papers, three keynotes and two specially invited contributions included here address the fundamental scientific and methodological challenges of adaptive, autonomous and intelligent pHealth approaches. Participants at this truly interdisciplinary conference included representatives from all relevant stakeholder communities, and the topics covered will be of interest to all those whose work involves improving the quality of medical services, optimizing industrial competitiveness and managing healthcare costs.




Computer Applications


Book Description

The two-volume set CCIS 1959 and 1960 constitutes the refereed post-conference proceedings of the 38th CCF National Conference on Computer Applications, CCF NCCA 2023, held in Suzhou, China, during July 16–20, 2023. The 39 revised full papers presented in these proceedings were carefully reviewed and selected from 197 submissions. The papers are organized in the following topical sections: Volume I: Artificial intelligence and application. Volume II: Data science and technology; pattern recognition and machine learning; network communication and security; frontier and comprehensive applications.




Fundamentals of Data Warehouses


Book Description

This book presents the first comparative review of the state of the art and the best current practices of data warehouses. It covers source and data integration, multidimensional aggregation, query optimization, metadata management, quality assessment, and design optimization. A conceptual framework is presented by which the architecture and quality of a data warehouse can be assessed and improved using enriched metadata management combined with advanced techniques from databases, business modeling, and artificial intelligence.




Computerworld


Book Description

For more than 40 years, Computerworld has been the leading source of technology news and information for IT influencers worldwide. Computerworld's award-winning Web site (Computerworld.com), twice-monthly publication, focused conference series and custom research form the hub of the world's largest global IT media network.