Data Engineering with AWS Cookbook


Book Description

Master AWS data engineering services and techniques for orchestrating pipelines, building layers, and managing migrations Key Features Get up to speed with the different AWS technologies for data engineering Learn the different aspects and considerations of building data lakes, such as security, storage, and operations Get hands on with key AWS services such as Glue, EMR, Redshift, QuickSight, and Athena for practical learning Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionPerforming data engineering with Amazon Web Services (AWS) combines AWS's scalable infrastructure with robust data processing tools, enabling efficient data pipelines and analytics workflows. This comprehensive guide to AWS data engineering will teach you all you need to know about data lake management, pipeline orchestration, and serving layer construction. Through clear explanations and hands-on exercises, you’ll master essential AWS services such as Glue, EMR, Redshift, QuickSight, and Athena. Additionally, you’ll explore various data platform topics such as data governance, data quality, DevOps, CI/CD, planning and performing data migration, and creating Infrastructure as Code. As you progress, you will gain insights into how to enrich your platform and use various AWS cloud services such as AWS EventBridge, AWS DataZone, and AWS SCT and DMS to solve data platform challenges. Each recipe in this book is tailored to a daily challenge that a data engineer team faces while building a cloud platform. By the end of this book, you will be well-versed in AWS data engineering and have gained proficiency in key AWS services and data processing techniques. You will develop the necessary skills to tackle large-scale data challenges with confidence.What you will learn Define your centralized data lake solution, and secure and operate it at scale Identify the most suitable AWS solution for your specific needs Build data pipelines using multiple ETL technologies Discover how to handle data orchestration and governance Explore how to build a high-performing data serving layer Delve into DevOps and data quality best practices Migrate your data from on-premises to AWS Who this book is for If you're involved in designing, building, or overseeing data solutions on AWS, this book provides proven strategies for addressing challenges in large-scale data environments. Data engineers as well as big data professionals looking to enhance their understanding of AWS features for optimizing their workflow, even if they're new to the platform, will find value. Basic familiarity with AWS security (users and roles) and command shell is recommended.




Data Engineering with AWS


Book Description

The missing expert-led manual for the AWS ecosystem — go from foundations to building data engineering pipelines effortlessly Purchase of the print or Kindle book includes a free eBook in the PDF format. Key Features Learn about common data architectures and modern approaches to generating value from big data Explore AWS tools for ingesting, transforming, and consuming data, and for orchestrating pipelines Learn how to architect and implement data lakes and data lakehouses for big data analytics from a data lakes expert Book DescriptionWritten by a Senior Data Architect with over twenty-five years of experience in the business, Data Engineering for AWS is a book whose sole aim is to make you proficient in using the AWS ecosystem. Using a thorough and hands-on approach to data, this book will give aspiring and new data engineers a solid theoretical and practical foundation to succeed with AWS. As you progress, you’ll be taken through the services and the skills you need to architect and implement data pipelines on AWS. You'll begin by reviewing important data engineering concepts and some of the core AWS services that form a part of the data engineer's toolkit. You'll then architect a data pipeline, review raw data sources, transform the data, and learn how the transformed data is used by various data consumers. You’ll also learn about populating data marts and data warehouses along with how a data lakehouse fits into the picture. Later, you'll be introduced to AWS tools for analyzing data, including those for ad-hoc SQL queries and creating visualizations. In the final chapters, you'll understand how the power of machine learning and artificial intelligence can be used to draw new insights from data. By the end of this AWS book, you'll be able to carry out data engineering tasks and implement a data pipeline on AWS independently.What you will learn Understand data engineering concepts and emerging technologies Ingest streaming data with Amazon Kinesis Data Firehose Optimize, denormalize, and join datasets with AWS Glue Studio Use Amazon S3 events to trigger a Lambda process to transform a file Run complex SQL queries on data lake data using Amazon Athena Load data into a Redshift data warehouse and run queries Create a visualization of your data using Amazon QuickSight Extract sentiment data from a dataset using Amazon Comprehend Who this book is for This book is for data engineers, data analysts, and data architects who are new to AWS and looking to extend their skills to the AWS cloud. Anyone new to data engineering who wants to learn about the foundational concepts while gaining practical experience with common data engineering services on AWS will also find this book useful. A basic understanding of big data-related topics and Python coding will help you get the most out of this book but it’s not a prerequisite. Familiarity with the AWS console and core services will also help you follow along.




Data Engineering with Databricks Cookbook


Book Description

Work through 70 recipes for implementing reliable data pipelines with Apache Spark, optimally store and process structured and unstructured data in Delta Lake, and use Databricks to orchestrate and govern your data Key Features Learn data ingestion, data transformation, and data management techniques using Apache Spark and Delta Lake Gain practical guidance on using Delta Lake tables and orchestrating data pipelines Implement reliable DataOps and DevOps practices, and enforce data governance policies on Databricks Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionWritten by a Senior Solutions Architect at Databricks, Data Engineering with Databricks Cookbook will show you how to effectively use Apache Spark, Delta Lake, and Databricks for data engineering, starting with comprehensive introduction to data ingestion and loading with Apache Spark. What makes this book unique is its recipe-based approach, which will help you put your knowledge to use straight away and tackle common problems. You’ll be introduced to various data manipulation and data transformation solutions that can be applied to data, find out how to manage and optimize Delta tables, and get to grips with ingesting and processing streaming data. The book will also show you how to improve the performance problems of Apache Spark apps and Delta Lake. Advanced recipes later in the book will teach you how to use Databricks to implement DataOps and DevOps practices, as well as how to orchestrate and schedule data pipelines using Databricks Workflows. You’ll also go through the full process of setup and configuration of the Unity Catalog for data governance. By the end of this book, you’ll be well-versed in building reliable and scalable data pipelines using modern data engineering technologies.What you will learn Perform data loading, ingestion, and processing with Apache Spark Discover data transformation techniques and custom user-defined functions (UDFs) in Apache Spark Manage and optimize Delta tables with Apache Spark and Delta Lake APIs Use Spark Structured Streaming for real-time data processing Optimize Apache Spark application and Delta table query performance Implement DataOps and DevOps practices on Databricks Orchestrate data pipelines with Delta Live Tables and Databricks Workflows Implement data governance policies with Unity Catalog Who this book is for This book is for data engineers, data scientists, and data practitioners who want to learn how to build efficient and scalable data pipelines using Apache Spark, Delta Lake, and Databricks. To get the most out of this book, you should have basic knowledge of data architecture, SQL, and Python programming.




Machine Learning Engineering on AWS


Book Description

Work seamlessly with production-ready machine learning systems and pipelines on AWS by addressing key pain points encountered in the ML life cycle Key FeaturesGain practical knowledge of managing ML workloads on AWS using Amazon SageMaker, Amazon EKS, and moreUse container and serverless services to solve a variety of ML engineering requirementsDesign, build, and secure automated MLOps pipelines and workflows on AWSBook Description There is a growing need for professionals with experience in working on machine learning (ML) engineering requirements as well as those with knowledge of automating complex MLOps pipelines in the cloud. This book explores a variety of AWS services, such as Amazon Elastic Kubernetes Service, AWS Glue, AWS Lambda, Amazon Redshift, and AWS Lake Formation, which ML practitioners can leverage to meet various data engineering and ML engineering requirements in production. This machine learning book covers the essential concepts as well as step-by-step instructions that are designed to help you get a solid understanding of how to manage and secure ML workloads in the cloud. As you progress through the chapters, you'll discover how to use several container and serverless solutions when training and deploying TensorFlow and PyTorch deep learning models on AWS. You'll also delve into proven cost optimization techniques as well as data privacy and model privacy preservation strategies in detail as you explore best practices when using each AWS. By the end of this AWS book, you'll be able to build, scale, and secure your own ML systems and pipelines, which will give you the experience and confidence needed to architect custom solutions using a variety of AWS services for ML engineering requirements. What you will learnFind out how to train and deploy TensorFlow and PyTorch models on AWSUse containers and serverless services for ML engineering requirementsDiscover how to set up a serverless data warehouse and data lake on AWSBuild automated end-to-end MLOps pipelines using a variety of servicesUse AWS Glue DataBrew and SageMaker Data Wrangler for data engineeringExplore different solutions for deploying deep learning models on AWSApply cost optimization techniques to ML environments and systemsPreserve data privacy and model privacy using a variety of techniquesWho this book is for This book is for machine learning engineers, data scientists, and AWS cloud engineers interested in working on production data engineering, machine learning engineering, and MLOps requirements using a variety of AWS services such as Amazon EC2, Amazon Elastic Kubernetes Service (EKS), Amazon SageMaker, AWS Glue, Amazon Redshift, AWS Lake Formation, and AWS Lambda -- all you need is an AWS account to get started. Prior knowledge of AWS, machine learning, and the Python programming language will help you to grasp the concepts covered in this book more effectively.




Azure Data Factory Cookbook


Book Description

Data Engineers guide to solve real-world problems encountered while building and transforming data pipelines using Azure's data integration tool Key Features Solve real-world data problems and create data-driven workflows with ease using Azure Data Factory Build an ADF pipeline that operates on pre-built ML model and Azure AI Get up and running with Fabric Data Explorer and extend ADF with Logic Apps and Azure functions Book DescriptionThis new edition of the Azure Data Factory book, fully updated to reflect ADS V2, will help you get up and running by showing you how to create and execute your first job in ADF. There are updated and new recipes throughout the book based on developments happening in Azure Synapse, Deployment with Azure DevOps, and Azure Purview. The current edition also runs you through Fabric Data Factory, Data Explorer, and some industry-grade best practices with specific chapters on each. You’ll learn how to branch and chain activities, create custom activities, and schedule pipelines, as well as discover the benefits of cloud data warehousing, Azure Synapse Analytics, and Azure Data Lake Gen2 Storage. With practical recipes, you’ll learn how to actively engage with analytical tools from Azure Data Services and leverage your on-premises infrastructure with cloud-native tools to get relevant business insights. You'll familiarize yourself with the common errors that you may encounter while working with ADF and find out the solutions to them. You’ll also understand error messages and resolve problems in connectors and data flows with the debugging capabilities of ADF. By the end of this book, you’ll be able to use ADF with its latest advancements as the main ETL and orchestration tool for your data warehouse projects.What you will learn Build and Manage data pipelines with ease using the latest version of ADF Configure, load data, and operate data flows with Azure Synapse Get up and running with Fabric Data Factory Working with Azure Data Factory and Azure Purview Create big data pipelines using Databricks and Delta tables Integrate ADF with commonly used Azure services such as Azure ML, Azure Logic Apps, and Azure Functions Learn industry-grade best practices for using Azure Data Factory Who this book is for This book is for ETL developers, data warehouse and ETL architects, software professionals, and anyone else who wants to learn about the common and not-so-common challenges faced while developing traditional and hybrid ETL solutions using Microsoft's Azure Data Factory. You’ll also find this book useful if you are looking for recipes to improve or enhance your existing ETL pipelines. Basic knowledge of data warehousing is a prerequisite.




Python Data Cleaning Cookbook


Book Description

Learn the intricacies of data description, issue identification, and practical problem-solving, armed with essential techniques and expert tips. Key Features Get to grips with new techniques for data preprocessing and cleaning for machine learning and NLP models Use new and updated AI tools and techniques for data cleaning tasks Clean, monitor, and validate large data volumes to diagnose problems using cutting-edge methodologies including Machine learning and AI Book DescriptionJumping into data analysis without proper data cleaning will certainly lead to incorrect results. The Python Data Cleaning Cookbook - Second Edition will show you tools and techniques for cleaning and handling data with Python for better outcomes. Fully updated to the latest version of Python and all relevant tools, this book will teach you how to manipulate and clean data to get it into a useful form. he current edition focuses on advanced techniques like machine learning and AI-specific approaches and tools for data cleaning along with the conventional ones. The book also delves into tips and techniques to process and clean data for ML, AI, and NLP models. You will learn how to filter and summarize data to gain insights and better understand what makes sense and what does not, along with discovering how to operate on data to address the issues you've identified. Next, you’ll cover recipes for using supervised learning and Naive Bayes analysis to identify unexpected values and classification errors and generate visualizations for exploratory data analysis (EDA) to identify unexpected values. Finally, you’ll build functions and classes that you can reuse without modification when you have new data. By the end of this Data Cleaning book, you'll know how to clean data and diagnose problems within it.What you will learn Using OpenAI tools for various data cleaning tasks Producing summaries of the attributes of datasets, columns, and rows Anticipating data-cleaning issues when importing tabular data into pandas Applying validation techniques for imported tabular data Improving your productivity in pandas by using method chaining Recognizing and resolving common issues like dates and IDs Setting up indexes to streamline data issue identification Using data cleaning to prepare your data for ML and AI models Who this book is for This book is for anyone looking for ways to handle messy, duplicate, and poor data using different Python tools and techniques. The book takes a recipe-based approach to help you to learn how to clean and manage data with practical examples. Working knowledge of Python programming is all you need to get the most out of the book.




Amazon Redshift Cookbook


Book Description

Discover how to build a cloud-based data warehouse at petabyte-scale that is burstable and built to scale for end-to-end analytical solutions Key FeaturesDiscover how to translate familiar data warehousing concepts into Redshift implementationUse impressive Redshift features to optimize development, productionizing, and operations processesFind out how to use advanced features such as concurrency scaling, Redshift Spectrum, and federated queriesBook Description Amazon Redshift is a fully managed, petabyte-scale AWS cloud data warehousing service. It enables you to build new data warehouse workloads on AWS and migrate on-premises traditional data warehousing platforms to Redshift. This book on Amazon Redshift starts by focusing on Redshift architecture, showing you how to perform database administration tasks on Redshift. You'll then learn how to optimize your data warehouse to quickly execute complex analytic queries against very large datasets. Because of the massive amount of data involved in data warehousing, designing your database for analytical processing lets you take full advantage of Redshift's columnar architecture and managed services. As you advance, you'll discover how to deploy fully automated and highly scalable extract, transform, and load (ETL) processes, which help minimize the operational efforts that you have to invest in managing regular ETL pipelines and ensure the timely and accurate refreshing of your data warehouse. Finally, you'll gain a clear understanding of Redshift use cases, data ingestion, data management, security, and scaling so that you can build a scalable data warehouse platform. By the end of this Redshift book, you'll be able to implement a Redshift-based data analytics solution and have understood the best practice solutions to commonly faced problems. What you will learnUse Amazon Redshift to build petabyte-scale data warehouses that are agile at scaleIntegrate your data warehousing solution with a data lake using purpose-built features and services on AWSBuild end-to-end analytical solutions from data sourcing to consumption with the help of useful recipesLeverage Redshift's comprehensive security capabilities to meet the most demanding business requirementsFocus on architectural insights and rationale when using analytical recipesDiscover best practices for working with big data to operate a fully managed solutionWho this book is for This book is for anyone involved in architecting, implementing, and optimizing an Amazon Redshift data warehouse, such as data warehouse developers, data analysts, database administrators, data engineers, and data scientists. Basic knowledge of data warehousing, database systems, and cloud concepts and familiarity with Redshift will be beneficial.




AWS Cookbook


Book Description

This practical guide provides over 70 self-contained recipes to help you creatively solve common AWS challenges you'll encounter on your cloud journey. If you're comfortable with rudimentary scripting and general cloud concepts, this cookbook provides what you need to address foundational tasks and create high-level capabilities. Authors John Culkin and Mike Zazon share real-world examples that incorporate best practices. Each recipe includes a diagram to visualize the components. Code is provided so that you can safely execute in an AWS account to ensure solutions work as described. From there, you can customize the code to help construct an application or fix an existing problem. Each recipe also includes a discussion to provide context, explain the approach, and challenge you to explore the possibilities further. Go beyond theory and learn the details you need to successfully build on AWS. The recipes help you: Redact personal identifiable information (PII) from text using Amazon Comprehend Automate password rotation for Amazon RDS databases Use VPC Reachability Analyzer to verify and troubleshoot network paths Lock down Amazon Simple Storage Service (S3) buckets Analyze AWS Identity and Access Management policies Autoscale a containerized service




Tableau 2019.x Cookbook


Book Description

Perform advanced dashboard, visualization, and analytical techniques with Tableau Desktop, Tableau Prep, and Tableau Server Key FeaturesUnique problem-solution approach to aid effective business decision-makingCreate interactive dashboards and implement powerful business intelligence solutionsIncludes best practices on using Tableau with modern cloud analytics servicesBook Description Tableau has been one of the most popular business intelligence solutions in recent times, thanks to its powerful and interactive data visualization capabilities. Tableau 2019.x Cookbook is full of useful recipes from industry experts, who will help you master Tableau skills and learn each aspect of Tableau's ecosystem. This book is enriched with features such as Tableau extracts, Tableau advanced calculations, geospatial analysis, and building dashboards. It will guide you with exciting data manipulation, storytelling, advanced filtering, expert visualization, and forecasting techniques using real-world examples. From basic functionalities of Tableau to complex deployment on Linux, you will cover it all. Moreover, you will learn advanced features of Tableau using R, Python, and various APIs. You will learn how to prepare data for analysis using the latest Tableau Prep. In the concluding chapters, you will learn how Tableau fits the modern world of analytics and works with modern data platforms such as Snowflake and Redshift. In addition, you will learn about the best practices of integrating Tableau with ETL using Matillion ETL. By the end of the book, you will be ready to tackle business intelligence challenges using Tableau's features. What you will learnUnderstand the basic and advanced skills of Tableau DesktopImplement best practices of visualization, dashboard, and storytellingLearn advanced analytics with the use of build in statisticsDeploy the multi-node server on Linux and WindowsUse Tableau with big data sources such as Hadoop, Athena, and SpectrumCover Tableau built-in functions for forecasting using R packagesCombine, shape, and clean data for analysis using Tableau PrepExtend Tableau’s functionalities with REST API and R/PythonWho this book is for Tableau 2019.x Cookbook is for data analysts, data engineers, BI developers, and users who are looking for quick solutions to common and not-so-common problems faced while using Tableau products. Put each recipe into practice by bringing the latest offerings of Tableau 2019.x to solve real-world analytics and business intelligence challenges. Some understanding of BI concepts and Tableau is required.




Python Natural Language Processing Cookbook


Book Description

Updated to include three new chapters on transformers, natural language understanding (NLU) with explainable AI, and dabbling with popular LLMs from Hugging Face and OpenAI Key Features Leverage ready-to-use recipes with the latest LLMs, including Mistral, Llama, and OpenAI models Use LLM-powered agents for custom tasks and real-world interactions Gain practical, in-depth knowledge of transformers and their role in implementing various NLP tasks with open-source and advanced LLMs Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionHarness the power of Natural Language Processing to overcome real-world text analysis challenges with this recipe-based roadmap written by two seasoned NLP experts with vast experience transforming various industries with their NLP prowess. You’ll be able to make the most of the latest NLP advancements, including large language models (LLMs), and leverage their capabilities through Hugging Face transformers. Through a series of hands-on recipes, you’ll master essential techniques such as extracting entities and visualizing text data. The authors will expertly guide you through building pipelines for sentiment analysis, topic modeling, and question-answering using popular libraries like spaCy, Gensim, and NLTK. You’ll also learn to implement RAG pipelines to draw out precise answers from a text corpus using LLMs. This second edition expands your skillset with new chapters on cutting-edge LLMs like GPT-4, Natural Language Understanding (NLU), and Explainable AI (XAI)—fostering trust and transparency in your NLP models. By the end of this book, you'll be equipped with the skills to apply advanced text processing techniques, use pre-trained transformer models, build custom NLP pipelines to extract valuable insights from text data to drive informed decision-making.What you will learn Understand fundamental NLP concepts along with their applications using examples in Python Classify text quickly and accurately with rule-based and supervised methods Train NER models and perform sentiment analysis to identify entities and emotions in text Explore topic modeling and text visualization to reveal themes and relationships within text Leverage Hugging Face and OpenAI LLMs to perform advanced NLP tasks Use question-answering techniques to handle both open and closed domains Apply XAI techniques to better understand your model predictions Who this book is for This updated edition of the Python Natural Language Processing Cookbook is for data scientists, machine learning engineers, and developers with a background in Python. Whether you’re looking to learn NLP techniques, extract valuable insights from textual data, or create foundational applications, this book will equip you with basic to intermediate skills. No prior NLP knowledge is necessary to get started. All you need is familiarity with basic programming principles. For seasoned developers, the updated sections offer the latest on transformers, explainable AI, and Generative AI with LLMs.