Pretrain Vision and Large Language Models in Python


Book Description

Master the art of training vision and large language models with conceptual fundaments and industry-expert guidance. Learn about AWS services and design patterns, with relevant coding examples Key Features Learn to develop, train, tune, and apply foundation models with optimized end-to-end pipelines Explore large-scale distributed training for models and datasets with AWS and SageMaker examples Evaluate, deploy, and operationalize your custom models with bias detection and pipeline monitoring Book Description Foundation models have forever changed machine learning. From BERT to ChatGPT, CLIP to Stable Diffusion, when billions of parameters are combined with large datasets and hundreds to thousands of GPUs, the result is nothing short of record-breaking. The recommendations, advice, and code samples in this book will help you pretrain and fine-tune your own foundation models from scratch on AWS and Amazon SageMaker, while applying them to hundreds of use cases across your organization. With advice from seasoned AWS and machine learning expert Emily Webber, this book helps you learn everything you need to go from project ideation to dataset preparation, training, evaluation, and deployment for large language, vision, and multimodal models. With step-by-step explanations of essential concepts and practical examples, you'll go from mastering the concept of pretraining to preparing your dataset and model, configuring your environment, training, fine-tuning, evaluating, deploying, and optimizing your foundation models. You will learn how to apply the scaling laws to distributing your model and dataset over multiple GPUs, remove bias, achieve high throughput, and build deployment pipelines. By the end of this book, you'll be well equipped to embark on your own project to pretrain and fine-tune the foundation models of the future. What you will learn Find the right use cases and datasets for pretraining and fine-tuning Prepare for large-scale training with custom accelerators and GPUs Configure environments on AWS and SageMaker to maximize performance Select hyperparameters based on your model and constraints Distribute your model and dataset using many types of parallelism Avoid pitfalls with job restarts, intermittent health checks, and more Evaluate your model with quantitative and qualitative insights Deploy your models with runtime improvements and monitoring pipelines Who this book is for If you're a machine learning researcher or enthusiast who wants to start a foundation modelling project, this book is for you. Applied scientists, data scientists, machine learning engineers, solution architects, product managers, and students will all benefit from this book. Intermediate Python is a must, along with introductory concepts of cloud computing. A strong understanding of deep learning fundamentals is needed, while advanced topics will be explained. The content covers advanced machine learning and cloud techniques, explaining them in an actionable, easy-to-understand way.




Learn Python Generative AI


Book Description

Learn to unleash the power of AI creativity KEY FEATURES ● Understand the core concepts related to generative AI. ● Different types of generative models and their applications. ● Learn how to design generative AI neural networks using Python and TensorFlow. DESCRIPTION This book researches the intricate world of generative Artificial Intelligence, offering readers an extensive understanding of various components and applications in this field. The book begins with an in-depth analysis of generative models, providing a solid foundation and exploring their combination nuances. It then focuses on enhancing TransVAE, a variational autoencoder, and introduces the Swin Transformer in generative AI. The inclusion of cutting edge applications like building an image search using Pinecone and a vector database further enriches its content. The narrative shifts to practical applications, showcasing GenAI's impact in healthcare, retail, and finance, with real-world examples and innovative solutions. In the healthcare sector, it emphasizes AI's transformative role in diagnostics and patient care. In retail and finance, it illustrates how AI revolutionizes customer engagement and decision making. The book concludes by synthesizing key learnings, offering insights into the future of generative AI, and making it a comprehensive guide for diverse industries. Readers will find themselves equipped with a profound understanding of generative AI, its current applications, and its boundless potential for future innovations. WHAT YOU WILL LEARN ● Acquire practical skills in designing and implementing various generative AI models. ● Gain expertise in vector databases and image embeddings, crucial for image search and data retrieval. ● Navigate challenges in healthcare, retail, and finance using sector specific insights. ● Generate images and text with VAEs, GANs, LLMs, and vector databases. ● Focus on both traditional and cutting edge techniques in generative AI. WHO THIS BOOK IS FOR This book is for current and aspiring emerging AI deep learning professionals, architects, students, and anyone who is starting and learning a rewarding career in generative AI. TABLE OF CONTENTS 1. Introducing Generative AI 2. Designing Generative Adversarial Networks 3. Training and Developing Generative Adversarial Networks 4. Architecting Auto Encoder for Generative AI 5. Building and Training Generative Autoencoders 6. Designing Generative Variation Auto Encoder 7. Building Variational Autoencoders for Generative AI 8. Fundamental of Designing New Age Generative Vision Transformer 9. Implementing Generative Vision Transformer 10. Architectural Refactoring for Generative Modeling 11. Major Technical Roadblocks in Generative AI and Way Forward 12. Overview and Application of Generative AI Models 13. Key Learnings




Introducing MLOps


Book Description

More than half of the analytics and machine learning (ML) models created by organizations today never make it into production. Some of the challenges and barriers to operationalization are technical, but others are organizational. Either way, the bottom line is that models not in production can't provide business impact. This book introduces the key concepts of MLOps to help data scientists and application engineers not only operationalize ML models to drive real business change but also maintain and improve those models over time. Through lessons based on numerous MLOps applications around the world, nine experts in machine learning provide insights into the five steps of the model life cycle--Build, Preproduction, Deployment, Monitoring, and Governance--uncovering how robust MLOps processes can be infused throughout. This book helps you: Fulfill data science value by reducing friction throughout ML pipelines and workflows Refine ML models through retraining, periodic tuning, and complete remodeling to ensure long-term accuracy Design the MLOps life cycle to minimize organizational risks with models that are unbiased, fair, and explainable Operationalize ML models for pipeline deployment and for external business systems that are more complex and less standardized




Practical MLOps


Book Description

Getting your models into production is the fundamental challenge of machine learning. MLOps offers a set of proven principles aimed at solving this problem in a reliable and automated way. This insightful guide takes you through what MLOps is (and how it differs from DevOps) and shows you how to put it into practice to operationalize your machine learning models. Current and aspiring machine learning engineers--or anyone familiar with data science and Python--will build a foundation in MLOps tools and methods (along with AutoML and monitoring and logging), then learn how to implement them in AWS, Microsoft Azure, and Google Cloud. The faster you deliver a machine learning system that works, the faster you can focus on the business problems you're trying to crack. This book gives you a head start. You'll discover how to: Apply DevOps best practices to machine learning Build production machine learning systems and maintain them Monitor, instrument, load-test, and operationalize machine learning systems Choose the correct MLOps tools for a given machine learning task Run machine learning models on a variety of platforms and devices, including mobile phones and specialized hardware




Machine Learning Design Patterns


Book Description

The design patterns in this book capture best practices and solutions to recurring problems in machine learning. The authors, three Google engineers, catalog proven methods to help data scientists tackle common problems throughout the ML process. These design patterns codify the experience of hundreds of experts into straightforward, approachable advice. In this book, you will find detailed explanations of 30 patterns for data and problem representation, operationalization, repeatability, reproducibility, flexibility, explainability, and fairness. Each pattern includes a description of the problem, a variety of potential solutions, and recommendations for choosing the best technique for your situation. You'll learn how to: Identify and mitigate common challenges when training, evaluating, and deploying ML models Represent data for different ML model types, including embeddings, feature crosses, and more Choose the right model type for specific problems Build a robust training loop that uses checkpoints, distribution strategy, and hyperparameter tuning Deploy scalable ML systems that you can retrain and update to reflect new data Interpret model predictions for stakeholders and ensure models are treating users fairly




Data Science on AWS


Book Description

With this practical book, AI and machine learning practitioners will learn how to successfully build and deploy data science projects on Amazon Web Services. The Amazon AI and machine learning stack unifies data science, data engineering, and application development to help level upyour skills. This guide shows you how to build and run pipelines in the cloud, then integrate the results into applications in minutes instead of days. Throughout the book, authors Chris Fregly and Antje Barth demonstrate how to reduce cost and improve performance. Apply the Amazon AI and ML stack to real-world use cases for natural language processing, computer vision, fraud detection, conversational devices, and more Use automated machine learning to implement a specific subset of use cases with SageMaker Autopilot Dive deep into the complete model development lifecycle for a BERT-based NLP use case including data ingestion, analysis, model training, and deployment Tie everything together into a repeatable machine learning operations pipeline Explore real-time ML, anomaly detection, and streaming analytics on data streams with Amazon Kinesis and Managed Streaming for Apache Kafka Learn security best practices for data science projects and workflows including identity and access management, authentication, authorization, and more




Practitioner’s Guide to Data Science


Book Description

Covers Data Science concepts, processes, and the real-world hands-on use cases. KEY FEATURES ● Covers the journey from a basic programmer to an effective Data Science developer. ● Applied use of Data Science native processes like CRISP-DM and Microsoft TDSP. ● Implementation of MLOps using Microsoft Azure DevOps. DESCRIPTION "How is the Data Science project to be implemented?" has never been more conceptually sounding, thanks to the work presented in this book. This book provides an in-depth look at the current state of the world's data and how Data Science plays a pivotal role in everything we do. This book explains and implements the entire Data Science lifecycle using well-known data science processes like CRISP-DM and Microsoft TDSP. The book explains the significance of these processes in connection with the high failure rate of Data Science projects. The book helps build a solid foundation in Data Science concepts and related frameworks. It teaches how to implement real-world use cases using data from the HMDA dataset. It explains Azure ML Service architecture, its capabilities, and implementation to the DS team, who will then be prepared to implement MLOps. The book also explains how to use Azure DevOps to make the process repeatable while we're at it. By the end of this book, you will learn strong Python coding skills, gain a firm grasp of concepts such as feature engineering, create insightful visualizations and become acquainted with techniques for building machine learning models. WHAT YOU WILL LEARN ● Organize Data Science projects using CRISP-DM and Microsoft TDSP. ● Learn to acquire and explore data using Python visualizations. ● Get well versed with the implementation of data pre-processing and Feature Engineering. ● Understand algorithm selection, model development, and model evaluation. ● Hands-on with Azure ML Service, its architecture, and capabilities. ● Learn to use Azure ML SDK and MLOps for implementing real-world use cases. WHO THIS BOOK IS FOR This book is intended for programmers who wish to pursue AI/ML development and build a solid conceptual foundation and familiarity with related processes and frameworks. Additionally, this book is an excellent resource for Software Architects and Managers involved in the design and delivery of Data Science-based solutions. TABLE OF CONTENTS 1. Data Science for Business 2. Data Science Project Methodologies and Team Processes 3. Business Understanding and Its Data Landscape 4. Acquire, Explore, and Analyze Data 5. Pre-processing and Preparing Data 6. Developing a Machine Learning Model 7. Lap Around Azure ML Service 8. Deploying and Managing Models




TinyML


Book Description

Deep learning networks are getting smaller. Much smaller. The Google Assistant team can detect words with a model just 14 kilobytes in size—small enough to run on a microcontroller. With this practical book you’ll enter the field of TinyML, where deep learning and embedded systems combine to make astounding things possible with tiny devices. Pete Warden and Daniel Situnayake explain how you can train models small enough to fit into any environment. Ideal for software and hardware developers who want to build embedded systems using machine learning, this guide walks you through creating a series of TinyML projects, step-by-step. No machine learning or microcontroller experience is necessary. Build a speech recognizer, a camera that detects people, and a magic wand that responds to gestures Work with Arduino and ultra-low-power microcontrollers Learn the essentials of ML and how to train your own models Train models to understand audio, image, and accelerometer data Explore TensorFlow Lite for Microcontrollers, Google’s toolkit for TinyML Debug applications and provide safeguards for privacy and security Optimize latency, energy usage, and model and binary size




Generative AI in Action


Book Description

Generative AI can transform your business by streamlining the process of creating text, images, and code. This book will show you how to get in on the action! Generative AI in Action is the comprehensive and concrete guide to generative AI you’ve been searching for. It introduces both AI’s fundamental principles and its practical applications in an enterprise context—from generating text and images for product catalogs and marketing campaigns, to technical reporting, and even writing software. Inside, author Amit Bahree shares his experience leading Generative AI projects at Microsoft for nearly a decade, starting well before the current GPT revolution. Inside Generative AI in Action you will find: • A practical overview of of generative AI applications • Architectural patterns, integration guidance, and best practices for generative AI • The latest techniques like RAG, prompt engineering, and multi-modality • The challenges and risks of generative AI like hallucinations and jailbreaks • How to integrate generative AI into your business and IT strategy Generative AI in Action is full of real-world use cases for generative AI, showing you where and how to start integrating this powerful technology into your products and workflows. You’ll benefit from tried-and-tested implementation advice, as well as application architectures to deploy GenAI in production at enterprise scale. Purchase of the print book includes a free eBook in PDF and ePub formats from Manning Publications. About the technology In controlled environments, deep learning systems routinely surpass humans in reading comprehension, image recognition, and language understanding. Large Language Models (LLMs) can deliver similar results in text and image generation and predictive reasoning. Outside the lab, though, generative AI can both impress and fail spectacularly. So how do you get the results you want? Keep reading! About the book Generative AI in Action presents concrete examples, insights, and techniques for using LLMs and other modern AI technologies successfully and safely. In it, you’ll find practical approaches for incorporating AI into marketing, software development, business report generation, data storytelling, and other typically-human tasks. You’ll explore the emerging patterns for GenAI apps, master best practices for prompt engineering, and learn how to address hallucination, high operating costs, the rapid pace of change and other common problems. What's inside • Best practices for deploying Generative AI apps • Production-quality RAG • Adapting GenAI models to your specific domain About the reader For enterprise architects, developers, and data scientists interested in upgrading their architectures with generative AI. About the author Amit Bahree is Principal Group Product Manager for the Azure AI engineering team at Microsoft. The technical editor on this book was Wee Hyong Tok. Table of Contents Part 1 1 Introduction to generative AI 2 Introduction to large language models 3 Working through an API: Generating text 4 From pixels to pictures: Generating images 5 What else can AI generate? Part 2 6 Guide to prompt engineering 7 Retrieval-augmented generation: The secret weapon 8 Chatting with your data 9 Tailoring models with model adaptation and fine-tuning Part 3 10 Application architecture for generative AI apps 11 Scaling up: Best practices for production deployment 12 Evaluations and benchmarks 13 Guide to ethical GenAI: Principles, practices, and pitfalls A The book’s GitHub repository B Responsible AI tools




High Performance Computing


Book Description

This book constitutes the proceedings of the 9th Latin American Conference on High Performance Computing, CARLA 2022, held in Porto Alegre, Brazil, in September 2022. The 16 full papers presented in this volume were carefully reviewed and selected from 56 submissions. CARLA, the Latin American High Performance Computing Conference, is an international academic meeting aimed at providing a forum to foster the growth and strength of the High Performance Computing (HPC) community in Latin America and the Caribbean through the exchange and dissemination of new ideas, techniques, and research in HPC and its application areas.