Mastering Large Datasets with Python


Book Description

Summary Modern data science solutions need to be clean, easy to read, and scalable. In Mastering Large Datasets with Python, author J.T. Wolohan teaches you how to take a small project and scale it up using a functionally influenced approach to Python coding. You’ll explore methods and built-in Python tools that lend themselves to clarity and scalability, like the high-performing parallelism method, as well as distributed technologies that allow for high data throughput. The abundant hands-on exercises in this practical tutorial will lock in these essential skills for any large-scale data science project. Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications. About the technology Programming techniques that work well on laptop-sized data can slow to a crawl—or fail altogether—when applied to massive files or distributed datasets. By mastering the powerful map and reduce paradigm, along with the Python-based tools that support it, you can write data-centric applications that scale efficiently without requiring codebase rewrites as your requirements change. About the book Mastering Large Datasets with Python teaches you to write code that can handle datasets of any size. You’ll start with laptop-sized datasets that teach you to parallelize data analysis by breaking large tasks into smaller ones that can run simultaneously. You’ll then scale those same programs to industrial-sized datasets on a cluster of cloud servers. With the map and reduce paradigm firmly in place, you’ll explore tools like Hadoop and PySpark to efficiently process massive distributed datasets, speed up decision-making with machine learning, and simplify your data storage with AWS S3. What's inside An introduction to the map and reduce paradigm Parallelization with the multiprocessing module and pathos framework Hadoop and Spark for distributed computing Running AWS jobs to process large datasets About the reader For Python programmers who need to work faster with more data. About the author J. T. Wolohan is a lead data scientist at Booz Allen Hamilton, and a PhD researcher at Indiana University, Bloomington. Table of Contents: PART 1 1 ¦ Introduction 2 ¦ Accelerating large dataset work: Map and parallel computing 3 ¦ Function pipelines for mapping complex transformations 4 ¦ Processing large datasets with lazy workflows 5 ¦ Accumulation operations with reduce 6 ¦ Speeding up map and reduce with advanced parallelization PART 2 7 ¦ Processing truly big datasets with Hadoop and Spark 8 ¦ Best practices for large data with Apache Streaming and mrjob 9 ¦ PageRank with map and reduce in PySpark 10 ¦ Faster decision-making with machine learning and PySpark PART 3 11 ¦ Large datasets in the cloud with Amazon Web Services and S3 12 ¦ MapReduce in the cloud with Amazon’s Elastic MapReduce




Mastering Large Datasets


Book Description

With an emphasis on clarity, style, and performance, author J.T. Wolohan expertly guides you through implementing a functionally-influenced approach to Python coding. You'll get familiar with Python's functional built-ins like the functools operator and itertools modules, as well as the toolz library. Mastering Large Datasets teaches you to write easily readable, easily scalable Python code that can efficiently process large volumes of structured and unstructured data. By the end of this comprehensive guide, you'll have a solid grasp on the tools and methods that will take your code beyond the laptop and your data science career to the next level! Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications.




Mastering Big Data


Book Description

Cybellium Ltd is dedicated to empowering individuals and organizations with the knowledge and skills they need to navigate the ever-evolving computer science landscape securely and learn only the latest information available on any subject in the category of computer science including: - Information Technology (IT) - Cyber Security - Information Security - Big Data - Artificial Intelligence (AI) - Engineering - Robotics - Standards and compliance Our mission is to be at the forefront of computer science education, offering a wide and comprehensive range of resources, including books, courses, classes and training programs, tailored to meet the diverse needs of any subject in computer science. Visit https://www.cybellium.com for more books.




Mastering Large Language Models


Book Description

Do not just talk AI, build it: Your guide to LLM application development KEY FEATURES ● Explore NLP basics and LLM fundamentals, including essentials, challenges, and model types. ● Learn data handling and pre-processing techniques for efficient data management. ● Understand neural networks overview, including NN basics, RNNs, CNNs, and transformers. ● Strategies and examples for harnessing LLMs. DESCRIPTION Transform your business landscape with the formidable prowess of large language models (LLMs). The book provides you with practical insights, guiding you through conceiving, designing, and implementing impactful LLM-driven applications. This book explores NLP fundamentals like applications, evolution, components and language models. It teaches data pre-processing, neural networks , and specific architectures like RNNs, CNNs, and transformers. It tackles training challenges, advanced techniques such as GANs, meta-learning, and introduces top LLM models like GPT-3 and BERT. It also covers prompt engineering. Finally, it showcases LLM applications and emphasizes responsible development and deployment. With this book as your compass, you will navigate the ever-evolving landscape of LLM technology, staying ahead of the curve with the latest advancements and industry best practices. WHAT YOU WILL LEARN ● Grasp fundamentals of natural language processing (NLP) applications. ● Explore advanced architectures like transformers and their applications. ● Master techniques for training large language models effectively. ● Implement advanced strategies, such as meta-learning and self-supervised learning. ● Learn practical steps to build custom language model applications. WHO THIS BOOK IS FOR This book is tailored for those aiming to master large language models, including seasoned researchers, data scientists, developers, and practitioners in natural language processing (NLP). TABLE OF CONTENTS 1. Fundamentals of Natural Language Processing 2. Introduction to Language Models 3. Data Collection and Pre-processing for Language Modeling 4. Neural Networks in Language Modeling 5. Neural Network Architectures for Language Modeling 6. Transformer-based Models for Language Modeling 7. Training Large Language Models 8. Advanced Techniques for Language Modeling 9. Top Large Language Models 10. Building First LLM App 11. Applications of LLMs 12. Ethical Considerations 13. Prompt Engineering 14. Future of LLMs and Its Impact




Extending Power BI with Python and R


Book Description

Perform more advanced analysis and manipulation of your data beyond what Power BI can do to unlock valuable insights using Python and R Key FeaturesGet the most out of Python and R with Power BI by implementing non-trivial codeLeverage the toolset of Python and R chunks to inject scripts into your Power BI dashboardsImplement new techniques for ingesting, enriching, and visualizing data with Python and R in Power BIBook Description Python and R allow you to extend Power BI capabilities to simplify ingestion and transformation activities, enhance dashboards, and highlight insights. With this book, you'll be able to make your artifacts far more interesting and rich in insights using analytical languages. You'll start by learning how to configure your Power BI environment to use your Python and R scripts. The book then explores data ingestion and data transformation extensions, and advances to focus on data augmentation and data visualization. You'll understand how to import data from external sources and transform them using complex algorithms. The book helps you implement personal data de-identification methods such as pseudonymization, anonymization, and masking in Power BI. You'll be able to call external APIs to enrich your data much more quickly using Python programming and R programming. Later, you'll learn advanced Python and R techniques to perform in-depth analysis and extract valuable information using statistics and machine learning. You'll also understand the main statistical features of datasets by plotting multiple visual graphs in the process of creating a machine learning model. By the end of this book, you'll be able to enrich your Power BI data models and visualizations using complex algorithms in Python and R. What you will learnDiscover best practices for using Python and R in Power BI productsUse Python and R to perform complex data manipulations in Power BIApply data anonymization and data pseudonymization in Power BILog data and load large datasets in Power BI using Python and REnrich your Power BI dashboards using external APIs and machine learning modelsExtract insights from your data using linear optimization and other algorithmsHandle outliers and missing values for multivariate and time-series dataCreate any visualization, as complex as you want, using R scriptsWho this book is for This book is for business analysts, business intelligence professionals, and data scientists who already use Microsoft Power BI and want to add more value to their analysis using Python and R. Working knowledge of Power BI is required to make the most of this book. Basic knowledge of Python and R will also be helpful.




Mastering Large Language Models with Python


Book Description

A Comprehensive Guide to Leverage Generative AI in the Modern Enterprise KEY FEATURES ● Gain a comprehensive understanding of LLMs within the framework of Generative AI, from foundational concepts to advanced applications. ● Dive into practical exercises and real-world applications, accompanied by detailed code walkthroughs in Python. ● Explore LLMOps with a dedicated focus on ensuring trustworthy AI and best practices for deploying, managing, and maintaining LLMs in enterprise settings. ● Prioritize the ethical and responsible use of LLMs, with an emphasis on building models that adhere to principles of fairness, transparency, and accountability, fostering trust in AI technologies. DESCRIPTION “Mastering Large Language Models with Python” is an indispensable resource that offers a comprehensive exploration of Large Language Models (LLMs), providing the essential knowledge to leverage these transformative AI models effectively. From unraveling the intricacies of LLM architecture to practical applications like code generation and AI-driven recommendation systems, readers will gain valuable insights into implementing LLMs in diverse projects. Covering both open-source and proprietary LLMs, the book delves into foundational concepts and advanced techniques, empowering professionals to harness the full potential of these models. Detailed discussions on quantization techniques for efficient deployment, operational strategies with LLMOps, and ethical considerations ensure a well-rounded understanding of LLM implementation. Through real-world case studies, code snippets, and practical examples, readers will navigate the complexities of LLMs with confidence, paving the way for innovative solutions and organizational growth. Whether you seek to deepen your understanding, drive impactful applications, or lead AI-driven initiatives, this book equips you with the tools and insights needed to excel in the dynamic landscape of artificial intelligence. WHAT WILL YOU LEARN ● In-depth study of LLM architecture and its versatile applications across industries. ● Harness open-source and proprietary LLMs to craft innovative solutions. ● Implement LLM APIs for a wide range of tasks spanning natural language processing, audio analysis, and visual recognition. ● Optimize LLM deployment through techniques such as quantization and operational strategies like LLMOps, ensuring efficient and scalable model usage. ● Master prompt engineering techniques to fine-tune LLM outputs, enhancing quality and relevance for diverse use cases. ● Navigate the complex landscape of ethical AI development, prioritizing responsible practices to drive impactful technology adoption and advancement. WHO IS THIS BOOK FOR? This book is tailored for software engineers, data scientists, AI researchers, and technology leaders with a foundational understanding of machine learning concepts and programming. It's ideal for those looking to deepen their knowledge of Large Language Models and their practical applications in the field of AI. If you aim to explore LLMs extensively for implementing inventive solutions or spearheading AI-driven projects, this book is tailored to your needs. TABLE OF CONTENTS 1. The Basics of Large Language Models and Their Applications 2. Demystifying Open-Source Large Language Models 3. Closed-Source Large Language Models 4. LLM APIs for Various Large Language Model Tasks 5. Integrating Cohere API in Google Sheets 6. Dynamic Movie Recommendation Engine Using LLMs 7. Document-and Web-based QA Bots with Large Language Models 8. LLM Quantization Techniques and Implementation 9. Fine-tuning and Evaluation of LLMs 10. Recipes for Fine-Tuning and Evaluating LLMs 11. LLMOps - Operationalizing LLMs at Scale 12. Implementing LLMOps in Practice Using MLflow on Databricks 13. Mastering the Art of Prompt Engineering 14. Prompt Engineering Essentials and Design Patterns 15. Ethical Considerations and Regulatory Frameworks for LLMs 16. Towards Trustworthy Generative AI (A Novel Framework Inspired by Symbolic Reasoning) Index




Mastering .NET Machine Learning


Book Description

Master the art of machine learning with .NET and gain insight into real-world applications About This Book Based on .NET framework 4.6.1, includes examples on ASP.NET Core 1.0 Set up your business application to start using machine learning techniques Familiarize the user with some of the more common .NET libraries for machine learning Implement several common machine learning techniques Evaluate, optimize and adjust machine learning models Who This Book Is For This book is targeted at .Net developers who want to build complex machine learning systems. Some basic understanding of data science is required. What You Will Learn Write your own machine learning applications and experiments using the latest .NET framework, including .NET Core 1.0 Set up your business application to start using machine learning. Accurately predict the future using regressions. Discover hidden patterns using decision trees. Acquire, prepare, and combine datasets to drive insights. Optimize business throughput using Bayes Classifier. Discover (more) hidden patterns using KNN and Naive Bayes. Discover (even more) hidden patterns using K-Means and PCA. Use Neural Networks to improve business decision making while using the latest ASP.NET technologies. Explore “Big Data”, distributed computing, and how to deploy machine learning models to IoT devices – making machines self-learning and adapting Along the way, learn about Open Data, Bing maps, and MBrace In Detail .Net is one of the widely used platforms for developing applications. With the meteoric rise of Machine learning, developers are now keen on finding out how can they make their .Net applications smarter. Also, .NET developers are interested into moving into the world of devices and how to apply machine learning techniques to, well, machines. This book is packed with real-world examples to easily use machine learning techniques in your business applications. You will begin with introduction to F# and prepare yourselves for machine learning using .NET framework. You will be writing a simple linear regression model using an example which predicts sales of a product. Forming a base with the regression model, you will start using machine learning libraries available in .NET framework such as Math.NET, Numl.NET and Accord.NET with the help of a sample application. You will then move on to writing multiple linear regressions and logistic regressions. You will learn what is open data and the awesomeness of type providers. Next, you are going to address some of the issues that we have been glossing over so far and take a deep dive into obtaining, cleaning, and organizing our data. You will compare the utility of building a KNN and Naive Bayes model to achieve best possible results. Implementation of Kmeans and PCA using Accord.NET and Numl.NET libraries is covered with the help of an example application. We will then look at many of issues confronting creating real-world machine learning models like overfitting and how to combat them using confusion matrixes, scaling, normalization, and feature selection. You will now enter into the world of Neural Networks and move your line of business application to a hybrid scientific application. After you have covered all the above machine learning models, you will see how to deal with very large datasets using MBrace and how to deploy machine learning models to Internet of Thing (IoT) devices so that the machine can learn and adapt on the fly Style and approach This book will guide you in learning everything about how to tackle the flood of data being encountered these days in your .NET applications with the help of popular machine learning libraries offered by the .NET framework.




Algorithms and Data Structures for Massive Datasets


Book Description

Massive modern datasets make traditional data structures and algorithms grind to a halt. This fun and practical guide introduces cutting-edge techniques that can reliably handle even the largest distributed datasets. In Algorithms and Data Structures for Massive Datasets you will learn: Probabilistic sketching data structures for practical problems Choosing the right database engine for your application Evaluating and designing efficient on-disk data structures and algorithms Understanding the algorithmic trade-offs involved in massive-scale systems Deriving basic statistics from streaming data Correctly sampling streaming data Computing percentiles with limited space resources Algorithms and Data Structures for Massive Datasets reveals a toolbox of new methods that are perfect for handling modern big data applications. You’ll explore the novel data structures and algorithms that underpin Google, Facebook, and other enterprise applications that work with truly massive amounts of data. These effective techniques can be applied to any discipline, from finance to text analysis. Graphics, illustrations, and hands-on industry examples make complex ideas practical to implement in your projects—and there’s no mathematical proofs to puzzle over. Work through this one-of-a-kind guide, and you’ll find the sweet spot of saving space without sacrificing your data’s accuracy. About the technology Standard algorithms and data structures may become slow—or fail altogether—when applied to large distributed datasets. Choosing algorithms designed for big data saves time, increases accuracy, and reduces processing cost. This unique book distills cutting-edge research papers into practical techniques for sketching, streaming, and organizing massive datasets on-disk and in the cloud. About the book Algorithms and Data Structures for Massive Datasets introduces processing and analytics techniques for large distributed data. Packed with industry stories and entertaining illustrations, this friendly guide makes even complex concepts easy to understand. You’ll explore real-world examples as you learn to map powerful algorithms like Bloom filters, Count-min sketch, HyperLogLog, and LSM-trees to your own use cases. What's inside Probabilistic sketching data structures Choosing the right database engine Designing efficient on-disk data structures and algorithms Algorithmic tradeoffs in massive-scale systems Computing percentiles with limited space resources About the reader Examples in Python, R, and pseudocode. About the author Dzejla Medjedovic earned her PhD in the Applied Algorithms Lab at Stony Brook University, New York. Emin Tahirovic earned his PhD in biostatistics from University of Pennsylvania. Illustrator Ines Dedovic earned her PhD at the Institute for Imaging and Computer Vision at RWTH Aachen University, Germany. Table of Contents 1 Introduction PART 1 HASH-BASED SKETCHES 2 Review of hash tables and modern hashing 3 Approximate membership: Bloom and quotient filters 4 Frequency estimation and count-min sketch 5 Cardinality estimation and HyperLogLog PART 2 REAL-TIME ANALYTICS 6 Streaming data: Bringing everything together 7 Sampling from data streams 8 Approximate quantiles on data streams PART 3 DATA STRUCTURES FOR DATABASES AND EXTERNAL MEMORY ALGORITHMS 9 Introducing the external memory model 10 Data structures for databases: B-trees, Bε-trees, and LSM-trees 11 External memory sorting




Mastering matplotlib


Book Description

matplotlib is a Python plotting library that provides a large feature set for a multitude of platforms. Given the depth of the library's legacy and the variety of related open source projects, gaining expert knowledge can be a time-consuming and often confusing process. You'll begin your exciting journey learning about the skills that are necessary in leading technical teams for a visualization project or to become a matplotlib contributor. Supported by highly-detailed IPython Notebooks, this book takes you through the conceptual components underlying the library and then provides a detailed overview of its APIs. From there, you will learn about event handling and how to code for interactive plots. Next you will move on to customization techniques, local configuration of matplotib, and then deployments in Cloud environments. The adventure culminates in an exploration of big data visualization and matplotlib clustering.




Data Science with Python and Dask


Book Description

Summary Dask is a native parallel analytics tool designed to integrate seamlessly with the libraries you're already using, including Pandas, NumPy, and Scikit-Learn. With Dask you can crunch and work with huge datasets, using the tools you already have. And Data Science with Python and Dask is your guide to using Dask for your data projects without changing the way you work! Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications. You'll find registration instructions inside the print book. About the Technology An efficient data pipeline means everything for the success of a data science project. Dask is a flexible library for parallel computing in Python that makes it easy to build intuitive workflows for ingesting and analyzing large, distributed datasets. Dask provides dynamic task scheduling and parallel collections that extend the functionality of NumPy, Pandas, and Scikit-learn, enabling users to scale their code from a single laptop to a cluster of hundreds of machines with ease. About the Book Data Science with Python and Dask teaches you to build scalable projects that can handle massive datasets. After meeting the Dask framework, you'll analyze data in the NYC Parking Ticket database and use DataFrames to streamline your process. Then, you'll create machine learning models using Dask-ML, build interactive visualizations, and build clusters using AWS and Docker. What's inside Working with large, structured and unstructured datasets Visualization with Seaborn and Datashader Implementing your own algorithms Building distributed apps with Dask Distributed Packaging and deploying Dask apps About the Reader For data scientists and developers with experience using Python and the PyData stack. About the Author Jesse Daniel is an experienced Python developer. He taught Python for Data Science at the University of Denver and leads a team of data scientists at a Denver-based media technology company. Table of Contents PART 1 - The Building Blocks of scalable computing Why scalable computing matters Introducing Dask PART 2 - Working with Structured Data using Dask DataFrames Introducing Dask DataFrames Loading data into DataFrames Cleaning and transforming DataFrames Summarizing and analyzing DataFrames Visualizing DataFrames with Seaborn Visualizing location data with Datashader PART 3 - Extending and deploying Dask Working with Bags and Arrays Machine learning with Dask-ML Scaling and deploying Dask