Advertisement
data engineering with dbt: Data Engineering with dbt Roberto Zagni, 2023-06-30 Use easy-to-apply patterns in SQL and Python to adopt modern analytics engineering to build agile platforms with dbt that are well-tested and simple to extend and run Purchase of the print or Kindle book includes a free PDF eBook Key Features Build a solid dbt base and learn data modeling and the modern data stack to become an analytics engineer Build automated and reliable pipelines to deploy, test, run, and monitor ELTs with dbt Cloud Guided dbt + Snowflake project to build a pattern-based architecture that delivers reliable datasets Book Descriptiondbt Cloud helps professional analytics engineers automate the application of powerful and proven patterns to transform data from ingestion to delivery, enabling real DataOps. This book begins by introducing you to dbt and its role in the data stack, along with how it uses simple SQL to build your data platform, helping you and your team work better together. You’ll find out how to leverage data modeling, data quality, master data management, and more to build a simple-to-understand and future-proof solution. As you advance, you’ll explore the modern data stack, understand how data-related careers are changing, and see how dbt enables this transition into the emerging role of an analytics engineer. The chapters help you build a sample project using the free version of dbt Cloud, Snowflake, and GitHub to create a professional DevOps setup with continuous integration, automated deployment, ELT run, scheduling, and monitoring, solving practical cases you encounter in your daily work. By the end of this dbt book, you’ll be able to build an end-to-end pragmatic data platform by ingesting data exported from your source systems, coding the needed transformations, including master data and the desired business rules, and building well-formed dimensional models or wide tables that’ll enable you to build reports with the BI tool of your choice.What you will learn Create a dbt Cloud account and understand the ELT workflow Combine Snowflake and dbt for building modern data engineering pipelines Use SQL to transform raw data into usable data, and test its accuracy Write dbt macros and use Jinja to apply software engineering principles Test data and transformations to ensure reliability and data quality Build a lightweight pragmatic data platform using proven patterns Write easy-to-maintain idempotent code using dbt materialization Who this book is for This book is for data engineers, analytics engineers, BI professionals, and data analysts who want to learn how to build simple, futureproof, and maintainable data platforms in an agile way. Project managers, data team managers, and decision makers looking to understand the importance of building a data platform and foster a culture of high-performing data teams will also find this book useful. Basic knowledge of SQL and data modeling will help you get the most out of the many layers of this book. The book also includes primers on many data-related subjects to help juniors get started. |
data engineering with dbt: Data Engineering with Python Paul Crickard, 2020-10-23 Build, monitor, and manage real-time data pipelines to create data engineering infrastructure efficiently using open-source Apache projects Key Features Become well-versed in data architectures, data preparation, and data optimization skills with the help of practical examples Design data models and learn how to extract, transform, and load (ETL) data using Python Schedule, automate, and monitor complex data pipelines in production Book DescriptionData engineering provides the foundation for data science and analytics, and forms an important part of all businesses. This book will help you to explore various tools and methods that are used for understanding the data engineering process using Python. The book will show you how to tackle challenges commonly faced in different aspects of data engineering. You’ll start with an introduction to the basics of data engineering, along with the technologies and frameworks required to build data pipelines to work with large datasets. You’ll learn how to transform and clean data and perform analytics to get the most out of your data. As you advance, you'll discover how to work with big data of varying complexity and production databases, and build data pipelines. Using real-world examples, you’ll build architectures on which you’ll learn how to deploy data pipelines. By the end of this Python book, you’ll have gained a clear understanding of data modeling techniques, and will be able to confidently build data engineering pipelines for tracking data, running quality checks, and making necessary changes in production.What you will learn Understand how data engineering supports data science workflows Discover how to extract data from files and databases and then clean, transform, and enrich it Configure processors for handling different file formats as well as both relational and NoSQL databases Find out how to implement a data pipeline and dashboard to visualize results Use staging and validation to check data before landing in the warehouse Build real-time pipelines with staging areas that perform validation and handle failures Get to grips with deploying pipelines in the production environment Who this book is for This book is for data analysts, ETL developers, and anyone looking to get started with or transition to the field of data engineering or refresh their knowledge of data engineering using Python. This book will also be useful for students planning to build a career in data engineering or IT professionals preparing for a transition. No previous knowledge of data engineering is required. |
data engineering with dbt: Mastering Snowflake Solutions Adam Morton, 2022-02-28 Design for large-scale, high-performance queries using Snowflake’s query processing engine to empower data consumers with timely, comprehensive, and secure access to data. This book also helps you protect your most valuable data assets using built-in security features such as end-to-end encryption for data at rest and in transit. It demonstrates key features in Snowflake and shows how to exploit those features to deliver a personalized experience to your customers. It also shows how to ingest the high volumes of both structured and unstructured data that are needed for game-changing business intelligence analysis. Mastering Snowflake Solutions starts with a refresher on Snowflake’s unique architecture before getting into the advanced concepts that make Snowflake the market-leading product it is today. Progressing through each chapter, you will learn how to leverage storage, query processing, cloning, data sharing, and continuous data protection features. This approach allows for greater operational agility in responding to the needs of modern enterprises, for example in supporting agile development techniques via database cloning. The practical examples and in-depth background on theory in this book help you unleash the power of Snowflake in building a high-performance system with little to no administrative overhead. Your result from reading will be a deep understanding of Snowflake that enables taking full advantage of Snowflake’s architecture to deliver value analytics insight to your business. What You Will Learn Optimize performance and costs associated with your use of the Snowflake data platform Enable data security to help in complying with consumer privacy regulations such as CCPA and GDPR Share data securely both inside your organization and with external partners Gain visibility to each interaction with your customers using continuous data feeds from Snowpipe Break down data silos to gain complete visibility your business-critical processes Transform customer experience and product quality through real-time analytics Who This Book Is for Data engineers, scientists, and architects who have had some exposure to the Snowflake data platform or bring some experience from working with another relational database. This book is for those beginning to struggle with new challenges as their Snowflake environment begins to mature, becoming more complex with ever increasing amounts of data, users, and requirements. New problems require a new approach and this book aims to arm you with the practical knowledge required to take advantage of Snowflake’s unique architecture to get the results you need. |
data engineering with dbt: The Data Warehouse Toolkit Ralph Kimball, Margy Ross, 2011-08-08 This old edition was published in 2002. The current and final edition of this book is The Data Warehouse Toolkit: The Definitive Guide to Dimensional Modeling, 3rd Edition which was published in 2013 under ISBN: 9781118530801. The authors begin with fundamental design recommendations and gradually progress step-by-step through increasingly complex scenarios. Clear-cut guidelines for designing dimensional models are illustrated using real-world data warehouse case studies drawn from a variety of business application areas and industries, including: Retail sales and e-commerce Inventory management Procurement Order management Customer relationship management (CRM) Human resources management Accounting Financial services Telecommunications and utilities Education Transportation Health care and insurance By the end of the book, you will have mastered the full range of powerful techniques for designing dimensional databases that are easy to understand and provide fast query response. You will also learn how to create an architected framework that integrates the distributed data warehouse using standardized dimensions and facts. |
data engineering with dbt: 97 Things Every Data Engineer Should Know Tobias Macey, 2021-06-11 Take advantage of today's sky-high demand for data engineers. With this in-depth book, current and aspiring engineers will learn powerful real-world best practices for managing data big and small. Contributors from notable companies including Twitter, Google, Stitch Fix, Microsoft, Capital One, and LinkedIn share their experiences and lessons learned for overcoming a variety of specific and often nagging challenges. Edited by Tobias Macey, host of the popular Data Engineering Podcast, this book presents 97 concise and useful tips for cleaning, prepping, wrangling, storing, processing, and ingesting data. Data engineers, data architects, data team managers, data scientists, machine learning engineers, and software engineers will greatly benefit from the wisdom and experience of their peers. Topics include: The Importance of Data Lineage - Julien Le Dem Data Security for Data Engineers - Katharine Jarmul The Two Types of Data Engineering and Data Engineers - Jesse Anderson Six Dimensions for Picking an Analytical Data Warehouse - Gleb Mezhanskiy The End of ETL as We Know It - Paul Singman Building a Career as a Data Engineer - Vijay Kiran Modern Metadata for the Modern Data Stack - Prukalpa Sankar Your Data Tests Failed! Now What? - Sam Bail |
data engineering with dbt: Streaming Systems Tyler Akidau, Slava Chernyak, Reuven Lax, 2018-07-16 Streaming data is a big deal in big data these days. As more and more businesses seek to tame the massive unbounded data sets that pervade our world, streaming systems have finally reached a level of maturity sufficient for mainstream adoption. With this practical guide, data engineers, data scientists, and developers will learn how to work with streaming data in a conceptual and platform-agnostic way. Expanded from Tyler Akidau’s popular blog posts Streaming 101 and Streaming 102, this book takes you from an introductory level to a nuanced understanding of the what, where, when, and how of processing real-time data streams. You’ll also dive deep into watermarks and exactly-once processing with co-authors Slava Chernyak and Reuven Lax. You’ll explore: How streaming and batch data processing patterns compare The core principles and concepts behind robust out-of-order data processing How watermarks track progress and completeness in infinite datasets How exactly-once data processing techniques ensure correctness How the concepts of streams and tables form the foundations of both batch and streaming data processing The practical motivations behind a powerful persistent state mechanism, driven by a real-world example How time-varying relations provide a link between stream processing and the world of SQL and relational algebra |
data engineering with dbt: Architecting Modern Data Platforms Jan Kunigk, Ian Buss, Paul Wilkinson, Lars George, 2018-12-05 There’s a lot of information about big data technologies, but splicing these technologies into an end-to-end enterprise data platform is a daunting task not widely covered. With this practical book, you’ll learn how to build big data infrastructure both on-premises and in the cloud and successfully architect a modern data platform. Ideal for enterprise architects, IT managers, application architects, and data engineers, this book shows you how to overcome the many challenges that emerge during Hadoop projects. You’ll explore the vast landscape of tools available in the Hadoop and big data realm in a thorough technical primer before diving into: Infrastructure: Look at all component layers in a modern data platform, from the server to the data center, to establish a solid foundation for data in your enterprise Platform: Understand aspects of deployment, operation, security, high availability, and disaster recovery, along with everything you need to know to integrate your platform with the rest of your enterprise IT Taking Hadoop to the cloud: Learn the important architectural aspects of running a big data platform in the cloud while maintaining enterprise security and high availability |
data engineering with dbt: Digital Business Transformation Nigel Vaz, 2021-01-05 Fuel your business' transition into the digital age with this insightful and comprehensive resource Digital Business Transformation: How Established Companies Sustain Competitive Advantage offers readers a framework for digital business transformation. Written by Nigel Vaz, the acclaimed CEO of Publicis Sapient, a global digital business transformation company, Digital Business Transformation delivers practical advice and approachable strategies to help businesses realize their digital potential. Digital Business Transformation provides readers with examples of the challenges faced by global organizations and the strategies they used to overcome them. The book also includes discussions of: How to decide whether to defend, differentiate, or disrupt your organization to meet digital challenges How to deconstruct decision-making throughout all levels of your organization How to combine strategy, product, experience, engineering, and data to produce digital results Perfect for anyone in a leadership position in a modern organization, particularly those who find themselves responsible for transformation-related decisions, Digital Business Transformation delivers a message that begs to be heard by everyone who hopes to help their organization meet the challenges of a changing world. |
data engineering with dbt: Analytics Engineering with SQL and dbt Rui Pedro Machado, Helder Russa, 2023-12-08 With the shift from data warehouses to data lakes, data now lands in repositories before it's been transformed, enabling engineers to model raw data into clean, well-defined datasets. dbt (data build tool) helps you take data further. This practical book shows data analysts, data engineers, BI developers, and data scientists how to create a true self-service transformation platform through the use of dynamic SQL. Authors Rui Machado from Monstarlab and Hélder Russa from Jumia show you how to quickly deliver new data products by focusing more on value delivery and less on architectural and engineering aspects. If you know your business well and have the technical skills to model raw data into clean, well-defined datasets, you'll learn how to design and deliver data models without any technical influence. With this book, you'll learn: What dbt is and how a dbt project is structured How dbt fits into the data engineering and analytics worlds How to collaborate on building data models The main tools and architectures for building useful, functional data models How to fit dbt into data warehousing and laking architecture How to build tests for data transformations |
data engineering with dbt: The Informed Company Dave Fowler, Matthew C. David, 2021-10-26 Learn how to manage a modern data stack and get the most out of data in your organization! Thanks to the emergence of new technologies and the explosion of data in recent years, we need new practices for managing and getting value out of data. In the modern, data driven competitive landscape the best guess approach—reading blog posts here and there and patching together data practices without any real visibility—is no longer going to hack it. The Informed Company provides definitive direction on how best to leverage the modern data stack, including cloud computing, columnar storage, cloud ETL tools, and cloud BI tools. You'll learn how to work with Agile methods and set up processes that's right for your company to use your data as a key weapon for your success . . . You'll discover best practices for every stage, from querying production databases at a small startup all the way to setting up data marts for different business lines of an enterprise. In their work at Chartio, authors Fowler and David have learned that most businesspeople are almost completely self-taught when it comes to data. If they are using resources, those resources are outdated, so they're missing out on the latest cloud technologies and advances in data analytics. This book will firm up your understanding of data and bring you into the present with knowledge around what works and what doesn't. Discover the data stack strategies that are working for today's successful small, medium, and enterprise companies Learn the different Agile stages of data organization, and the right one for your team Learn how to maintain Data Lakes and Data Warehouses for effective, accessible data storage Gain the knowledge you need to architect Data Warehouses and Data Marts Understand your business's level of data sophistication and the steps you can take to get to level up your data The Informed Company is the definitive data book for anyone who wants to work faster and more nimbly, armed with actionable decision-making data. |
data engineering with dbt: Building a Scalable Data Warehouse with Data Vault 2.0 Daniel Linstedt, Michael Olschimke, 2015-09-15 The Data Vault was invented by Dan Linstedt at the U.S. Department of Defense, and the standard has been successfully applied to data warehousing projects at organizations of different sizes, from small to large-size corporations. Due to its simplified design, which is adapted from nature, the Data Vault 2.0 standard helps prevent typical data warehousing failures. Building a Scalable Data Warehouse covers everything one needs to know to create a scalable data warehouse end to end, including a presentation of the Data Vault modeling technique, which provides the foundations to create a technical data warehouse layer. The book discusses how to build the data warehouse incrementally using the agile Data Vault 2.0 methodology. In addition, readers will learn how to create the input layer (the stage layer) and the presentation layer (data mart) of the Data Vault 2.0 architecture including implementation best practices. Drawing upon years of practical experience and using numerous examples and an easy to understand framework, Dan Linstedt and Michael Olschimke discuss: - How to load each layer using SQL Server Integration Services (SSIS), including automation of the Data Vault loading processes. - Important data warehouse technologies and practices. - Data Quality Services (DQS) and Master Data Services (MDS) in the context of the Data Vault architecture. - Provides a complete introduction to data warehousing, applications, and the business context so readers can get-up and running fast - Explains theoretical concepts and provides hands-on instruction on how to build and implement a data warehouse - Demystifies data vault modeling with beginning, intermediate, and advanced techniques - Discusses the advantages of the data vault approach over other techniques, also including the latest updates to Data Vault 2.0 and multiple improvements to Data Vault 1.0 |
data engineering with dbt: Joe Celko's SQL for Smarties Joe Celko, 2000 An industry consultant shares his most useful tips and tricks for advanced SQL programming to help the working programmer gain performance and work around system deficiencies. |
data engineering with dbt: Jumpstart Snowflake Dmitry Anoshin, Dmitry Shirokov, Donna Strok, 2019-12-20 Explore the modern market of data analytics platforms and the benefits of using Snowflake computing, the data warehouse built for the cloud. With the rise of cloud technologies, organizations prefer to deploy their analytics using cloud providers such as Amazon Web Services (AWS), Microsoft Azure, or Google Cloud Platform. Cloud vendors are offering modern data platforms for building cloud analytics solutions to collect data and consolidate into single storage solutions that provide insights for business users. The core of any analytics framework is the data warehouse, and previously customers did not have many choices of platform to use. Snowflake was built specifically for the cloud and it is a true game changer for the analytics market. This book will help onboard you to Snowflake, present best practices to deploy, and use the Snowflake data warehouse. In addition, it covers modern analytics architecture and use cases. It provides use cases of integration with leading analytics software such as Matillion ETL, Tableau, and Databricks. Finally, it covers migration scenarios for on-premise legacy data warehouses. What You Will Learn Know the key functionalities of Snowflake Set up security and access with cluster Bulk load data into Snowflake using the COPY command Migrate from a legacy data warehouse to Snowflake integrate the Snowflake data platform with modern business intelligence (BI) and data integration tools Who This Book Is For Those working with data warehouse and business intelligence (BI) technologies, and existing and potential Snowflake users |
data engineering with dbt: Information Dashboard Design Stephen Few, 2006 Dashboards have become popular in recent years as uniquely powerful tools for communicating important information at a glance. Although dashboards are potentially powerful, this potential is rarely realized. The greatest display technology in the world won't solve this if you fail to use effective visual design. And if a dashboard fails to tell you precisely what you need to know in an instant, you'll never use it, even if it's filled with cute gauges, meters, and traffic lights. Don't let your investment in dashboard technology go to waste. This book will teach you the visual design skills you need to create dashboards that communicate clearly, rapidly, and compellingly. Information Dashboard Design will explain how to: Avoid the thirteen mistakes common to dashboard design Provide viewers with the information they need quickly and clearly Apply what we now know about visual perception to the visual presentation of information Minimize distractions, cliches, and unnecessary embellishments that create confusion Organize business information to support meaning and usability Create an aesthetically pleasing viewing experience Maintain consistency of design to provide accurate interpretation Optimize the power of dashboard technology by pairing it with visual effectiveness Stephen Few has over 20 years of experience as an IT innovator, consultant, and educator. As Principal of the consultancy Perceptual Edge, Stephen focuses on data visualization for analyzing and communicating quantitative business information. He provides consulting and training services, speaks frequently at conferences, and teaches in the MBA program at the University ofCalifornia in Berkeley. He is also the author of Show Me the Numbers: Designing Tables and Graphs to Enlighten. Visit his website at www.perceptualedge.com. |
data engineering with dbt: Statistics for Engineers and Scientists William Cyrus Navidi, 2008 |
data engineering with dbt: Agile Data Warehouse Design Lawrence Corr, Jim Stagnitto, 2011-11 Agile Data Warehouse Design is a step-by-step guide for capturing data warehousing/business intelligence (DW/BI) requirements and turning them into high performance dimensional models in the most direct way: by modelstorming (data modeling + brainstorming) with BI stakeholders. This book describes BEAM✲, an agile approach to dimensional modeling, for improving communication between data warehouse designers, BI stakeholders and the whole DW/BI development team. BEAM✲ provides tools and techniques that will encourage DW/BI designers and developers to move away from their keyboards and entity relationship based tools and model interactively with their colleagues. The result is everyone thinks dimensionally from the outset! Developers understand how to efficiently implement dimensional modeling solutions. Business stakeholders feel ownership of the data warehouse they have created, and can already imagine how they will use it to answer their business questions. Within this book, you will learn: ✲ Agile dimensional modeling using Business Event Analysis & Modeling (BEAM✲) ✲ Modelstorming: data modeling that is quicker, more inclusive, more productive, and frankly more fun! ✲ Telling dimensional data stories using the 7Ws (who, what, when, where, how many, why and how) ✲ Modeling by example not abstraction; using data story themes, not crow's feet, to describe detail ✲ Storyboarding the data warehouse to discover conformed dimensions and plan iterative development ✲ Visual modeling: sketching timelines, charts and grids to model complex process measurement - simply ✲ Agile design documentation: enhancing star schemas with BEAM✲ dimensional shorthand notation ✲ Solving difficult DW/BI performance and usability problems with proven dimensional design patterns Lawrence Corr is a data warehouse designer and educator. As Principal of DecisionOne Consulting, he helps clients to review and simplify their data warehouse designs, and advises vendors on visual data modeling techniques. He regularly teaches agile dimensional modeling courses worldwide and has taught dimensional DW/BI skills to thousands of students. Jim Stagnitto is a data warehouse and master data management architect specializing in the healthcare, financial services, and information service industries. He is the founder of the data warehousing and data mining consulting firm Llumino. |
data engineering with dbt: The Unified Star Schema Bill Inmon, Francesco Puppini, 2020-10 Master the most agile and resilient design for building analytics applications: the Unified Star Schema (USS) approach. The USS has many benefits over traditional dimensional modeling. Witness the power of the USS as a single star schema that serves as a foundation for all present and future business requirements of your organization. |
data engineering with dbt: Think Stats Allen B. Downey, 2014-10-16 If you know how to program, you have the skills to turn data into knowledge, using tools of probability and statistics. This concise introduction shows you how to perform statistical analysis computationally, rather than mathematically, with programs written in Python. By working with a single case study throughout this thoroughly revised book, you’ll learn the entire process of exploratory data analysis—from collecting data and generating statistics to identifying patterns and testing hypotheses. You’ll explore distributions, rules of probability, visualization, and many other tools and concepts. New chapters on regression, time series analysis, survival analysis, and analytic methods will enrich your discoveries. Develop an understanding of probability and statistics by writing and testing code Run experiments to test statistical behavior, such as generating samples from several distributions Use simulations to understand concepts that are hard to grasp mathematically Import data from most sources with Python, rather than rely on data that’s cleaned and formatted for statistics tools Use statistical inference to answer questions about real-world data |
data engineering with dbt: The Oxford Handbook of Dialectical Behaviour Therapy Michaela A. Swales, 2018 Dialectical behavior therapy (DBT) is a specific type of cognitive-behavioral psychotherapy developed in the late 1980s by psychologist Marsha M. Linehan to help better treat borderline personality disorder. Since its development, it has also been used for the treatment of other kinds of mental health disorders. The Oxford Handbook of DBT charts the development of DBT from its early inception to the current cutting edge state of knowledge about both the theoretical underpinnings of the treatment and its clinical application across a range of disorders and adaptations to new clinical groups. Experts in the treatment address the current state of the evidence with respect to the efficacy of the treatment, its effectiveness in routine clinical practice and central issues in the clinical and programmatic implementation of the treatment. In sum this volume provides a desk reference for clinicians and academics keen to understand the origins and current state of the science, and the art, of DBT. |
data engineering with dbt: Building a Life Worth Living Marsha M. Linehan, 2021-01-05 Marsha Linehan tells the story of her journey from suicidal teenager to world-renowned developer of the life-saving behavioral therapy DBT, using her own struggle to develop life skills for others. “This book is a victory on both sides of the page.”—Gloria Steinem “Are you one of us?” a patient once asked Marsha Linehan, the world-renowned psychologist who developed Dialectical Behavior Therapy. “Because if you were, it would give all of us so much hope.” Over the years, DBT had saved the lives of countless people fighting depression and suicidal thoughts, but Linehan had never revealed that her pioneering work was inspired by her own desperate struggles as a young woman. Only when she received this question did she finally decide to tell her story. In this remarkable and inspiring memoir, Linehan describes how, when she was eighteen years old, she began an abrupt downward spiral from popular teenager to suicidal young woman. After several miserable years in a psychiatric institute, Linehan made a vow that if she could get out of emotional hell, she would try to find a way to help others get out of hell too, and to build a life worth living. She went on to put herself through night school and college, living at a YWCA and often scraping together spare change to buy food. She went on to get her PhD in psychology, specializing in behavior therapy. In the 1980s, she achieved a breakthrough when she developed Dialectical Behavioral Therapy, a therapeutic approach that combines acceptance of the self and ways to change. Linehan included mindfulness as a key component in therapy treatment, along with original and specific life-skill techniques. She says, You can't think yourself into new ways of acting; you can only act yourself into new ways of thinking. Throughout her extraordinary scientific career, Marsha Linehan remained a woman of deep spirituality. Her powerful and moving story is one of faith and perseverance. Linehan shows, in Building a Life Worth Living, how the principles of DBT really work—and how, using her life skills and techniques, people can build lives worth living. |
data engineering with dbt: Data Science in Production Ben Weber, 2020 Putting predictive models into production is one of the most direct ways that data scientists can add value to an organization. By learning how to build and deploy scalable model pipelines, data scientists can own more of the model production process and more rapidly deliver data products. This book provides a hands-on approach to scaling up Python code to work in distributed environments in order to build robust pipelines. Readers will learn how to set up machine learning models as web endpoints, serverless functions, and streaming pipelines using multiple cloud environments. It is intended for analytics practitioners with hands-on experience with Python libraries such as Pandas and scikit-learn, and will focus on scaling up prototype models to production. From startups to trillion dollar companies, data science is playing an important role in helping organizations maximize the value of their data. This book helps data scientists to level up their careers by taking ownership of data products with applied examples that demonstrate how to: Translate models developed on a laptop to scalable deployments in the cloud Develop end-to-end systems that automate data science workflows Own a data product from conception to production The accompanying Jupyter notebooks provide examples of scalable pipelines across multiple cloud environments, tools, and libraries (github.com/bgweber/DS_Production). Book Contents Here are the topics covered by Data Science in Production: Chapter 1: Introduction - This chapter will motivate the use of Python and discuss the discipline of applied data science, present the data sets, models, and cloud environments used throughout the book, and provide an overview of automated feature engineering. Chapter 2: Models as Web Endpoints - This chapter shows how to use web endpoints for consuming data and hosting machine learning models as endpoints using the Flask and Gunicorn libraries. We'll start with scikit-learn models and also set up a deep learning endpoint with Keras. Chapter 3: Models as Serverless Functions - This chapter will build upon the previous chapter and show how to set up model endpoints as serverless functions using AWS Lambda and GCP Cloud Functions. Chapter 4: Containers for Reproducible Models - This chapter will show how to use containers for deploying models with Docker. We'll also explore scaling up with ECS and Kubernetes, and building web applications with Plotly Dash. Chapter 5: Workflow Tools for Model Pipelines - This chapter focuses on scheduling automated workflows using Apache Airflow. We'll set up a model that pulls data from BigQuery, applies a model, and saves the results. Chapter 6: PySpark for Batch Modeling - This chapter will introduce readers to PySpark using the community edition of Databricks. We'll build a batch model pipeline that pulls data from a data lake, generates features, applies a model, and stores the results to a No SQL database. Chapter 7: Cloud Dataflow for Batch Modeling - This chapter will introduce the core components of Cloud Dataflow and implement a batch model pipeline for reading data from BigQuery, applying an ML model, and saving the results to Cloud Datastore. Chapter 8: Streaming Model Workflows - This chapter will introduce readers to Kafka and PubSub for streaming messages in a cloud environment. After working through this material, readers will learn how to use these message brokers to create streaming model pipelines with PySpark and Dataflow that provide near real-time predictions. Excerpts of these chapters are available on Medium (@bgweber), and a book sample is available on Leanpub. |
data engineering with dbt: Structural Design for Physical Security Task Committee on Structural Design for Physical Security, 1999-01-01 Prepared by the Task Committee on Structural Design for Physical Security of the Structural Engineering Institute of ASCE. This report provides guidance to structural engineers in the design of civil structures to resist the effects of terrorist bombings. As dramatized by the bombings of the World Trade Center in New York City and the Murrah Building in Oklahoma City, civil engineers today need guidance on designing structures to resist hostile acts. The U.S. military services and foreign embassy facilities developed requirements for their unique needs, but these the documents are restricted. Thus, no widely available document exists to provide engineers with the technical data necessary to design civil structures for enhanced physical security. The unrestricted government information included in this report is assembled collectively for the first time and rephrased for application to civilian facilities. Topics include: determination of the threat, methods by which structural loadings are derived for the determined threat, the behavior and selection of structural systems, the design of structural components, the design of security doors, the design of utility openings, and the retrofitting of existing structures. This report transfers this technology to the civil sector and provides complete methods, guidance, and references for structural engineers challenged with a physical security problem. |
data engineering with dbt: Building the Data Lakehouse Bill Inmon, Ranjeet Srivastava, Mary Levins, 2021-10 The data lakehouse is the next generation of the data warehouse and data lake, designed to meet today's complex and ever-changing analytics, machine learning, and data science requirements. Learn about the features and architecture of the data lakehouse, along with its powerful analytical infrastructure. Appreciate how the universal common connector blends structured, textual, analog, and IoT data. Maintain the lakehouse for future generations through Data Lakehouse Housekeeping and Data Future-proofing. Know how to incorporate the lakehouse into an existing data governance strategy. Incorporate data catalogs, data lineage tools, and open source software into your architecture to ensure your data scientists, analysts, and end users live happily ever after. |
data engineering with dbt: Data Engineering with AWS Gareth Eagar, 2023-10-31 Looking to revolutionize your data transformation game with AWS? Look no further! From strong foundations to hands-on building of data engineering pipelines, our expert-led manual has got you covered. Key Features Delve into robust AWS tools for ingesting, transforming, and consuming data, and for orchestrating pipelines Stay up to date with a comprehensive revised chapter on Data Governance Build modern data platforms with a new section covering transactional data lakes and data mesh Book DescriptionThis book, authored by a seasoned Senior Data Architect with 25 years of experience, aims to help you achieve proficiency in using the AWS ecosystem for data engineering. This revised edition provides updates in every chapter to cover the latest AWS services and features, takes a refreshed look at data governance, and includes a brand-new section on building modern data platforms which covers; implementing a data mesh approach, open-table formats (such as Apache Iceberg), and using DataOps for automation and observability. You'll begin by reviewing the key concepts and essential AWS tools in a data engineer's toolkit and getting acquainted with modern data management approaches. You'll then architect a data pipeline, review raw data sources, transform the data, and learn how that transformed data is used by various data consumers. You’ll learn how to ensure strong data governance, and about populating data marts and data warehouses along with how a data lakehouse fits into the picture. After that, you'll be introduced to AWS tools for analyzing data, including those for ad-hoc SQL queries and creating visualizations. Then, you'll explore how the power of machine learning and artificial intelligence can be used to draw new insights from data. In the final chapters, you'll discover transactional data lakes, data meshes, and how to build a cutting-edge data platform on AWS. By the end of this AWS book, you'll be able to execute data engineering tasks and implement a data pipeline on AWS like a pro!What you will learn Seamlessly ingest streaming data with Amazon Kinesis Data Firehose Optimize, denormalize, and join datasets with AWS Glue Studio Use Amazon S3 events to trigger a Lambda process to transform a file Load data into a Redshift data warehouse and run queries with ease Visualize and explore data using Amazon QuickSight Extract sentiment data from a dataset using Amazon Comprehend Build transactional data lakes using Apache Iceberg with Amazon Athena Learn how a data mesh approach can be implemented on AWS Who this book is forThis book is for data engineers, data analysts, and data architects who are new to AWS and looking to extend their skills to the AWS cloud. Anyone new to data engineering who wants to learn about the foundational concepts, while gaining practical experience with common data engineering services on AWS, will also find this book useful. A basic understanding of big data-related topics and Python coding will help you get the most out of this book, but it’s not a prerequisite. Familiarity with the AWS console and core services will also help you follow along. |
data engineering with dbt: Learning Spark Jules S. Damji, Brooke Wenig, Tathagata Das, Denny Lee, 2020-07-16 Data is bigger, arrives faster, and comes in a variety of formats—and it all needs to be processed at scale for analytics or machine learning. But how can you process such varied workloads efficiently? Enter Apache Spark. Updated to include Spark 3.0, this second edition shows data engineers and data scientists why structure and unification in Spark matters. Specifically, this book explains how to perform simple and complex data analytics and employ machine learning algorithms. Through step-by-step walk-throughs, code snippets, and notebooks, you’ll be able to: Learn Python, SQL, Scala, or Java high-level Structured APIs Understand Spark operations and SQL Engine Inspect, tune, and debug Spark operations with Spark configurations and Spark UI Connect to data sources: JSON, Parquet, CSV, Avro, ORC, Hive, S3, or Kafka Perform analytics on batch and streaming data using Structured Streaming Build reliable data pipelines with open source Delta Lake and Spark Develop machine learning pipelines with MLlib and productionize models using MLflow |
data engineering with dbt: The Self-Service Data Roadmap Sandeep Uttamchandani, 2020-09-10 Data-driven insights are a key competitive advantage for any industry today, but deriving insights from raw data can still take days or weeks. Most organizations can’t scale data science teams fast enough to keep up with the growing amounts of data to transform. What’s the answer? Self-service data. With this practical book, data engineers, data scientists, and team managers will learn how to build a self-service data science platform that helps anyone in your organization extract insights from data. Sandeep Uttamchandani provides a scorecard to track and address bottlenecks that slow down time to insight across data discovery, transformation, processing, and production. This book bridges the gap between data scientists bottlenecked by engineering realities and data engineers unclear about ways to make self-service work. Build a self-service portal to support data discovery, quality, lineage, and governance Select the best approach for each self-service capability using open source cloud technologies Tailor self-service for the people, processes, and technology maturity of your data platform Implement capabilities to democratize data and reduce time to insight Scale your self-service portal to support a large number of users within your organization |
data engineering with dbt: Hands on Data Science for Biologists Using Python Yasha Hasija, Rajkumar Chakraborty, 2021-04-08 Hands-on Data Science for Biologists using Python has been conceptualized to address the massive data handling needs of modern-day biologists. With the advent of high throughput technologies and consequent availability of omics data, biological science has become a data-intensive field. This hands-on textbook has been written with the inception of easing data analysis by providing an interactive, problem-based instructional approach in Python programming language. The book starts with an introduction to Python and steadily delves into scrupulous techniques of data handling, preprocessing, and visualization. The book concludes with machine learning algorithms and their applications in biological data science. Each topic has an intuitive explanation of concepts and is accompanied with biological examples. Features of this book: The book contains standard templates for data analysis using Python, suitable for beginners as well as advanced learners. This book shows working implementations of data handling and machine learning algorithms using real-life biological datasets and problems, such as gene expression analysis; disease prediction; image recognition; SNP association with phenotypes and diseases. Considering the importance of visualization for data interpretation, especially in biological systems, there is a dedicated chapter for the ease of data visualization and plotting. Every chapter is designed to be interactive and is accompanied with Jupyter notebook to prompt readers to practice in their local systems. Other avant-garde component of the book is the inclusion of a machine learning project, wherein various machine learning algorithms are applied for the identification of genes associated with age-related disorders. A systematic understanding of data analysis steps has always been an important element for biological research. This book is a readily accessible resource that can be used as a handbook for data analysis, as well as a platter of standard code templates for building models. |
data engineering with dbt: Big Data Rajkumar Buyya, Rodrigo N. Calheiros, Amir Vahid Dastjerdi, 2016-06-07 Big Data: Principles and Paradigms captures the state-of-the-art research on the architectural aspects, technologies, and applications of Big Data. The book identifies potential future directions and technologies that facilitate insight into numerous scientific, business, and consumer applications. To help realize Big Data's full potential, the book addresses numerous challenges, offering the conceptual and technological solutions for tackling them. These challenges include life-cycle data management, large-scale storage, flexible processing infrastructure, data modeling, scalable machine learning, data analysis algorithms, sampling techniques, and privacy and ethical issues. - Covers computational platforms supporting Big Data applications - Addresses key principles underlying Big Data computing - Examines key developments supporting next generation Big Data platforms - Explores the challenges in Big Data computing and ways to overcome them - Contains expert contributors from both academia and industry |
data engineering with dbt: Biodefense in the Age of Synthetic Biology National Academies of Sciences, Engineering, and Medicine, Division on Earth and Life Studies, Board on Life Sciences, Board on Chemical Sciences and Technology, Committee on Strategies for Identifying and Addressing Potential Biodefense Vulnerabilities Posed by Synthetic Biology, 2019-01-05 Scientific advances over the past several decades have accelerated the ability to engineer existing organisms and to potentially create novel ones not found in nature. Synthetic biology, which collectively refers to concepts, approaches, and tools that enable the modification or creation of biological organisms, is being pursued overwhelmingly for beneficial purposes ranging from reducing the burden of disease to improving agricultural yields to remediating pollution. Although the contributions synthetic biology can make in these and other areas hold great promise, it is also possible to imagine malicious uses that could threaten U.S. citizens and military personnel. Making informed decisions about how to address such concerns requires a realistic assessment of the capabilities that could be misused. Biodefense in the Age of Synthetic Biology explores and envisions potential misuses of synthetic biology. This report develops a framework to guide an assessment of the security concerns related to advances in synthetic biology, assesses the levels of concern warranted for such advances, and identifies options that could help mitigate those concerns. |
data engineering with dbt: Computer-Aided Materials Selection During Structural Design National Research Council, Division on Engineering and Physical Sciences, National Materials Advisory Board, Commission on Engineering and Technical Systems, Committee on Application of Expert Systems to Materials Selection During Structural Design, 1995-04-03 The selection of the proper materials for a structural component is a critical activity that is governed by many, often conflicting factors. Incorporating materials expert systems into CAD/CAM operations could assist designers by suggesting potential manufacturing processes for particular products to facilitate concurrent engineering, recommending various materials for a specific part based on a given set of characteristics, or proposing possible modifications of a design if suitable materials for a particular part do not exist. This book reviews the structural design process, determines the elements, and capabilities required for a materials selection expert system to assist design engineers, and recommends the areas of expert system and materials modeling research and development required to devise a materials-specific design system. |
data engineering with dbt: Docker: Up & Running Sean P. Kane, Karl Matthias, 2018-09-07 Docker is rapidly changing the way organizations deploy software at scale. However, understanding how Linux containers fit into your workflow—and getting the integration details right—is not a trivial task. With the updated edition of this practical guide, you’ll learn how to use Docker to package your applications with all of their dependencies and then test, ship, scale, and support your containers in production. This edition includes significant updates to the examples and explanations that reflect the substantial changes that have occurred over the past couple of years. Sean Kane and Karl Matthias have added a complete chapter on Docker Compose, deeper coverage of Docker Swarm mode, introductions to both Kubernetes and AWS Fargate, examples on how to optimize your Docker images, and much more. Learn how Docker simplifies dependency management and deployment workflow for your applications Start working with Docker images, containers, and command line tools Use practical techniques to deploy and test Docker containers in production Debug containers by understanding their composition and internal processes Deploy production containers at scale inside your data center or cloud environment Explore advanced Docker topics, including deployment tools, networking, orchestration, security, and configuration |
data engineering with dbt: Trino: The Definitive Guide Matt Fuller, Manfred Moser, Martin Traverso, 2021-04-14 Perform fast interactive analytics against different data sources using the Trino high-performance distributed SQL query engine. With this practical guide, you'll learn how to conduct analytics on data where it lives, whether it's Hive, Cassandra, a relational database, or a proprietary data store. Analysts, software engineers, and production engineers will learn how to manage, use, and even develop with Trino. Initially developed by Facebook, open source Trino is now used by Netflix, Airbnb, LinkedIn, Twitter, Uber, and many other companies. Matt Fuller, Manfred Moser, and Martin Traverso show you how a single Trino query can combine data from multiple sources to allow for analytics across your entire organization. Get started: Explore Trino's use cases and learn about tools that will help you connect to Trino and query data Go deeper: Learn Trino's internal workings, including how to connect to and query data sources with support for SQL statements, operators, functions, and more Put Trino in production: Secure Trino, monitor workloads, tune queries, and connect more applications; learn how other organizations apply Trino |
data engineering with dbt: Darkening Skies Eden Hudson, 2019-08-05 A boy abandoned at an ancient school of warrior arts. A princess bartered away as a treaty bride. A lurking evil with the power to topple the world. Ji Yu Raijin is happy to live out his life serving his school, cultivating his life force, and learning the Path of Darkening Skies, an ancient and noble warrior art-until the day comes when he has to choose: remain true to the teachings of his path and allow his art to die out or commit an unforgivable sin and save the world. Second princess Shyong San Koida was born the only cripple in a dynasty of powerful warriors. With a broken life force, the only way Koida can contribute to her family's empire is by sealing a strategic alliance to a barbarian leader with her hand in marriage. But hidden forces within the court conspire to stop the union and obliterate the Shyong San dynasty at all costs. At the intersection of the servant boy and the princess's stories lies a secret, malignant art bent on destroying not only them, but the entire world. Darkening Skies is the first novel in the Path of the Thunderbird series, for fans of wuxia, xanxia, cultivation novels, demon beasts, and all styles of martial arts. Darkening Skies is written by eden Hudson, enthusiastic but terrible martial arts student and best-selling author of Jubal Van Zandt and the litRPG fantasy Rogue Dungeon. |
data engineering with dbt: Data Pipelines with Apache Airflow Bas P. Harenslak, Julian de Ruiter, 2021-04-27 This book teaches you how to build and maintain effective data pipelines. Youll explore the most common usage patterns, including aggregating multiple data sources, connecting to and from data lakes, and cloud deployment. -- |
data engineering with dbt: Engineering Innovative Products Roger Woods, Karen Rafferty, Julian Murphy, Paul Hermon, 2014-07-28 The book provides vital insights into commercial development for engineering students in a highly practical and applied manner. Over the past 3 years, application of the book’s material has allowed the students to develop their commercial literacy and ambition in the University. —Steve Orr, Director, Northern Ireland Science Park CONNECT program which looks to accelerate the growth of knowledge-based companies in Northern Ireland Engineering Innovative Products: A Practical Experience is a pioneering book that will be of key use to senior undergraduate and graduate engineering students who are being encouraged to explore innovation and commercialization as part of their courses. The book will teach the essential skills of entrepreneurship and address the fundamental requirements needed to establish a successful technology company. As well as providing the crucial background and insights enabling students to identify a key market, it also offers a highly practical guide to undertaking genuine product validation and producing a feasibility study, as well as providing vital insights into the challenges and demands in forming a technology based company. Key features: Outlines how to develop and grow an engineering solution which has market potential and covers key business aspects of giving the perfect pitch, sales and marketing, protection of ideas and finance, to offer a complete and practical guide to commercializing ideas. Provides vital insight into the design and innovation processes within engineering and the challenges and pitfalls in translating good ideas into great products. Features contributions from leading experts in marketing, finance, company formation, sales and intellectual protection which provides details of the challenges faced by innovators when commercializing ideas. Includes Ccase studies from engineering students who give insights into how they have successfully developed their own ideas into companies. |
data engineering with dbt: Data Engineering with Scala and Spark Eric Tome, Rupam Bhattacharjee, David Radford, 2024-01-31 Take your data engineering skills to the next level by learning how to utilize Scala and functional programming to create continuous and scheduled pipelines that ingest, transform, and aggregate data Key Features Transform data into a clean and trusted source of information for your organization using Scala Build streaming and batch-processing pipelines with step-by-step explanations Implement and orchestrate your pipelines by following CI/CD best practices and test-driven development (TDD) Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionMost data engineers know that performance issues in a distributed computing environment can easily lead to issues impacting the overall efficiency and effectiveness of data engineering tasks. While Python remains a popular choice for data engineering due to its ease of use, Scala shines in scenarios where the performance of distributed data processing is paramount. This book will teach you how to leverage the Scala programming language on the Spark framework and use the latest cloud technologies to build continuous and triggered data pipelines. You’ll do this by setting up a data engineering environment for local development and scalable distributed cloud deployments using data engineering best practices, test-driven development, and CI/CD. You’ll also get to grips with DataFrame API, Dataset API, and Spark SQL API and its use. Data profiling and quality in Scala will also be covered, alongside techniques for orchestrating and performance tuning your end-to-end pipelines to deliver data to your end users. By the end of this book, you will be able to build streaming and batch data pipelines using Scala while following software engineering best practices.What you will learn Set up your development environment to build pipelines in Scala Get to grips with polymorphic functions, type parameterization, and Scala implicits Use Spark DataFrames, Datasets, and Spark SQL with Scala Read and write data to object stores Profile and clean your data using Deequ Performance tune your data pipelines using Scala Who this book is for This book is for data engineers who have experience in working with data and want to understand how to transform raw data into a clean, trusted, and valuable source of information for their organization using Scala and the latest cloud technologies. |
data engineering with dbt: Fundamentals of Data Engineering Joe Reis, Matt Housley, 2022-06-22 Data engineering has grown rapidly in the past decade, leaving many software engineers, data scientists, and analysts looking for a comprehensive view of this practice. With this practical book, you'll learn how to plan and build systems to serve the needs of your organization and customers by evaluating the best technologies available through the framework of the data engineering lifecycle. Authors Joe Reis and Matt Housley walk you through the data engineering lifecycle and show you how to stitch together a variety of cloud technologies to serve the needs of downstream data consumers. You'll understand how to apply the concepts of data generation, ingestion, orchestration, transformation, storage, and governance that are critical in any data environment regardless of the underlying technology. This book will help you: Get a concise overview of the entire data engineering landscape Assess data engineering problems using an end-to-end framework of best practices Cut through marketing hype when choosing data technologies, architecture, and processes Use the data engineering lifecycle to design and build a robust architecture Incorporate data governance and security across the data engineering lifecycle |
data engineering with dbt: Enterprise Integration Patterns Gregor Hohpe, Bobby Woolf, 2012-03-09 Enterprise Integration Patterns provides an invaluable catalog of sixty-five patterns, with real-world solutions that demonstrate the formidable of messaging and help you to design effective messaging solutions for your enterprise. The authors also include examples covering a variety of different integration technologies, such as JMS, MSMQ, TIBCO ActiveEnterprise, Microsoft BizTalk, SOAP, and XSL. A case study describing a bond trading system illustrates the patterns in practice, and the book offers a look at emerging standards, as well as insights into what the future of enterprise integration might hold. This book provides a consistent vocabulary and visual notation framework to describe large-scale integration solutions across many technologies. It also explores in detail the advantages and limitations of asynchronous messaging architectures. The authors present practical advice on designing code that connects an application to a messaging system, and provide extensive information to help you determine when to send a message, how to route it to the proper destination, and how to monitor the health of a messaging system. If you want to know how to manage, monitor, and maintain a messaging system once it is in use, get this book. |
data engineering with dbt: Practical DataOps Harvinder Atwal, 2019-12-09 Gain a practical introduction to DataOps, a new discipline for delivering data science at scale inspired by practices at companies such as Facebook, Uber, LinkedIn, Twitter, and eBay. Organizations need more than the latest AI algorithms, hottest tools, and best people to turn data into insight-driven action and useful analytical data products. Processes and thinking employed to manage and use data in the 20th century are a bottleneck for working effectively with the variety of data and advanced analytical use cases that organizations have today. This book provides the approach and methods to ensure continuous rapid use of data to create analytical data products and steer decision making. Practical DataOps shows you how to optimize the data supply chain from diverse raw data sources to the final data product, whether the goal is a machine learning model or other data-orientated output. The book provides an approach to eliminate wasted effort and improve collaboration between data producers, data consumers, and the rest of the organization through the adoption of lean thinking and agile software development principles. This book helps you to improve the speed and accuracy of analytical application development through data management and DevOps practices that securely expand data access, and rapidly increase the number of reproducible data products through automation, testing, and integration. The book also shows how to collect feedback and monitor performance to manage and continuously improve your processes and output. What You Will LearnDevelop a data strategy for your organization to help it reach its long-term goals Recognize and eliminate barriers to delivering data to users at scale Work on the right things for the right stakeholders through agile collaboration Create trust in data via rigorous testing and effective data management Build a culture of learning and continuous improvement through monitoring deployments and measuring outcomes Create cross-functional self-organizing teams focused on goals not reporting lines Build robust, trustworthy, data pipelines in support of AI, machine learning, and other analytical data products Who This Book Is For Data science and advanced analytics experts, CIOs, CDOs (chief data officers), chief analytics officers, business analysts, business team leaders, and IT professionals (data engineers, developers, architects, and DBAs) supporting data teams who want to dramatically increase the value their organization derives from data. The book is ideal for data professionals who want to overcome challenges of long delivery time, poor data quality, high maintenance costs, and scaling difficulties in getting data science output and machine learning into customer-facing production. |
data engineering with dbt: Spark in Action Jean-Georges Perrin, 2020-05-12 Summary The Spark distributed data processing platform provides an easy-to-implement tool for ingesting, streaming, and processing data from any source. In Spark in Action, Second Edition, you’ll learn to take advantage of Spark’s core features and incredible processing speed, with applications including real-time computation, delayed evaluation, and machine learning. Spark skills are a hot commodity in enterprises worldwide, and with Spark’s powerful and flexible Java APIs, you can reap all the benefits without first learning Scala or Hadoop. Foreword by Rob Thomas. About the technology Analyzing enterprise data starts by reading, filtering, and merging files and streams from many sources. The Spark data processing engine handles this varied volume like a champ, delivering speeds 100 times faster than Hadoop systems. Thanks to SQL support, an intuitive interface, and a straightforward multilanguage API, you can use Spark without learning a complex new ecosystem. About the book Spark in Action, Second Edition, teaches you to create end-to-end analytics applications. In this entirely new book, you’ll learn from interesting Java-based examples, including a complete data pipeline for processing NASA satellite data. And you’ll discover Java, Python, and Scala code samples hosted on GitHub that you can explore and adapt, plus appendixes that give you a cheat sheet for installing tools and understanding Spark-specific terms. What's inside Writing Spark applications in Java Spark application architecture Ingestion through files, databases, streaming, and Elasticsearch Querying distributed datasets with Spark SQL About the reader This book does not assume previous experience with Spark, Scala, or Hadoop. About the author Jean-Georges Perrin is an experienced data and software architect. He is France’s first IBM Champion and has been honored for 12 consecutive years. Table of Contents PART 1 - THE THEORY CRIPPLED BY AWESOME EXAMPLES 1 So, what is Spark, anyway? 2 Architecture and flow 3 The majestic role of the dataframe 4 Fundamentally lazy 5 Building a simple app for deployment 6 Deploying your simple app PART 2 - INGESTION 7 Ingestion from files 8 Ingestion from databases 9 Advanced ingestion: finding data sources and building your own 10 Ingestion through structured streaming PART 3 - TRANSFORMING YOUR DATA 11 Working with SQL 12 Transforming your data 13 Transforming entire documents 14 Extending transformations with user-defined functions 15 Aggregating your data PART 4 - GOING FURTHER 16 Cache and checkpoint: Enhancing Spark’s performances 17 Exporting data and building full data pipelines 18 Exploring deployment |
Data Engineering with dbt - O'Reilly Media
Build a solid dbt base and learn data modeling and the modern data stack to become an analytics engineer; Build automated and reliable pipelines to deploy, test, run, and monitor ELTs with …
Analytics Engineering with SQL and dbt - O'Reilly Media
This practical book shows data analysts, data engineers, BI developers, and data scientists how to create a true self-service transformation platform through the use of dynamic SQL. Authors …
Data Engineering with dbt [Book] - O'Reilly Media
Get full access to Data Engineering with dbt and 60K+ other titles, with a free 10-day trial of O'Reilly. There are also live events, courses curated by job role, and more.
Preface - Data Engineering with dbt [Book] - O'Reilly Media
This book begins by introducing you to dbt and its role in the data stack, along with how it uses simple SQL to build your data platform, helping you and your team work better together. You’ll …
Chapter 5: Transforming Data with dbt - Data Engineering with …
Transforming Data with dbt. In this chapter, we will start to build a data application in dbt and touch on how to manage the data life cycle and organize the layers of our data application.
Analytics Engineering as the New Core of Data Engineering
Analytics engineering transforms raw data from disparate company data sources into information ready for use with tools that analysts and ... Get Data Engineering with dbt now with the …
B18572_TOC_ePUb - Data Engineering with dbt [Book] - O'Reilly …
Get full access to Data Engineering with dbt and 60K+ other titles, with a free 10-day trial of O'Reilly. There are also live events, courses curated by job role, and more.
Part 1: The Foundations of Data Engineering - Data Engineering …
In this section, you will get started on your path of learning how to build a data platform by learning the basics of SQL, modeling, and data engineering. This section includes the …
Data Engineering with dbt - O'Reilly Media
In this chapter, we will start to work hands-on with dbt. We will start by setting up a development environment, first creating a free GitHub account with a repository for your dbt project, then …
Data Engineering with dbt - O'Reilly Media
In previous chapters, we saw how to use the basic dbt functionalities to build an end-to-end data platform that transforms source data into usable information. In this chapter and the next one, …
Data Engineering with dbt - O'Reilly Media
Build a solid dbt base and learn data modeling and the modern data stack to become an analytics engineer; Build automated and reliable pipelines to …
Analytics Engineering with SQL and dbt - O'Reilly Media
This practical book shows data analysts, data engineers, BI developers, and data scientists how to create a true self-service transformation platform …
Data Engineering with dbt [Book] - O'Reilly Media
Get full access to Data Engineering with dbt and 60K+ other titles, with a free 10-day trial of O'Reilly. There are also live events, courses curated by job …
Preface - Data Engineering with dbt [Book] - O'Reilly Media
This book begins by introducing you to dbt and its role in the data stack, along with how it uses simple SQL to build your data platform, helping you and …
Chapter 5: Transforming Data with dbt - Data Engineering wi…
Transforming Data with dbt. In this chapter, we will start to build a data application in dbt and touch on how to manage the data life cycle and …