Use coupon code “SUMMER20” for a 20% discount on all items! Valid until 2024-08-31

Site Logo
Search Suggestions

      Royal Mail  express delivery to UK destinations

      Regular sales and promotions

      Stock updates every 20 minutes!

      Data Pipelines with Apache Airflow

      12 in stock

      Firm sale: non returnable item
      SKU 9781617296901 Categories ,
      Select Guide Rating
      Pipelines can be challenging to manage, especially when your data has to flow through a collection of application components, servers, and cloud services. Airflow lets you schedule, restart, and backfill pipelines, and its easy-to-use UI and workflows with Python scripting has...

      £39.99

      Buy new:

      Delivery: UK delivery Only. Usually dispatched in 1-2 working days.

      Shipping costs: All shipping costs calculated in the cart or during the checkout process.

      Standard service (normally 2-3 working days): 48hr Tracked service.

      Premium service (next working day): 24hr Tracked service – signature service included.

      Royal mail: 24 & 48hr Tracked: Trackable items weighing up to 20kg are tracked to door and are inclusive of text and email with ‘Leave in Safe Place’ options, but are non-signature services. Examples of service expected: Standard 48hr service – if ordered before 3pm on Thursday then expected delivery would be on Saturday. If Premium 24hr service used, then expected delivery would be Friday.

      Signature Service: This service is only available for tracked items.

      Leave in Safe Place: This option is available at no additional charge for tracked services.

      Description

      Product ID:9781617296901
      Product Form:Paperback / softback
      Country of Manufacture:GB
      Title:Data Pipelines with Apache Airflow
      Authors:Author: Bas Harenslak, Julian Ruiter
      Page Count:425
      Subjects:Programming and scripting languages: general, Programming & scripting languages: general, Cloud computing, Information visualization, Cloud computing, Information visualization
      Description:Select Guide Rating
      Pipelines can be challenging to manage, especially when your data has to flow through a collection of application components, servers, and cloud services. Airflow lets you schedule, restart, and backfill pipelines, and its easy-to-use UI and workflows with Python scripting has users praising its incredible flexibility. Data Pipelines with Apache Airflow takes you through best practices for creating pipelines for multiple tasks, including data lakes, cloud deployments, and data science.   Data Pipelines with Apache Airflow teaches you the ins-and-outs of the Directed Acyclic Graphs (DAGs) that power Airflow, and how to write your own DAGs to meet the needs of your projects. With complete coverage of both foundational and lesser-known features, when you’re done you’ll be set to start using Airflow for seamless data pipeline development and management.   Key Features Framework foundation and best practices Airflow's execution and dependency system Testing Airflow DAGs Running Airflow in production   For data-savvy developers, DevOps and data engineers, and system administrators with intermediate Python skills.   About the technology Data pipelines are used to extract, transform and load data to and from multiple sources, routing it wherever it’s needed -- whether that’s visualisation tools, business intelligence dashboards, or machine learning models. Airflow streamlines the whole process, giving you one tool for programmatically developing and monitoring batch data pipelines, and integrating all the pieces you use in your data stack.   Bas Harenslak and Julian de Ruiter are data engineers with extensive experience using Airflow to develop pipelines for major companies including Heineken, Unilever, and Booking.com. Bas is a committer, and both Bas and Julian are active contributors to Apache Airflow.

      Pipelines can be challenging to manage, especially when your data has to flow through a collection of application components, servers, and cloud services. Airflow lets you schedule, restart, and backfill pipelines, and its easy-to-use UI and workflows with Python scripting has users praising its incredible flexibility. Data Pipelines with Apache Airflow takes you through best practices for creating pipelines for multiple tasks, including data lakes, cloud deployments, and data science.

       

      Data Pipelines with Apache Airflow teaches you the ins-and-outs of the Directed Acyclic Graphs (DAGs) that power Airflow, and how to write your own DAGs to meet the needs of your projects. With complete coverage of both foundational and lesser-known features, when you’re done you’ll be set to start using Airflow for seamless data pipeline development and management.

       

      Key Features

      Framework foundation and best practices

      Airflow''s execution and dependency system

      Testing Airflow DAGs

      Running Airflow in production

       

      For data-savvy developers, DevOps and data engineers, and system

      administrators with intermediate Python skills.

       

      About the technology

      Data pipelines are used to extract, transform and load data to and from multiple sources, routing it wherever it’s needed -- whether that’s visualisation tools, business intelligence dashboards, or machine learning models. Airflow streamlines the whole process, giving you one tool for programmatically developing and monitoring batch data pipelines, and integrating all the pieces you use in your data stack.

       

      Bas Harenslak and Julian de Ruiter are data engineers with extensive experience using Airflow to develop pipelines for major companies including Heineken, Unilever, and Booking.com. Bas is a committer, and both Bas and Julian are active contributors to Apache Airflow.


      Imprint Name:Manning Publications
      Publisher Name:Manning Publications
      Country of Publication:GB
      Publishing Date:2021-06-07

      Additional information

      Weight872 g
      Dimensions188 × 237 × 30 mm