Serverless Data Processing with Dataflow: Develop Pipelines

Por: Coursera . en: , ,

  • Introduction
    • This module covers the course outline
  • Beam Concepts Review
    • Review main concepts of Apache Beam, and how to apply them to write your own data processing pipelines.
  • Windows, Watermarks Triggers
    • In this module, you will learn about how to process data in streaming with Dataflow. For that, there are three main concepts that you need to learn: how to group data in windows, the importance of watermark to know when the window is ready to produce results, and how you can control when and how many times the window will emit output.
  • Sources & Sinks
    • In this module, you will learn about what makes sources and sinks in Google Cloud Dataflow. The module will go over some examples of Text IO, FileIO, BigQueryIO, PubSub IO, KafKa IO, BigTable IO, Avro IO, and Splittable DoFn. The module will also point out some useful features associated with each IO.
  • Schemas
    • This module will introduce schemas, which give developers a way to express structured data in their Beam pipelines.
  • State and Timers
    • This module covers State and Timers, two powerful features that you can use in your DoFn to implement stateful transformations.
  • Best Practices
    • This module will discuss best practices and review common patterns that maximize performance for your Dataflow pipelines.
  • Dataflow SQL & DataFrames
    • This modules introduces two new APIs to represent your business logic in Beam: SQL and Dataframes.
  • Beam Notebooks
    • This module will cover Beam notebooks, an interface for Python developers to onboard onto the Beam SDK and develop their pipelines iteratively in a Jupyter notebook environment.
  • Summary
    • This module provides a recap of the course