TrueschoTruescho
All Courses
Ensure Consistency in Streaming Pipelines
Coursera
Course
Unknown

Ensure Consistency in Streaming Pipelines

Coursera

Master the design and implementation of consistent streaming data pipelines using Apache Kafka, Spark, and Flink with proper delivery guarantees and failure handling.

Unknown3 weeksEnglish

About this Course

Master the design and implementation of consistent streaming data pipelines using Apache Kafka, Spark, and Flink. In this hands-on course, you'll apply systematic decision frameworks to select appropriate delivery guarantees (at-most-once, at-least-once, exactly-once) based on business requirements and failure scenario analysis. You'll implement end-to-end exactly-once processing by configuring Kafka producer transactions, Spark Structured Streaming checkpoints, and Hudi transactional tables, then validate your implementation through integration testing with failure injection. Finally, you'll evaluate watermarking strategies by analyzing event arrival patterns to optimize the latency-completeness tradeoff and meet specific SLA requirements. Through realistic scenarios—from preventing duplicate billing in order processing to optimizing IoT event pipelines for sub-10-second P95 latency—you'll develop the skills to architect production streaming systems that balance correctness, performance, and operational simplicity. Intermediate data and platform engineers using Kafka, Spark, or Flink who want to design production streaming pipelines with correct delivery guarantees, exactly-once semantics, and low-latency processing. Foundational knowledge of distributed systems; basic experience with Apache Kafka or similar messaging systems; familiarity with SQL; and introductory experience with stream or batch data processing concepts. By the end of this course, you will be able to design and validate production-ready streaming pipelines with correct delivery guarantees, exactly-once semantics, and low-latency event-time processing

What You'll Learn

  • Design streaming pipelines by analyzing failure scenarios and business needs
  • Implement exactly-once processing using transactions, checkpoints, and idempotency
  • Evaluate watermarking and windowing to optimize latency and completeness

Prerequisites

  • Basic familiarity with relevant technical terms
  • Willingness to engage in practical exercises and case studies

Instructors

S

Starweaver

Global Leaders in Professional & Technology Education

R

Ritesh Vajariya

Advisor | Leader | Speaker |Author

Topics

Software Development
Computer Science
Probability and Statistics
Data Science
Service Level
Data Integrity
Apache Kafka
Production Management
Verification And Validation
System Design and Implementation

Course Info

PlatformCoursera
LevelUnknown
PacingUnknown
PriceFree

Skills

تطوير البرمجيات
علوم الحاسوب
الاحتمالات والإحصاء
علوم البيانات
مستوى الخدمة
سلامة البيانات
أباتشي كافكا
إدارة الإنتاج
Verification And Validation
System Design and Implementation

Start Learning Now