Session

Better Together: Change Data Feed in a Streaming Data Flow

Overview

ExperienceIn Person
TypeBreakout
TrackData Engineering and Streaming
IndustryRetail and CPG - Food
TechnologiesApache Spark, Delta Lake, DLT
Skill LevelIntermediate

Traditional streaming works great when your data source is append-only, but what if your data source includes updates and deletes? At 84.51 we used DLT and Delta Lake to build a streaming data flow that consumes inserts, updates and deletes while still taking advantage of streaming checkpoints. We combined this flow with a materialized view and Enzyme incremental refresh for a low-code, efficient and robust end-to-end data flow.

 

We process around 8 million sales transactions each day with 80 million items purchased. This flow not only handles new transactions but also handles updates to previous transactions. Join us to learn how 84.51 combined change data feed, data streaming and materialized views to deliver a “better together” solution.

 

84.51 is a retail insights, media & marketing company. We use first-party retail data from 60 million households sourced through a loyalty card program to drive Kroger’s customer-centric journey.

Session Speakers

IMAGE COMING SOON

Scott Gordon

/Data Engineer
84.51˚