SlideShare a Scribd company logo
Apache Kafka and the Data Mesh
Ben Stopford, Michael G. Noll
Office of the CTO, Confluent
Kafka Summit Europe, May 11-12, 2021
What is a Data Mesh?
2
1. An implementation Pattern (not unlike Microservices)
2. 4 Principles
3. Technology agnostic
Several historical influences
3
DDD Microservices
Data Marts Event Streaming
Data on the Inside /
Data on the Outside
Data “in practice” Needs More Discipline
4
Data as a Practice
… is not on the same level.
Software as a Practice
Spaghetti: Data architectures often lack rigour
5
Kafka provides a solution. The implementation.
6
Kafka
Centralize an immutable stream of facts. Decentralize the freedom to act, adapt, and change.
Data Mesh with Event Streaming
Domain-driven
Decentralization
Local Autonomy Per
Domain
(Organizational
Concerns)
Data as a
First-class Product
Product thinking,
“Microservice for
Data”
Federated
Governance
Interoperability Across
Domains, Network
Effects
(Organizational
Concerns)
Self-serve
Data Platform
Infrastructure as a
Platform
Across Domains
1 2 3 4
The Principles of a Data Mesh
10
Data Mesh
A First Look
Domain
Inventory
Shipments
Orders
...
Data
Product
Domain-driven
Decentralization
Local Autonomy
(Organizational
Concerns)
Data as a
First-class Product
Product thinking,
“Microservice for
Data”
Federated
Governance
Interoperability,
Network Effects
(Organizational
Concerns)
Self-serve
Data Platform
Infra Tooling,
Across Domains
1 2 3 4
The Principles of a Data Mesh
Principle 1: Domain-driven Decentralization
Anti-pattern: responsibility for data
becomes the domain of the DWH team
Pattern: Ownership of a data asset given to the
“local” team that is most familiar with it
Centralized
Data Ownership
Decentralized
Data Ownership
Objective: Ensure data is owned by those that truly understand it
13
Billing Recommendations
Inventory
Orders
Domain
Shipment
Domain
Shipping Data
Order Data
1. Joe in Inventory has a problem with
Order data.
2. Inventory items are going negative,
because of bad Order data.
3. He could fix the data up locally in the
Inventory domain, and get on with his
job.
4. Or, better, he contacts Alice in
Orders and get it fixed at the source.
This is more reliable as Joe doesn’t
fully understand the Orders process.
5. Ergo, Alice needs be an responsible
& responsive “Data Product Owner”,
so everyone benefits from the fix to
Joe’s problem.
Joe
Alice
Practical example
Recommendations: Domain-driven Decentralization
14
Learn from DDD:
- Use a standard language and nomenclature for data.
- Business users should understand a data flow diagram.
- The stream of events should create a shared narrative that is business-user comprehensible.
Domain-driven
Decentralization
Local Autonomy
(Organizational
Concerns)
Data as a
First-class Product
Product thinking,
“Microservice for
Data”
Federated
Governance
Interoperability,
Network Effects
(Organizational
Concerns)
Self-serve
Data Platform
Infra Tooling,
Across Domains
1 2 3 4
The Principles of a Data Mesh
Principle 2: Data as a First-Class Product
16
• Objective: Make shared data discoverable, addressable, trustworthy, secure,
so other teams can make good use of it.
• Data is treated as a true product, not a by-product.
This product thinking is important to prevent data chauvinism.
Infra
Code
Data product, a “microservice for the data world”
17
• Data product is a node on the data mesh, situated within a domain.
• Produces—and possibly consumes—high-quality data within the mesh.
• Encapsulates all the elements required for its function, namely data + code + infrastructure.
Data
Creates, manipulates,
serves, etc. that data
Powers the data (e.g., storage) and the code
(e.g., run, deploy, monitor)
“Items about to expire”
Data Product
Data and metadata,
including history
Domain
18
Data Mesh
Data
Product
Connectivity within the mesh lends itself...
Domain
19
Data
Product
...naturally to Event Streaming with Kafka
Data Mesh
20
Event Streaming is Pub/Sub, not Point-to-Point
Data
Product
Data
Product
Data
Product
Data
Product
stream
(persisted) other
streams
write
(publish)
read
(consume)
independently
Data producers are scalably decoupled from consumers.
Data Product
Data Product
Why is Event Streaming a good fit for meshing?
21
0 1 2 3 4 5 6 1
7
Streams are real-time, low latency ⇒ Propagate data immediately.
Streams are highly scalable ⇒ Handle today’s massive data
volumes.
Streams are stored, replayable ⇒ Capture real-time & historical
data.
Streams are immutable ⇒ Auditable source of
How to get data into & out of a data product
22
Data Product
Input
Data
Ports
Output
Data
Ports
Snapshot via
Nightly ETL
Snapshot via
Nighty ETL
Continuous
Stream
Snapshot via
Req/Res API
Snapshot via
Req/Res API
A
B
C
Continuous
Stream
Onboarding existing data
23
Data
Product
Input
Data
Ports
Source
Connectors
Use Kafka connectors to stream data from cloud services and existing systems into the
mesh.
https://www.confluent.io/hub/
Data product: what’s happening inside
24
Input
Data
Ports
Output
Data
Ports
…pick your favorites...
Data on the Inside: HOW the domain team solves specific problems
internally? This doesn’t matter to other domains.
Event Streaming inside a data product
25
Input
Data
Ports
Output
Data
Ports
ksqlDB to filter,
process, join,
aggregate, analyze
Stream data from other
DPs or internal systems
into ksqlDB
1 2 Stream data to internal
systems or the outside.
Pull queries can drive a
req/res API.
3
Req/Res API
Pull Queries
Use ksqlDB, Kafka Streams apps, etc. for processing data in motion.
Event Streaming inside a data product
26
Input
Data
Ports
Output
Data
Ports
MySQL
Sink Connector Source
Connector
DB client apps work
as usual
Stream data from other
Data Products into your
local DB
Stream data to the outside with
CDC and e.g. the Outbox
Pattern, ksqlDB, etc.
1 3
2
Use Kafka connectors and CDC to “streamify” classic databases.
Dealing with data change: schemas & versioning
27
Data
Product
Output
Data
Ports
V1 - user, product, quantity
V2 - userAnonymized, product, quantity
Publish evolving streams with back/forward-compatible schemas.
Publish versioned streams for breaking changes.
Also, when needed, data can be fully reprocessed by replaying history.
Recommendations: Data as a First-class Product
28
1. Data-on-the-Outside is harder to change, but it has more value in a holistic sense.
a. Use schemas as a contract.
b. Handle incompatible schema changes using Dual Schema Upgrade Window pattern.
2. Get data from the source, not from intermediaries. Think: Demeter's law applied to data.
a. Otherwise, proliferation of ‘slightly corrupt’ data within the mesh. “Game of Telephone”.
b. Event Streaming makes it easy to subscribe to data from authoritative sources.
3. Change data at the source, including error fixes. Don’t “fix data up” locally.
4. Some data sources will be difficult to turn into first-class data products. Example: Batch-based
sources that lose event-level data or are not reproducible.
a. Use Event Streaming plus CDC, Outbox Pattern, etc. to integrate these into the mesh.
Domain-driven
Decentralization
Local Autonomy
(Organizational
Concerns)
Data as a
First-class Product
Product thinking,
“Microservice for
Data”
Federated
Governance
Interoperability,
Network Effects
(Organizational
Concerns)
Self-serve
Data Platform
Infra Tooling,
Across Domains
1 2 3 4
The Principles of a Data Mesh
Why Self-service Matters
30
Trade Surveillance System
● Data from 13 sources
● Some sources publish events
● Needed both historical and real-time data
● Historical data from database extracts arranged with dev
team.
● Format of events different to format of extracts
● 9 months of effort to get 13 sources into the new system.
Principle 3: Self-serve Data Platform
31
Central infrastructure that provides real-time and historical data on demand
Objective: Make domains autonomous in their execution through rapid data provisioning
Consuming real-time & historical data from the mesh
32
1) Separate Systems for Real-time and Historical Data (Lambda Architecture)
- Considerations:
- Difficulty to correlate real-time with historical “snapshot” data
- Two systems to manage
- Unlike event streams, snapshots have less granularity
1) One System for Real-time and Historical Data (Kappa Architecture)
- Considerations:
- Operational complexity (addressed in Confluent Cloud)
- Downsides of immutability of regular streams: e.g. altering or deleting events
- Storage cost (addressed in Confluent Cloud, in Apache Kafka with KIP-405)
What this can look like in practice
33
Browse Schemas
Implementation: Database Inside-Out
DB
DB
CONNECTOR
CONNECTOR
STREAM
PROCESSOR
DB/
View
Messaging that
Remembers
ksqlDB
35
With ksqlDB the Data Mesh is Queryable and
Decentralized
Data
Product
STREAM
PROCESSOR
ksqlDB
Query is the interface to
the mesh
Events are the interface to the
mesh
Think: Infrastructure as code, but for data
36
Code
Container
Image
+ Same APP
every time
Code
Event
Streams
+ Same DATA
every time
37
Mesh is one logical cluster. Data product has another.
Data
Product
Data Product has its own
cluster for internal use
Domain-driven
Decentralization
Local Autonomy
(Organizational
Concerns)
Data as a
First-class Product
Product thinking,
“Microservice for
Data”
Federated
Governance
Interoperability,
Network Effects
(Organizational
Concerns)
Self-serve
Data Platform
Infra Tooling,
Across Domains
1 2 3 4
The Principles of a Data Mesh
Principle 4: Federated Governance
39
• Objective: Independent data products can interoperate and create network effects.
• Establish global standards, like governance, that apply to all data products in the mesh.
• Ideally, these global standards and rules are applied automatically by the platform.
Domain Domain Domain Domain
Self-serve Data Platform
What is decided locally
by a domain?
What is globally?
(implemented and
enforced by platform)
Must balance between Decentralization vs. Centralization. No silver bullet!
Example standard: Identifying customers globally
• Define how data is represented, so you can join and correlate data across different domains.
• Use data contracts, schemas, registries, etc. to implement and enforce such standards.
• Use Event Streaming to retrofit historical data to new requirements, standards.
40
customerId=29639
customerId=29639
customerId=29639
customerId=29639
SELECT … FROM orders o
LEFT JOIN shipments s
ON o.customerId = s.customerId
EMIT CHANGES;
Example standard: Detect errors and recover with Streams
41
• Use strategies like logging, data profiling, data lineage, etc. to detect errors in the mesh.
• Streams are very helpful to detect errors and identify cause-effect relationships.
• Streams let you recover and fix errors: e.g., replay & reprocess historical data.
Data
Product
Output
Data
Ports
0 1 2 3 4 5 6 7 8 9
My App
Bug? Error? Rewind to
start of stream, then
reprocess.
If needed, tell the origin data product to fix problematic data at the source.
Event Streams give
you a powerful
Time Machine.
Example standard: Tracking data lineage with Streams
42
• Lineage must work across domains and data products—and systems, clouds, data centers.
• Event streaming is a foundational technology for this.
On-premise
Recommendations: Federated Governance
43
1. Be pragmatic: Don’t expect governance systems to be perfect.
a. They are a map that helps you navigate the data-landscape of your company.
b. But there will always be roads that have changed or have not been mapped.
2. Governance is more a process—i.e., an organizational concern—than a technology.
1. Beware of centralized data models, which can become slow to change. Where they must exist, use
processes & tooling like GitHub to collaborate and change quickly. Good luck! 🙂
Data mesh journey
44
Principle 1
Data should have one owner: the
team that creates it.
Principle 2
Data is your product:
All exposed data should be
good data.
Principle 3
Get access to any data
immediately and painlessly, be it
historical or real-time.
Principle 4: Governance, with standards, security, lineage, etc.
(cross-cutting concerns)
Difficulty
to execute
Start Here
1
2
3
Thank you!
@benstopford
@miguno
Ad

More Related Content

What's hot (20)

Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Tristan Baker
 
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
HostedbyConfluent
 
Data Lake Overview
Data Lake OverviewData Lake Overview
Data Lake Overview
James Serra
 
Architect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh ArchitectureArchitect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh Architecture
Databricks
 
Data Mesh 101
Data Mesh 101Data Mesh 101
Data Mesh 101
ChrisFord803185
 
Data Mesh
Data MeshData Mesh
Data Mesh
Piethein Strengholt
 
DW Migration Webinar-March 2022.pptx
DW Migration Webinar-March 2022.pptxDW Migration Webinar-March 2022.pptx
DW Migration Webinar-March 2022.pptx
Databricks
 
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
HostedbyConfluent
 
Webinar future dataintegration-datamesh-and-goldengatekafka
Webinar future dataintegration-datamesh-and-goldengatekafkaWebinar future dataintegration-datamesh-and-goldengatekafka
Webinar future dataintegration-datamesh-and-goldengatekafka
Jeffrey T. Pollock
 
Snowflake Architecture.pptx
Snowflake Architecture.pptxSnowflake Architecture.pptx
Snowflake Architecture.pptx
chennakesava44
 
From Insights to Action, How to build and maintain a Data Driven Organization...
From Insights to Action, How to build and maintain a Data Driven Organization...From Insights to Action, How to build and maintain a Data Driven Organization...
From Insights to Action, How to build and maintain a Data Driven Organization...
Amazon Web Services Korea
 
Microsoft Data Platform - What's included
Microsoft Data Platform - What's includedMicrosoft Data Platform - What's included
Microsoft Data Platform - What's included
James Serra
 
Building a modern data warehouse
Building a modern data warehouseBuilding a modern data warehouse
Building a modern data warehouse
James Serra
 
Data Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to MeshData Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to Mesh
Jeffrey T. Pollock
 
Unleashing the value of metadata with Talend
Unleashing the value of metadata with Talend Unleashing the value of metadata with Talend
Unleashing the value of metadata with Talend
Jean-Michel Franco
 
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
Lace Lofranco
 
Achieving Lakehouse Models with Spark 3.0
Achieving Lakehouse Models with Spark 3.0Achieving Lakehouse Models with Spark 3.0
Achieving Lakehouse Models with Spark 3.0
Databricks
 
Evolution from EDA to Data Mesh: Data in Motion
Evolution from EDA to Data Mesh: Data in MotionEvolution from EDA to Data Mesh: Data in Motion
Evolution from EDA to Data Mesh: Data in Motion
confluent
 
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
DATAVERSITY
 
Microsoft Azure Data Factory Hands-On Lab Overview Slides
Microsoft Azure Data Factory Hands-On Lab Overview SlidesMicrosoft Azure Data Factory Hands-On Lab Overview Slides
Microsoft Azure Data Factory Hands-On Lab Overview Slides
Mark Kromer
 
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Tristan Baker
 
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
How to Build the Data Mesh Foundation: A Principled Approach | Zhamak Dehghan...
HostedbyConfluent
 
Data Lake Overview
Data Lake OverviewData Lake Overview
Data Lake Overview
James Serra
 
Architect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh ArchitectureArchitect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh Architecture
Databricks
 
DW Migration Webinar-March 2022.pptx
DW Migration Webinar-March 2022.pptxDW Migration Webinar-March 2022.pptx
DW Migration Webinar-March 2022.pptx
Databricks
 
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
Standing on the Shoulders of Open-Source Giants: The Serverless Realtime Lake...
HostedbyConfluent
 
Webinar future dataintegration-datamesh-and-goldengatekafka
Webinar future dataintegration-datamesh-and-goldengatekafkaWebinar future dataintegration-datamesh-and-goldengatekafka
Webinar future dataintegration-datamesh-and-goldengatekafka
Jeffrey T. Pollock
 
Snowflake Architecture.pptx
Snowflake Architecture.pptxSnowflake Architecture.pptx
Snowflake Architecture.pptx
chennakesava44
 
From Insights to Action, How to build and maintain a Data Driven Organization...
From Insights to Action, How to build and maintain a Data Driven Organization...From Insights to Action, How to build and maintain a Data Driven Organization...
From Insights to Action, How to build and maintain a Data Driven Organization...
Amazon Web Services Korea
 
Microsoft Data Platform - What's included
Microsoft Data Platform - What's includedMicrosoft Data Platform - What's included
Microsoft Data Platform - What's included
James Serra
 
Building a modern data warehouse
Building a modern data warehouseBuilding a modern data warehouse
Building a modern data warehouse
James Serra
 
Data Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to MeshData Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to Mesh
Jeffrey T. Pollock
 
Unleashing the value of metadata with Talend
Unleashing the value of metadata with Talend Unleashing the value of metadata with Talend
Unleashing the value of metadata with Talend
Jean-Michel Franco
 
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
DataOps for the Modern Data Warehouse on Microsoft Azure @ NDCOslo 2020 - Lac...
Lace Lofranco
 
Achieving Lakehouse Models with Spark 3.0
Achieving Lakehouse Models with Spark 3.0Achieving Lakehouse Models with Spark 3.0
Achieving Lakehouse Models with Spark 3.0
Databricks
 
Evolution from EDA to Data Mesh: Data in Motion
Evolution from EDA to Data Mesh: Data in MotionEvolution from EDA to Data Mesh: Data in Motion
Evolution from EDA to Data Mesh: Data in Motion
confluent
 
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...
DATAVERSITY
 
Microsoft Azure Data Factory Hands-On Lab Overview Slides
Microsoft Azure Data Factory Hands-On Lab Overview SlidesMicrosoft Azure Data Factory Hands-On Lab Overview Slides
Microsoft Azure Data Factory Hands-On Lab Overview Slides
Mark Kromer
 

Similar to Apache Kafka and the Data Mesh | Ben Stopford and Michael Noll, Confluent (20)

Apache Kafka® and the Data Mesh
Apache Kafka® and the Data MeshApache Kafka® and the Data Mesh
Apache Kafka® and the Data Mesh
ConfluentInc1
 
Apache Kafka and the Data Mesh | Michael Noll, Confluent
Apache Kafka and the Data Mesh | Michael Noll, ConfluentApache Kafka and the Data Mesh | Michael Noll, Confluent
Apache Kafka and the Data Mesh | Michael Noll, Confluent
HostedbyConfluent
 
data-mesh-101.pptx
data-mesh-101.pptxdata-mesh-101.pptx
data-mesh-101.pptx
TarekHamdi8
 
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
Denodo
 
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
confluent
 
Off-Label Data Mesh: A Prescription for Healthier Data
Off-Label Data Mesh: A Prescription for Healthier DataOff-Label Data Mesh: A Prescription for Healthier Data
Off-Label Data Mesh: A Prescription for Healthier Data
HostedbyConfluent
 
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Denodo
 
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Denodo
 
The Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
The Enterprise Guide to Building a Data Mesh - Introducing SpecMeshThe Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
The Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
IanFurlong4
 
From Single Purpose to Multi Purpose Data Lakes - Broadening End Users
From Single Purpose to Multi Purpose Data Lakes - Broadening End UsersFrom Single Purpose to Multi Purpose Data Lakes - Broadening End Users
From Single Purpose to Multi Purpose Data Lakes - Broadening End Users
Denodo
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Denodo
 
A Logical Architecture is Always a Flexible Architecture (ASEAN)
A Logical Architecture is Always a Flexible Architecture (ASEAN)A Logical Architecture is Always a Flexible Architecture (ASEAN)
A Logical Architecture is Always a Flexible Architecture (ASEAN)
Denodo
 
Bridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need ItBridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need It
Denodo
 
Building a Logical Data Fabric using Data Virtualization (ASEAN)
Building a Logical Data Fabric using Data Virtualization (ASEAN)Building a Logical Data Fabric using Data Virtualization (ASEAN)
Building a Logical Data Fabric using Data Virtualization (ASEAN)
Denodo
 
Why Data Virtualization? An Introduction
Why Data Virtualization? An IntroductionWhy Data Virtualization? An Introduction
Why Data Virtualization? An Introduction
Denodo
 
Data virtualization an introduction
Data virtualization an introductionData virtualization an introduction
Data virtualization an introduction
Denodo
 
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Denodo
 
Flash session -streaming--ses1243-lon
Flash session -streaming--ses1243-lonFlash session -streaming--ses1243-lon
Flash session -streaming--ses1243-lon
Jeffrey T. Pollock
 
Best Practices in the Cloud for Data Management (US)
Best Practices in the Cloud for Data Management (US)Best Practices in the Cloud for Data Management (US)
Best Practices in the Cloud for Data Management (US)
Denodo
 
Data Virtualization: Introduction and Business Value (UK)
Data Virtualization: Introduction and Business Value (UK)Data Virtualization: Introduction and Business Value (UK)
Data Virtualization: Introduction and Business Value (UK)
Denodo
 
Apache Kafka® and the Data Mesh
Apache Kafka® and the Data MeshApache Kafka® and the Data Mesh
Apache Kafka® and the Data Mesh
ConfluentInc1
 
Apache Kafka and the Data Mesh | Michael Noll, Confluent
Apache Kafka and the Data Mesh | Michael Noll, ConfluentApache Kafka and the Data Mesh | Michael Noll, Confluent
Apache Kafka and the Data Mesh | Michael Noll, Confluent
HostedbyConfluent
 
data-mesh-101.pptx
data-mesh-101.pptxdata-mesh-101.pptx
data-mesh-101.pptx
TarekHamdi8
 
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
Denodo
 
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
Event Driven Architecture with a RESTful Microservices Architecture (Kyle Ben...
confluent
 
Off-Label Data Mesh: A Prescription for Healthier Data
Off-Label Data Mesh: A Prescription for Healthier DataOff-Label Data Mesh: A Prescription for Healthier Data
Off-Label Data Mesh: A Prescription for Healthier Data
HostedbyConfluent
 
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Denodo
 
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)
Denodo
 
The Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
The Enterprise Guide to Building a Data Mesh - Introducing SpecMeshThe Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
The Enterprise Guide to Building a Data Mesh - Introducing SpecMesh
IanFurlong4
 
From Single Purpose to Multi Purpose Data Lakes - Broadening End Users
From Single Purpose to Multi Purpose Data Lakes - Broadening End UsersFrom Single Purpose to Multi Purpose Data Lakes - Broadening End Users
From Single Purpose to Multi Purpose Data Lakes - Broadening End Users
Denodo
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Denodo
 
A Logical Architecture is Always a Flexible Architecture (ASEAN)
A Logical Architecture is Always a Flexible Architecture (ASEAN)A Logical Architecture is Always a Flexible Architecture (ASEAN)
A Logical Architecture is Always a Flexible Architecture (ASEAN)
Denodo
 
Bridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need ItBridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need It
Denodo
 
Building a Logical Data Fabric using Data Virtualization (ASEAN)
Building a Logical Data Fabric using Data Virtualization (ASEAN)Building a Logical Data Fabric using Data Virtualization (ASEAN)
Building a Logical Data Fabric using Data Virtualization (ASEAN)
Denodo
 
Why Data Virtualization? An Introduction
Why Data Virtualization? An IntroductionWhy Data Virtualization? An Introduction
Why Data Virtualization? An Introduction
Denodo
 
Data virtualization an introduction
Data virtualization an introductionData virtualization an introduction
Data virtualization an introduction
Denodo
 
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Data Fabric - Why Should Organizations Implement a Logical and Not a Physical...
Denodo
 
Flash session -streaming--ses1243-lon
Flash session -streaming--ses1243-lonFlash session -streaming--ses1243-lon
Flash session -streaming--ses1243-lon
Jeffrey T. Pollock
 
Best Practices in the Cloud for Data Management (US)
Best Practices in the Cloud for Data Management (US)Best Practices in the Cloud for Data Management (US)
Best Practices in the Cloud for Data Management (US)
Denodo
 
Data Virtualization: Introduction and Business Value (UK)
Data Virtualization: Introduction and Business Value (UK)Data Virtualization: Introduction and Business Value (UK)
Data Virtualization: Introduction and Business Value (UK)
Denodo
 
Ad

More from HostedbyConfluent (20)

Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
HostedbyConfluent
 
Renaming a Kafka Topic | Kafka Summit London
Renaming a Kafka Topic | Kafka Summit LondonRenaming a Kafka Topic | Kafka Summit London
Renaming a Kafka Topic | Kafka Summit London
HostedbyConfluent
 
Evolution of NRT Data Ingestion Pipeline at Trendyol
Evolution of NRT Data Ingestion Pipeline at TrendyolEvolution of NRT Data Ingestion Pipeline at Trendyol
Evolution of NRT Data Ingestion Pipeline at Trendyol
HostedbyConfluent
 
Ensuring Kafka Service Resilience: A Dive into Health-Checking Techniques
Ensuring Kafka Service Resilience: A Dive into Health-Checking TechniquesEnsuring Kafka Service Resilience: A Dive into Health-Checking Techniques
Ensuring Kafka Service Resilience: A Dive into Health-Checking Techniques
HostedbyConfluent
 
Exactly-once Stream Processing with Arroyo and Kafka
Exactly-once Stream Processing with Arroyo and KafkaExactly-once Stream Processing with Arroyo and Kafka
Exactly-once Stream Processing with Arroyo and Kafka
HostedbyConfluent
 
Fish Plays Pokemon | Kafka Summit London
Fish Plays Pokemon | Kafka Summit LondonFish Plays Pokemon | Kafka Summit London
Fish Plays Pokemon | Kafka Summit London
HostedbyConfluent
 
Tiered Storage 101 | Kafla Summit London
Tiered Storage 101 | Kafla Summit LondonTiered Storage 101 | Kafla Summit London
Tiered Storage 101 | Kafla Summit London
HostedbyConfluent
 
Building a Self-Service Stream Processing Portal: How And Why
Building a Self-Service Stream Processing Portal: How And WhyBuilding a Self-Service Stream Processing Portal: How And Why
Building a Self-Service Stream Processing Portal: How And Why
HostedbyConfluent
 
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
HostedbyConfluent
 
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
HostedbyConfluent
 
Navigating Private Network Connectivity Options for Kafka Clusters
Navigating Private Network Connectivity Options for Kafka ClustersNavigating Private Network Connectivity Options for Kafka Clusters
Navigating Private Network Connectivity Options for Kafka Clusters
HostedbyConfluent
 
Apache Flink: Building a Company-wide Self-service Streaming Data Platform
Apache Flink: Building a Company-wide Self-service Streaming Data PlatformApache Flink: Building a Company-wide Self-service Streaming Data Platform
Apache Flink: Building a Company-wide Self-service Streaming Data Platform
HostedbyConfluent
 
Explaining How Real-Time GenAI Works in a Noisy Pub
Explaining How Real-Time GenAI Works in a Noisy PubExplaining How Real-Time GenAI Works in a Noisy Pub
Explaining How Real-Time GenAI Works in a Noisy Pub
HostedbyConfluent
 
TL;DR Kafka Metrics | Kafka Summit London
TL;DR Kafka Metrics | Kafka Summit LondonTL;DR Kafka Metrics | Kafka Summit London
TL;DR Kafka Metrics | Kafka Summit London
HostedbyConfluent
 
A Window Into Your Kafka Streams Tasks | KSL
A Window Into Your Kafka Streams Tasks | KSLA Window Into Your Kafka Streams Tasks | KSL
A Window Into Your Kafka Streams Tasks | KSL
HostedbyConfluent
 
Mastering Kafka Producer Configs: A Guide to Optimizing Performance
Mastering Kafka Producer Configs: A Guide to Optimizing PerformanceMastering Kafka Producer Configs: A Guide to Optimizing Performance
Mastering Kafka Producer Configs: A Guide to Optimizing Performance
HostedbyConfluent
 
Data Contracts Management: Schema Registry and Beyond
Data Contracts Management: Schema Registry and BeyondData Contracts Management: Schema Registry and Beyond
Data Contracts Management: Schema Registry and Beyond
HostedbyConfluent
 
Code-First Approach: Crafting Efficient Flink Apps
Code-First Approach: Crafting Efficient Flink AppsCode-First Approach: Crafting Efficient Flink Apps
Code-First Approach: Crafting Efficient Flink Apps
HostedbyConfluent
 
Debezium vs. the World: An Overview of the CDC Ecosystem
Debezium vs. the World: An Overview of the CDC EcosystemDebezium vs. the World: An Overview of the CDC Ecosystem
Debezium vs. the World: An Overview of the CDC Ecosystem
HostedbyConfluent
 
Beyond Tiered Storage: Serverless Kafka with No Local Disks
Beyond Tiered Storage: Serverless Kafka with No Local DisksBeyond Tiered Storage: Serverless Kafka with No Local Disks
Beyond Tiered Storage: Serverless Kafka with No Local Disks
HostedbyConfluent
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
HostedbyConfluent
 
Renaming a Kafka Topic | Kafka Summit London
Renaming a Kafka Topic | Kafka Summit LondonRenaming a Kafka Topic | Kafka Summit London
Renaming a Kafka Topic | Kafka Summit London
HostedbyConfluent
 
Evolution of NRT Data Ingestion Pipeline at Trendyol
Evolution of NRT Data Ingestion Pipeline at TrendyolEvolution of NRT Data Ingestion Pipeline at Trendyol
Evolution of NRT Data Ingestion Pipeline at Trendyol
HostedbyConfluent
 
Ensuring Kafka Service Resilience: A Dive into Health-Checking Techniques
Ensuring Kafka Service Resilience: A Dive into Health-Checking TechniquesEnsuring Kafka Service Resilience: A Dive into Health-Checking Techniques
Ensuring Kafka Service Resilience: A Dive into Health-Checking Techniques
HostedbyConfluent
 
Exactly-once Stream Processing with Arroyo and Kafka
Exactly-once Stream Processing with Arroyo and KafkaExactly-once Stream Processing with Arroyo and Kafka
Exactly-once Stream Processing with Arroyo and Kafka
HostedbyConfluent
 
Fish Plays Pokemon | Kafka Summit London
Fish Plays Pokemon | Kafka Summit LondonFish Plays Pokemon | Kafka Summit London
Fish Plays Pokemon | Kafka Summit London
HostedbyConfluent
 
Tiered Storage 101 | Kafla Summit London
Tiered Storage 101 | Kafla Summit LondonTiered Storage 101 | Kafla Summit London
Tiered Storage 101 | Kafla Summit London
HostedbyConfluent
 
Building a Self-Service Stream Processing Portal: How And Why
Building a Self-Service Stream Processing Portal: How And WhyBuilding a Self-Service Stream Processing Portal: How And Why
Building a Self-Service Stream Processing Portal: How And Why
HostedbyConfluent
 
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
From the Trenches: Improving Kafka Connect Source Connector Ingestion from 7 ...
HostedbyConfluent
 
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
Future with Zero Down-Time: End-to-end Resiliency with Chaos Engineering and ...
HostedbyConfluent
 
Navigating Private Network Connectivity Options for Kafka Clusters
Navigating Private Network Connectivity Options for Kafka ClustersNavigating Private Network Connectivity Options for Kafka Clusters
Navigating Private Network Connectivity Options for Kafka Clusters
HostedbyConfluent
 
Apache Flink: Building a Company-wide Self-service Streaming Data Platform
Apache Flink: Building a Company-wide Self-service Streaming Data PlatformApache Flink: Building a Company-wide Self-service Streaming Data Platform
Apache Flink: Building a Company-wide Self-service Streaming Data Platform
HostedbyConfluent
 
Explaining How Real-Time GenAI Works in a Noisy Pub
Explaining How Real-Time GenAI Works in a Noisy PubExplaining How Real-Time GenAI Works in a Noisy Pub
Explaining How Real-Time GenAI Works in a Noisy Pub
HostedbyConfluent
 
TL;DR Kafka Metrics | Kafka Summit London
TL;DR Kafka Metrics | Kafka Summit LondonTL;DR Kafka Metrics | Kafka Summit London
TL;DR Kafka Metrics | Kafka Summit London
HostedbyConfluent
 
A Window Into Your Kafka Streams Tasks | KSL
A Window Into Your Kafka Streams Tasks | KSLA Window Into Your Kafka Streams Tasks | KSL
A Window Into Your Kafka Streams Tasks | KSL
HostedbyConfluent
 
Mastering Kafka Producer Configs: A Guide to Optimizing Performance
Mastering Kafka Producer Configs: A Guide to Optimizing PerformanceMastering Kafka Producer Configs: A Guide to Optimizing Performance
Mastering Kafka Producer Configs: A Guide to Optimizing Performance
HostedbyConfluent
 
Data Contracts Management: Schema Registry and Beyond
Data Contracts Management: Schema Registry and BeyondData Contracts Management: Schema Registry and Beyond
Data Contracts Management: Schema Registry and Beyond
HostedbyConfluent
 
Code-First Approach: Crafting Efficient Flink Apps
Code-First Approach: Crafting Efficient Flink AppsCode-First Approach: Crafting Efficient Flink Apps
Code-First Approach: Crafting Efficient Flink Apps
HostedbyConfluent
 
Debezium vs. the World: An Overview of the CDC Ecosystem
Debezium vs. the World: An Overview of the CDC EcosystemDebezium vs. the World: An Overview of the CDC Ecosystem
Debezium vs. the World: An Overview of the CDC Ecosystem
HostedbyConfluent
 
Beyond Tiered Storage: Serverless Kafka with No Local Disks
Beyond Tiered Storage: Serverless Kafka with No Local DisksBeyond Tiered Storage: Serverless Kafka with No Local Disks
Beyond Tiered Storage: Serverless Kafka with No Local Disks
HostedbyConfluent
 
Ad

Recently uploaded (20)

Big Data Analytics Quick Research Guide by Arthur Morgan
Big Data Analytics Quick Research Guide by Arthur MorganBig Data Analytics Quick Research Guide by Arthur Morgan
Big Data Analytics Quick Research Guide by Arthur Morgan
Arthur Morgan
 
Into The Box Conference Keynote Day 1 (ITB2025)
Into The Box Conference Keynote Day 1 (ITB2025)Into The Box Conference Keynote Day 1 (ITB2025)
Into The Box Conference Keynote Day 1 (ITB2025)
Ortus Solutions, Corp
 
What is Model Context Protocol(MCP) - The new technology for communication bw...
What is Model Context Protocol(MCP) - The new technology for communication bw...What is Model Context Protocol(MCP) - The new technology for communication bw...
What is Model Context Protocol(MCP) - The new technology for communication bw...
Vishnu Singh Chundawat
 
How Can I use the AI Hype in my Business Context?
How Can I use the AI Hype in my Business Context?How Can I use the AI Hype in my Business Context?
How Can I use the AI Hype in my Business Context?
Daniel Lehner
 
Technology Trends in 2025: AI and Big Data Analytics
Technology Trends in 2025: AI and Big Data AnalyticsTechnology Trends in 2025: AI and Big Data Analytics
Technology Trends in 2025: AI and Big Data Analytics
InData Labs
 
Procurement Insights Cost To Value Guide.pptx
Procurement Insights Cost To Value Guide.pptxProcurement Insights Cost To Value Guide.pptx
Procurement Insights Cost To Value Guide.pptx
Jon Hansen
 
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfThe Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
Abi john
 
Heap, Types of Heap, Insertion and Deletion
Heap, Types of Heap, Insertion and DeletionHeap, Types of Heap, Insertion and Deletion
Heap, Types of Heap, Insertion and Deletion
Jaydeep Kale
 
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxDevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
Justin Reock
 
Cybersecurity Identity and Access Solutions using Azure AD
Cybersecurity Identity and Access Solutions using Azure ADCybersecurity Identity and Access Solutions using Azure AD
Cybersecurity Identity and Access Solutions using Azure AD
VICTOR MAESTRE RAMIREZ
 
Complete Guide to Advanced Logistics Management Software in Riyadh.pdf
Complete Guide to Advanced Logistics Management Software in Riyadh.pdfComplete Guide to Advanced Logistics Management Software in Riyadh.pdf
Complete Guide to Advanced Logistics Management Software in Riyadh.pdf
Software Company
 
Cyber Awareness overview for 2025 month of security
Cyber Awareness overview for 2025 month of securityCyber Awareness overview for 2025 month of security
Cyber Awareness overview for 2025 month of security
riccardosl1
 
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-UmgebungenHCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
panagenda
 
TrsLabs - Fintech Product & Business Consulting
TrsLabs - Fintech Product & Business ConsultingTrsLabs - Fintech Product & Business Consulting
TrsLabs - Fintech Product & Business Consulting
Trs Labs
 
Splunk Security Update | Public Sector Summit Germany 2025
Splunk Security Update | Public Sector Summit Germany 2025Splunk Security Update | Public Sector Summit Germany 2025
Splunk Security Update | Public Sector Summit Germany 2025
Splunk
 
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Impelsys Inc.
 
Generative Artificial Intelligence (GenAI) in Business
Generative Artificial Intelligence (GenAI) in BusinessGenerative Artificial Intelligence (GenAI) in Business
Generative Artificial Intelligence (GenAI) in Business
Dr. Tathagat Varma
 
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager APIUiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPathCommunity
 
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep DiveDesigning Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
ScyllaDB
 
Semantic Cultivators : The Critical Future Role to Enable AI
Semantic Cultivators : The Critical Future Role to Enable AISemantic Cultivators : The Critical Future Role to Enable AI
Semantic Cultivators : The Critical Future Role to Enable AI
artmondano
 
Big Data Analytics Quick Research Guide by Arthur Morgan
Big Data Analytics Quick Research Guide by Arthur MorganBig Data Analytics Quick Research Guide by Arthur Morgan
Big Data Analytics Quick Research Guide by Arthur Morgan
Arthur Morgan
 
Into The Box Conference Keynote Day 1 (ITB2025)
Into The Box Conference Keynote Day 1 (ITB2025)Into The Box Conference Keynote Day 1 (ITB2025)
Into The Box Conference Keynote Day 1 (ITB2025)
Ortus Solutions, Corp
 
What is Model Context Protocol(MCP) - The new technology for communication bw...
What is Model Context Protocol(MCP) - The new technology for communication bw...What is Model Context Protocol(MCP) - The new technology for communication bw...
What is Model Context Protocol(MCP) - The new technology for communication bw...
Vishnu Singh Chundawat
 
How Can I use the AI Hype in my Business Context?
How Can I use the AI Hype in my Business Context?How Can I use the AI Hype in my Business Context?
How Can I use the AI Hype in my Business Context?
Daniel Lehner
 
Technology Trends in 2025: AI and Big Data Analytics
Technology Trends in 2025: AI and Big Data AnalyticsTechnology Trends in 2025: AI and Big Data Analytics
Technology Trends in 2025: AI and Big Data Analytics
InData Labs
 
Procurement Insights Cost To Value Guide.pptx
Procurement Insights Cost To Value Guide.pptxProcurement Insights Cost To Value Guide.pptx
Procurement Insights Cost To Value Guide.pptx
Jon Hansen
 
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfThe Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdf
Abi john
 
Heap, Types of Heap, Insertion and Deletion
Heap, Types of Heap, Insertion and DeletionHeap, Types of Heap, Insertion and Deletion
Heap, Types of Heap, Insertion and Deletion
Jaydeep Kale
 
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxDevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptx
Justin Reock
 
Cybersecurity Identity and Access Solutions using Azure AD
Cybersecurity Identity and Access Solutions using Azure ADCybersecurity Identity and Access Solutions using Azure AD
Cybersecurity Identity and Access Solutions using Azure AD
VICTOR MAESTRE RAMIREZ
 
Complete Guide to Advanced Logistics Management Software in Riyadh.pdf
Complete Guide to Advanced Logistics Management Software in Riyadh.pdfComplete Guide to Advanced Logistics Management Software in Riyadh.pdf
Complete Guide to Advanced Logistics Management Software in Riyadh.pdf
Software Company
 
Cyber Awareness overview for 2025 month of security
Cyber Awareness overview for 2025 month of securityCyber Awareness overview for 2025 month of security
Cyber Awareness overview for 2025 month of security
riccardosl1
 
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-UmgebungenHCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungen
panagenda
 
TrsLabs - Fintech Product & Business Consulting
TrsLabs - Fintech Product & Business ConsultingTrsLabs - Fintech Product & Business Consulting
TrsLabs - Fintech Product & Business Consulting
Trs Labs
 
Splunk Security Update | Public Sector Summit Germany 2025
Splunk Security Update | Public Sector Summit Germany 2025Splunk Security Update | Public Sector Summit Germany 2025
Splunk Security Update | Public Sector Summit Germany 2025
Splunk
 
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...
Impelsys Inc.
 
Generative Artificial Intelligence (GenAI) in Business
Generative Artificial Intelligence (GenAI) in BusinessGenerative Artificial Intelligence (GenAI) in Business
Generative Artificial Intelligence (GenAI) in Business
Dr. Tathagat Varma
 
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager APIUiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager API
UiPathCommunity
 
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep DiveDesigning Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep Dive
ScyllaDB
 
Semantic Cultivators : The Critical Future Role to Enable AI
Semantic Cultivators : The Critical Future Role to Enable AISemantic Cultivators : The Critical Future Role to Enable AI
Semantic Cultivators : The Critical Future Role to Enable AI
artmondano
 

Apache Kafka and the Data Mesh | Ben Stopford and Michael Noll, Confluent

  • 1. Apache Kafka and the Data Mesh Ben Stopford, Michael G. Noll Office of the CTO, Confluent Kafka Summit Europe, May 11-12, 2021
  • 2. What is a Data Mesh? 2 1. An implementation Pattern (not unlike Microservices) 2. 4 Principles 3. Technology agnostic
  • 3. Several historical influences 3 DDD Microservices Data Marts Event Streaming Data on the Inside / Data on the Outside
  • 4. Data “in practice” Needs More Discipline 4 Data as a Practice … is not on the same level. Software as a Practice
  • 5. Spaghetti: Data architectures often lack rigour 5
  • 6. Kafka provides a solution. The implementation. 6 Kafka Centralize an immutable stream of facts. Decentralize the freedom to act, adapt, and change.
  • 7. Data Mesh with Event Streaming
  • 8. Domain-driven Decentralization Local Autonomy Per Domain (Organizational Concerns) Data as a First-class Product Product thinking, “Microservice for Data” Federated Governance Interoperability Across Domains, Network Effects (Organizational Concerns) Self-serve Data Platform Infrastructure as a Platform Across Domains 1 2 3 4 The Principles of a Data Mesh
  • 9. 10 Data Mesh A First Look Domain Inventory Shipments Orders ... Data Product
  • 10. Domain-driven Decentralization Local Autonomy (Organizational Concerns) Data as a First-class Product Product thinking, “Microservice for Data” Federated Governance Interoperability, Network Effects (Organizational Concerns) Self-serve Data Platform Infra Tooling, Across Domains 1 2 3 4 The Principles of a Data Mesh
  • 11. Principle 1: Domain-driven Decentralization Anti-pattern: responsibility for data becomes the domain of the DWH team Pattern: Ownership of a data asset given to the “local” team that is most familiar with it Centralized Data Ownership Decentralized Data Ownership Objective: Ensure data is owned by those that truly understand it
  • 12. 13 Billing Recommendations Inventory Orders Domain Shipment Domain Shipping Data Order Data 1. Joe in Inventory has a problem with Order data. 2. Inventory items are going negative, because of bad Order data. 3. He could fix the data up locally in the Inventory domain, and get on with his job. 4. Or, better, he contacts Alice in Orders and get it fixed at the source. This is more reliable as Joe doesn’t fully understand the Orders process. 5. Ergo, Alice needs be an responsible & responsive “Data Product Owner”, so everyone benefits from the fix to Joe’s problem. Joe Alice Practical example
  • 13. Recommendations: Domain-driven Decentralization 14 Learn from DDD: - Use a standard language and nomenclature for data. - Business users should understand a data flow diagram. - The stream of events should create a shared narrative that is business-user comprehensible.
  • 14. Domain-driven Decentralization Local Autonomy (Organizational Concerns) Data as a First-class Product Product thinking, “Microservice for Data” Federated Governance Interoperability, Network Effects (Organizational Concerns) Self-serve Data Platform Infra Tooling, Across Domains 1 2 3 4 The Principles of a Data Mesh
  • 15. Principle 2: Data as a First-Class Product 16 • Objective: Make shared data discoverable, addressable, trustworthy, secure, so other teams can make good use of it. • Data is treated as a true product, not a by-product. This product thinking is important to prevent data chauvinism.
  • 16. Infra Code Data product, a “microservice for the data world” 17 • Data product is a node on the data mesh, situated within a domain. • Produces—and possibly consumes—high-quality data within the mesh. • Encapsulates all the elements required for its function, namely data + code + infrastructure. Data Creates, manipulates, serves, etc. that data Powers the data (e.g., storage) and the code (e.g., run, deploy, monitor) “Items about to expire” Data Product Data and metadata, including history
  • 18. Domain 19 Data Product ...naturally to Event Streaming with Kafka Data Mesh
  • 19. 20 Event Streaming is Pub/Sub, not Point-to-Point Data Product Data Product Data Product Data Product stream (persisted) other streams write (publish) read (consume) independently Data producers are scalably decoupled from consumers.
  • 20. Data Product Data Product Why is Event Streaming a good fit for meshing? 21 0 1 2 3 4 5 6 1 7 Streams are real-time, low latency ⇒ Propagate data immediately. Streams are highly scalable ⇒ Handle today’s massive data volumes. Streams are stored, replayable ⇒ Capture real-time & historical data. Streams are immutable ⇒ Auditable source of
  • 21. How to get data into & out of a data product 22 Data Product Input Data Ports Output Data Ports Snapshot via Nightly ETL Snapshot via Nighty ETL Continuous Stream Snapshot via Req/Res API Snapshot via Req/Res API A B C Continuous Stream
  • 22. Onboarding existing data 23 Data Product Input Data Ports Source Connectors Use Kafka connectors to stream data from cloud services and existing systems into the mesh. https://www.confluent.io/hub/
  • 23. Data product: what’s happening inside 24 Input Data Ports Output Data Ports …pick your favorites... Data on the Inside: HOW the domain team solves specific problems internally? This doesn’t matter to other domains.
  • 24. Event Streaming inside a data product 25 Input Data Ports Output Data Ports ksqlDB to filter, process, join, aggregate, analyze Stream data from other DPs or internal systems into ksqlDB 1 2 Stream data to internal systems or the outside. Pull queries can drive a req/res API. 3 Req/Res API Pull Queries Use ksqlDB, Kafka Streams apps, etc. for processing data in motion.
  • 25. Event Streaming inside a data product 26 Input Data Ports Output Data Ports MySQL Sink Connector Source Connector DB client apps work as usual Stream data from other Data Products into your local DB Stream data to the outside with CDC and e.g. the Outbox Pattern, ksqlDB, etc. 1 3 2 Use Kafka connectors and CDC to “streamify” classic databases.
  • 26. Dealing with data change: schemas & versioning 27 Data Product Output Data Ports V1 - user, product, quantity V2 - userAnonymized, product, quantity Publish evolving streams with back/forward-compatible schemas. Publish versioned streams for breaking changes. Also, when needed, data can be fully reprocessed by replaying history.
  • 27. Recommendations: Data as a First-class Product 28 1. Data-on-the-Outside is harder to change, but it has more value in a holistic sense. a. Use schemas as a contract. b. Handle incompatible schema changes using Dual Schema Upgrade Window pattern. 2. Get data from the source, not from intermediaries. Think: Demeter's law applied to data. a. Otherwise, proliferation of ‘slightly corrupt’ data within the mesh. “Game of Telephone”. b. Event Streaming makes it easy to subscribe to data from authoritative sources. 3. Change data at the source, including error fixes. Don’t “fix data up” locally. 4. Some data sources will be difficult to turn into first-class data products. Example: Batch-based sources that lose event-level data or are not reproducible. a. Use Event Streaming plus CDC, Outbox Pattern, etc. to integrate these into the mesh.
  • 28. Domain-driven Decentralization Local Autonomy (Organizational Concerns) Data as a First-class Product Product thinking, “Microservice for Data” Federated Governance Interoperability, Network Effects (Organizational Concerns) Self-serve Data Platform Infra Tooling, Across Domains 1 2 3 4 The Principles of a Data Mesh
  • 29. Why Self-service Matters 30 Trade Surveillance System ● Data from 13 sources ● Some sources publish events ● Needed both historical and real-time data ● Historical data from database extracts arranged with dev team. ● Format of events different to format of extracts ● 9 months of effort to get 13 sources into the new system.
  • 30. Principle 3: Self-serve Data Platform 31 Central infrastructure that provides real-time and historical data on demand Objective: Make domains autonomous in their execution through rapid data provisioning
  • 31. Consuming real-time & historical data from the mesh 32 1) Separate Systems for Real-time and Historical Data (Lambda Architecture) - Considerations: - Difficulty to correlate real-time with historical “snapshot” data - Two systems to manage - Unlike event streams, snapshots have less granularity 1) One System for Real-time and Historical Data (Kappa Architecture) - Considerations: - Operational complexity (addressed in Confluent Cloud) - Downsides of immutability of regular streams: e.g. altering or deleting events - Storage cost (addressed in Confluent Cloud, in Apache Kafka with KIP-405)
  • 32. What this can look like in practice 33 Browse Schemas
  • 34. 35 With ksqlDB the Data Mesh is Queryable and Decentralized Data Product STREAM PROCESSOR ksqlDB Query is the interface to the mesh Events are the interface to the mesh
  • 35. Think: Infrastructure as code, but for data 36 Code Container Image + Same APP every time Code Event Streams + Same DATA every time
  • 36. 37 Mesh is one logical cluster. Data product has another. Data Product Data Product has its own cluster for internal use
  • 37. Domain-driven Decentralization Local Autonomy (Organizational Concerns) Data as a First-class Product Product thinking, “Microservice for Data” Federated Governance Interoperability, Network Effects (Organizational Concerns) Self-serve Data Platform Infra Tooling, Across Domains 1 2 3 4 The Principles of a Data Mesh
  • 38. Principle 4: Federated Governance 39 • Objective: Independent data products can interoperate and create network effects. • Establish global standards, like governance, that apply to all data products in the mesh. • Ideally, these global standards and rules are applied automatically by the platform. Domain Domain Domain Domain Self-serve Data Platform What is decided locally by a domain? What is globally? (implemented and enforced by platform) Must balance between Decentralization vs. Centralization. No silver bullet!
  • 39. Example standard: Identifying customers globally • Define how data is represented, so you can join and correlate data across different domains. • Use data contracts, schemas, registries, etc. to implement and enforce such standards. • Use Event Streaming to retrofit historical data to new requirements, standards. 40 customerId=29639 customerId=29639 customerId=29639 customerId=29639 SELECT … FROM orders o LEFT JOIN shipments s ON o.customerId = s.customerId EMIT CHANGES;
  • 40. Example standard: Detect errors and recover with Streams 41 • Use strategies like logging, data profiling, data lineage, etc. to detect errors in the mesh. • Streams are very helpful to detect errors and identify cause-effect relationships. • Streams let you recover and fix errors: e.g., replay & reprocess historical data. Data Product Output Data Ports 0 1 2 3 4 5 6 7 8 9 My App Bug? Error? Rewind to start of stream, then reprocess. If needed, tell the origin data product to fix problematic data at the source. Event Streams give you a powerful Time Machine.
  • 41. Example standard: Tracking data lineage with Streams 42 • Lineage must work across domains and data products—and systems, clouds, data centers. • Event streaming is a foundational technology for this. On-premise
  • 42. Recommendations: Federated Governance 43 1. Be pragmatic: Don’t expect governance systems to be perfect. a. They are a map that helps you navigate the data-landscape of your company. b. But there will always be roads that have changed or have not been mapped. 2. Governance is more a process—i.e., an organizational concern—than a technology. 1. Beware of centralized data models, which can become slow to change. Where they must exist, use processes & tooling like GitHub to collaborate and change quickly. Good luck! 🙂
  • 43. Data mesh journey 44 Principle 1 Data should have one owner: the team that creates it. Principle 2 Data is your product: All exposed data should be good data. Principle 3 Get access to any data immediately and painlessly, be it historical or real-time. Principle 4: Governance, with standards, security, lineage, etc. (cross-cutting concerns) Difficulty to execute Start Here 1 2 3

Editor's Notes

  • #6: Small scale data pipelines constantly broken. Large scale: finance and risk have completely different numbers. Story of one path for books in an investment bank. Boose Allen Hamilton. 3 months analysis. 2 hours to explain.
  • #7: This allows the applications to connect around data in motion Acts as a kind of central nervous system Let’s something happening in one part of the company, trigger the right updates and response everywhere else as it occurs
  • #9: Thanks Ben for setting the stage and covering the influences of the data mesh concept. Let us now dive into building a data mesh with event streaming and Kafka.
  • #10: Ben and I will discuss the four data mesh principles: (1) giving local autonomy to teams via domain-driven decentralization, (2) establishing product thinking to treat data as a first-class product, (3) providing a self-serve data platform so the local domains can easily participate in the data mesh, and (4) federated governance to tie these independent domains and data assets together so they actually “mesh” and interoperate.
  • #11: The data mesh concept has both organizational and technical dimensions. But, essentially, we can think of a data mesh as a network to exchange data about the business, with nodes and connections. The nodes in the mesh are the “data products”—we’ll learn more about them later, but for now let’s just imagine a node could be a microservice, a database, an application, etc. The nodes function is to produce and consume high-quality, locally-curated data within the mesh for the benefits of other nodes. “Related” nodes, so to speak, are grouped into domains, like “Inventory”. And the data flows from node-to-node and thus from domain-to-domain as needed.
  • #13: A centralized approach, essentially, does not work at non-trivial scale (might work for a small startup, but not an enterprise).
  • #14: OUTRO: This ensures that data does not diverge in different ways across the organization; e.g., the risk department has different closing position to the finance department.
  • #16: Let’s talk about the second principle: treating data as a first-class product. This is arguably the most important principle of the four.
  • #17: This principle is about product thinking: Your domain’s shared data is managed as a true product. And your objective is to provide this data product in the right way so that other teams can make good use of it. In other words, you treat the other teams as internal customers of your data. This product thinking is really important to prevent data chauvinism, which is a risk of principle 1 that Ben just covered. That’s because decentralization and granting local autonomy to domains may cause the creation of even more data silos along with data quality problems if the origin domains only care about themselves. So, what is “a data product” then? Sounds a bit abstract, doesn’t it?
  • #18: Think of a data product as a microservice for the data world. It is a node on the data mesh, situated within a particular domain. The role of the data product is to produce—and possibly consume—high-quality data within the mesh. And, like a microservice, it encapsulates everything it needs for its functioning: (1) the actual data that we want to share with the outside world, (2) the code that creates, manipulates, and shares that data, (3) the infrastructure that we need to store the data and to run the code. So, as you can see, it’s really similar to microservices. [OUTRO] Now, before we take a closer look at a data product, let’s first zoom out again to the bigger picture of the data mesh.
  • #19: Because a key question is: How is data shared across data products in the mesh? That is, how do we implement the links and connections between the nodes in the network? And, even visually, we can see that this interconnectivity lends itself very naturally to … [NEXT SLIDE]
  • #20: ...Event Streaming with Kafka. Here, data is provided to other data products through streams in Kafka. And any data product can consume via Kafka from the high-quality data streams of other data products. As we can see, this idea of a data mesh is very similar to the idea of a Central Nervous System, where data is continuously flowing, being processed, analyzed, acted upon. Now, we must remember that the data mesh shown here is a LOGICAL view, not a physical one. [OUTRO] If you know Kafka, you know that the reality looks a bit different and...a bit better.
  • #21: Event streaming is not a point-to-point architecture that leads to this spaghetti mess we have showed in the beginning of this talk. Instead, it is an architecture that decouples data producers from data consumers in a way that is highly scalable, fault-tolerant, and persistent (because streams are durably stored). Here, a data product talks to other data products only INDIRECTLY, and this layer of indirection is event streams in Kafka. So this setup is Write-Once, Read-Many. And all a data product has to do to participate in a data mesh is to publish its “public” data as streams, and then other data products can independently consume this data in the way they see fit. [OUTRO] Now, as a quick refresher, let’s recap the key properties of event streams and why they are such a great fit to build a data mesh?
  • #22: Here are some of the reasons. For example, streams are real-time, so you can propagate data throughout the mesh immediately, as soon as new information is available. Streams are also persisted and replayable, so they let you capture both real-time AND historical data with one setup. And because they are immutable, they make for a great source of record, which is useful for governance purposes.
  • #23: So, let’s get back to data products and the question of how do we get data into and out of a data product? There are three main options. A) Use streams. We already discussed why this is a good fit in a data mesh. A strong advantage is that they continuously import and export data as soon as new information is available in the business. B) Use Request-Response APIs to request a data snapshot on demand, like a file export. C) Use traditional ETL for batch import/export of nightly snapshots. This gives you a fixed cadence of importing and exporting these snapshots, but, compared to option B, less control over exactly what snapshot data you’ll be working with.
  • #24: With event streaming, you can quickly onboard data from existing systems through Kafka connectors that stream data from these systems into the mesh in real-time. There are two hundred of such connectors readily available, so in most cases you don’t have to write anything yourself. So this is a great way to set data in motion when you are bootstrapping a data mesh.
  • #25: Now that there is some data flowing into our data product, what’s happening then? Well, really, what’s happening inside a data product is of no concern to the data mesh. Apart from the team responsible for a data product, everybody else in the data mesh only cares about the public data that is shared by this domain. HOW this public data is created internally doesn’t matter, as long as it meets criteria such as being addressable, trustworthy, have good quality, and so on.
  • #26: That being said, the local team does care about how the data is being made. And, among other interesting technologies, event streaming is also very useful to implement the insides of a data product. Naturally, you’d use streaming tools like ksqlDB and Kafka Streams to process any streaming data that enters or is created within a data product, and streaming data that leaves and thus is shared with the outside world in the data mesh.
  • #27: Likewise, if you use non-streaming systems like databases and cloud storage services within a data product, you can “streamify” them via Kafka connectors. For example, as shown here, you can feed real-time updates into relational database like MySQL from upstream data products to keep the database current with the latest business information, and any existing applications that interact with the database can continue to work as usual. And similarly, any changes to the database itself—like updates to a customer profile—can be captured and shared as event streams with the outside world in the data mesh.
  • #28: This brings us to the last point: In which way should we share our data product’s streams within the mesh? The two main considerations here are the use of schemas and the use of versioning. When you publish a stream, you should use schemas with appropriate backwards and/or forwards compatibility. This ensures, for example, that when you add a new data field to a stream, that any existing consumers of that stream can continue to operate without having to make any changes. However, if you do need to introduce breaking changes to the schema of a stream, then use versioned streams. For example, if you are being required to obfuscate certain data fields moving forward, which likely DOES break existing consumers, then publish a new version of your stream, and deprecate the original version.
  • #35: This is the architecture of most event streaming applications that have any UI component (to show sth to a human). Maybe you streaming-import data from other databases into Kafka, then perform stream processing on top of that, and then use Kafka Connect to streaming-export the results into a traditional database, against which a UI application runs lookup queries or other tasks. This architecture is good, and many great applications have been built with it. But: it does suffer from unnecessary complexity.
  • #36: ksqlDB turns the data mesh into something you can query, while still having all the benefits of being decentralized
  • #39: Let’s talk about the last principle: Federated governance.
  • #40: Like principle 1 (Decentralization), this principle is largely an organizational concern. The objective here is to ensure that all these independent, autonomous teams and data products within the mesh can work actually together—that is, we want to ensure that the various constituents in the mesh actually mesh so that we generate network effects for all this data we share across the organization. Predominantly, this task is about establishing global standards like data governance, and applying these across the full mesh. So, even though we started with decentralization in principles 1 and 2, we now want to strike the right balance between decentralization and centralization to tie everything together. This isn’t easy, and in practice the execution will likely differ from company to company. Let’s take a look at concrete examples.
  • #41: One important task is to ensure that “the same” data or data attribute is represented identically across different domains. Otherwise we can’t join or correlate, for example, data about our customers across the various domains like Inventory, Orders, and Shipments. We won’t have the time in this talk to cover the how to define and agree upon such data models. But what we can say is that—once you have them—then data contracts, schemas, and schema registries will help you to implement and enforce such standards throughout the data mesh.
  • #42: Another example is how data products should detect and recover from errors. Here, we can use common strategies like application logs, data profiling, and data lineage to unearth errors in the mesh. Streams are very useful here because they capture live AND historical data in a sequence of events, which let you identify cause-and-effect relationships much more easily—particularly if you can join and correlate streams from different domains as discussed in the slide before. Also, streams let you then recover and fix any such errors by replaying and reprocessing historical data.
  • #43: In this final example, we can see again that there are lots of data streams within a data mesh. These data streams may span across systems, data centers, clouds, and so on. For the purpose of tracking data lineage, we ideally want to cover the full mesh, so we must follow the data. Event streaming is again a key technology to implement this in practice, because it lets you track data-in-motion all the way from its origins to intermediate and to the final destinations.
  • #44: Before we wrap up, let’s talk about some further considerations for federated governance. In my opinion, the most important recommendation is having the right mindset: be pragmatic! Don’t expect a governance system to be perfect. It only needs to be sufficiently good. There will always be parts of a data architecture that is not covered or that was just recently changed. So governance is really more an process than a technology. Lastly, we talked quite a bit about decentralization vs. centralization. Governance is on the centralized side. So while you probably want to standardize a number of things, please be aware of the downsides of centralized DATA MODELS, as they will eventually become too slow to change. If you really need something like this, use processes and tools like GitHub to collaborate efficiently, and have good luck! :-)
  • #45: Michael: As Ben just said, the data mesh is a concept, and you should really think of [building a data mesh in your company] as a journey. You didn’t have one last week, and you will not have a finished data mesh next week. So get the necessary management commitment, and then ideally start small with the first concrete use cases, and iterate from there. And finally, while the data mesh is certainly a valuable concept, it’s not the holy grail. The data mesh sits next to other useful concepts, like microservices and DDD. And with its focus on analytical data, the data mesh is but one of the ways you can apply event streaming & Kafka to drive your business, just like you can use event streaming with microservices to implement the more operational systems. Apply these concepts in the way you see fit so that you can make technology useful in your company. [OUTRO] Ben and I hope you enjoyed our session on data mesh and event streaming, and we wish you a lot of fun with the rest of Kafka Summit.
  • #48: [OUTRO] So, what’s a Data Product then?