SlideShare a Scribd company logo
SPARK SHUFFLE
INTRODUCTION
天火@蘑菇街
About Me
Spark / Hadoop / Hbase / Phoenix
contributor
For spark mainly contributes in:
•  Yarn
•  Shuffle
•  BlockManager
•  Scala2.10 update
•  Standalone HA
•  Various other fixes.
tianhuo@mogujie.com
Weibo @冷冻蚂蚁
blog.csdn.net/colorant
Why Spark is fast(er)
•  Whom do we compare to?
•  What do we mean by fast?
•  fast to write
•  fast to run
Why Spark is fast(er) cont.
•  But the figure in previous page is some how misleading.
•  The key is the flexible programming mode.
•  Which lead to more reasonable data flow.
•  Which lead to less IO operation.
•  Especially for iterative heavy workloads like ML.
•  Which potentially cut off a lot of shuffle operations needed.
•  But, you won’t always be lucky.
•  Many app logic did need to exchange a lot of data.
•  In the end, you will still need to deal with shuffle
•  And which usually impact performance a lot.
Spark shuffle introduction
What is shuffle
7
Shuffle overview
Aggregator
Aggregator
AggregatorAggregator
AggregatorAggregator
How does shuffle come into the picture
•  Spark run job stage by stage.
•  Stages are build up by DAGScheduler according to RDD’s
ShuffleDependency
•  e.g. ShuffleRDD / CoGroupedRDD will have a ShuffleDependency
•  Many operator will create ShuffleRDD / CoGroupedRDD under
the hook.
•  Repartition/CombineByKey/GroupBy/ReduceByKey/cogroup
•  many other operator will further call into the above operators
•  e.g. various join operator will call cogroup.
•  Each ShuffleDependency maps to one stage in Spark Job
and then will lead to a shuffle.
So everyone should have seen this before
join	
  
union	
  
groupBy	
  
map	
  
Stage	
  3	
  
Stage	
  1	
  
Stage	
  2	
  
A:	
   B:	
  
C:	
   D:	
  
E:	
  
F:	
  
G:	
  
why shuffle is expensive
•  When doing shuffle, data no longer stay in memory only
•  For spark, shuffle process might involve
•  data partition: which might involve very expensive data sorting
works etc.
•  data ser/deser: to enable data been transfer through network or
across processes.
•  data compression: to reduce IO bandwidth etc.
•  DISK IO: probably multiple times on one single data block
•  E.g. Shuffle Spill, Merge combine
History
•  Spark 0.6-0.7, same code path with RDD’s persistent
method, can choose MEMORY_ONLY and DISK_ONLY
(default).
•  Spark 0.8-0.9:
•  separate shuffle code path from BM and create
ShuffleBlockManager and BlockObjectWriter only for shuffle, now
shuffle data can only be written to disk.
•  Shuffle optimization: Consolidate shuffle write.
•  Spark 1.0, pluggable shuffle framework.
•  Spark 1.1, sort-based shuffle implementation.
•  Spark 1.2 netty transfer service reimplementation. sort-
based shuffle by default
•  Spark 1.2+ on the go: external shuffle service etc.
LOOK
INSIDE
13
Pluggable Shuffle Framework
•  ShuffleManager
•  Manage shuffle related components, registered in SparkEnv,
configured through SparkConf, default is sort (pre 1.2 is hash),
•  ShuffleWriter
•  Handle shuffle data output logics. Will return MapStatus to be
tracked by MapOutputTracker.
•  ShuffleReader
•  Fetch shuffle data to be used by e.g. ShuffleRDD
•  ShuffleBlockManager
•  Manage the mapping relation between abstract bucket and
materialized data block.
High level data flow
BlockManager
HashShuffleManager
DiskBlockManager
FileShuffleBlockManager
Local File System
SortShuffleManager
IndexShuffleBlockManager
GetBlockData
BlockTransferService
GetBlockData
Direct mapping or
mapping by File Groups
Map to One Data File and
One Index File per mapId
Just do one-one File mapping
15
Hash Based Shuffle - Shuffle Writer
•  Basic shuffle writer
Map	
  Task Map	
  Task Map	
  Task Map	
  Task
File
File
File
File
File
File
File
File
File
File
File
File
File
File
File
File
Aggregator AggregatorAggregatorAggregator
Each bucket is mapping to a single file
16
Hash Based Shuffle - Shuffle Writer
•  Consolidate Shuffle Writer
Each bucket is mapping to a segment of file
Aggregator
Aggregator
Aggregator
Aggregator
17
Hash Based Shuffle - Shuffle Writer
•  Basic Shuffle Writer
•  M * R shuffle spill files
•  Concurrent C * R opened shuffle files.
•  If shuffle spill enabled, could generate more tmp spill files say N.
•  Consolidate Shuffle Writer
•  Reduce the total spilled files into C * R if (M >> C)
•  Concurrent opened is the same as the basic shuffle writer.
•  Memory consumption
•  Thus Concurrent C * R + N file handlers.
•  Each file handler could take up to 32~100KB+ Memory for various
buffers across the writer stream chain.
18
Sort Based Shuffle - Shuffle Writer
•  Sort Shuffle Writer
Map	
  Task Map	
  Task Map	
  Task Map	
  Task
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileSegment
FileFile
FileFile FileFile
FileFile
ExternalSorter ExternalSorterExternalSorterExternalSorter
19
Sort Based Shuffle - Shuffle Writer
•  Each map task generates 1 shuffle data file + 1 index file
•  Utilize ExternalSorter to do the sort works.
•  If map-side combine is required, data will be sorted by
key and partition for aggregation. Otherwise data will
only be sorted by partition.
•  If reducer number <= 200 and no need to do aggregation
or ordering, data will not be sorted at all.
•  Will go with hash way and spill to separate files for each reduce
partition, then merge them into one per map for final output.
20
Hash Based Shuffle - Shuffle Reader
•  Actually, at present, Sort Based Shuffle also go with
HashShuffleReader
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Bucket
Reduce	
  TaskReduce	
  Task Reduce	
  Task Reduce	
  Task Reduce	
  Task
Aggregator AggregatorAggregatorAggregator
BLOCK
TRANSFER
SERVICE
Related conceptions
•  BlockTransferService
•  Provide a general interface for ShuffleFetcher and working with
BlockDataManager to get local data.
•  ShuffleClient
•  Wrap up the fetching data process for the client side, say setup
TransportContext, new TransportClient etc.
•  TransportContext
•  Context to setup the transport layer
•  TransportServer
•  low-level streaming service server
•  TransportClient
•  Client for fetching consecutive chunks TransportServer
ShuffleManager
Data Flow
23
BlockManager
NioBlockTransferService
GetBlockData
BlockDataManager
ConnectionManager
NioBlockTransferService
ConnectionManager
GetBlock
GotBlock
BlockStoreShuffleFetcher
ShuffleBlockFetcherIterator
Block
Manager
Local Blocks
Remote Blocks
Local Remote
HashShuffleReader
fetch
ShuffleManager
ShuffleBlockManager
GetBlockData
Can Switch to different BlockTransferService
ShuffleManager
Data Flow
24
BlockManager
NettyBlockTransferService
GetBlockData
BlockDataManager
TransportClient
NettyBlockTransferService
TransportServer
BlockStoreShuffleFetcher
ShuffleBlockFetcherIterator
Block
Manager
Local Blocks
Remote Blocks
Local Remote
HashShuffleReader
fetch
ShuffleManager
ShuffleBlockManager
GetBlockData
clientHandler
TransportChannel
HandlerclientHandler
TransportChannel
Handler
Fetch
Request
Fetch
Results
Spark shuffle introduction
External Shuffle Service
•  Design goal
•  allow for the service to be long-running
•  possibly much longer-running than Spark
•  support multiple version of Spark simultaneously etc.
•  can be integrated into YARN NodeManager, Standalone Worker, or
on its own
•  The entire service been ported to Java
•  do not include Spark's dependencies
•  full control over the binary compatibility of the components
•  not depend on the Scala runtime or version.
External Shuffle Service
•  Current Status
•  Basic framework seems ready.
•  A Network module extracted from the core module
•  BlockManager could be configured with executor built-in shuffle
service or external standalone shuffle service
•  A standaloneWorkerShuffleService could be launched by worker
•  Disabled by default.
•  How it works
•  Shuffle data is still written by the shuffleWriter to local disks.
•  The external shuffle service knows how to read these files on disks
(executor will registered related info to it, e.g. shuffle manager type,
file dir layout etc.), it follow the same rules applied for written these
file, so it could serve the data correctly.
28
Sort Merge Shuffle Reader
•  Background:
•  Current HashShuffleReader does not utilize the sort result within
partition in map-side.
•  The actual by key sort work is always done at reduce side.
•  While the map side will do by-partition sort anyway ( sort shuffle )
•  Change it to a by-key-and-partition sort does not bring many extra
overhead.
•  Current Status
•  [WIP] https://github.com/apache/spark/pull/3438
Some shuffle related configs
•  spark.shuffle.spill (true)
•  spark.shuffle.memoryFraction (0.2)
•  spark.shuffle.manager [sort]/hash
•  spark.shuffle.sort.bypassMergeThreshold (200)
•  spark.shuffle.blockTransferService [netty]/nio
•  spark.shuffle.consolidateFiles (false)
•  spark.shuffle.service.enabled (false)
What’s next?
•  Other custom shuffle logic?
•  Alternative way to save shuffle data blocks
•  E.g. in memory (again)
•  Other transport mechanism?
•  Break stage barrier?
•  To fetch shuffle data when part of the map tasks are done.
•  Push mode instead of pull mode?
Thanks to Jerry Shao
•  Some of this ppt’s material came from Jerry Shao@Intel
weibo: @saisai_shao
•  Jerry also contributes a lot of essential patches for spark
core / spark streaming etc.
Join Us !
• 加盟 / 合作 / 讨论 统统欢迎
• 数据平台开发,大数据相关技术,只要够
Cool,我们都玩
• tianhuo@mogujie.com
Spark shuffle introduction
Ad

More Related Content

What's hot (20)

Apache Spark Architecture
Apache Spark ArchitectureApache Spark Architecture
Apache Spark Architecture
Alexey Grishchenko
 
Physical Plans in Spark SQL
Physical Plans in Spark SQLPhysical Plans in Spark SQL
Physical Plans in Spark SQL
Databricks
 
Top 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark ApplicationsTop 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark Applications
Spark Summit
 
How to understand and analyze Apache Hive query execution plan for performanc...
How to understand and analyze Apache Hive query execution plan for performanc...How to understand and analyze Apache Hive query execution plan for performanc...
How to understand and analyze Apache Hive query execution plan for performanc...
DataWorks Summit/Hadoop Summit
 
Apache Spark Core – Practical Optimization
Apache Spark Core – Practical OptimizationApache Spark Core – Practical Optimization
Apache Spark Core – Practical Optimization
Databricks
 
Introduction to Apache Spark
Introduction to Apache SparkIntroduction to Apache Spark
Introduction to Apache Spark
Rahul Jain
 
Apache Spark in Depth: Core Concepts, Architecture & Internals
Apache Spark in Depth: Core Concepts, Architecture & InternalsApache Spark in Depth: Core Concepts, Architecture & Internals
Apache Spark in Depth: Core Concepts, Architecture & Internals
Anton Kirillov
 
Dive into PySpark
Dive into PySparkDive into PySpark
Dive into PySpark
Mateusz Buśkiewicz
 
Apache Spark At Scale in the Cloud
Apache Spark At Scale in the CloudApache Spark At Scale in the Cloud
Apache Spark At Scale in the Cloud
Databricks
 
Beyond SQL: Speeding up Spark with DataFrames
Beyond SQL: Speeding up Spark with DataFramesBeyond SQL: Speeding up Spark with DataFrames
Beyond SQL: Speeding up Spark with DataFrames
Databricks
 
Hive + Tez: A Performance Deep Dive
Hive + Tez: A Performance Deep DiveHive + Tez: A Performance Deep Dive
Hive + Tez: A Performance Deep Dive
DataWorks Summit
 
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the CloudAmazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Noritaka Sekiyama
 
Top 5 Mistakes to Avoid When Writing Apache Spark Applications
Top 5 Mistakes to Avoid When Writing Apache Spark ApplicationsTop 5 Mistakes to Avoid When Writing Apache Spark Applications
Top 5 Mistakes to Avoid When Writing Apache Spark Applications
Cloudera, Inc.
 
PySpark dataframe
PySpark dataframePySpark dataframe
PySpark dataframe
Jaemun Jung
 
Introduction to Spark Internals
Introduction to Spark InternalsIntroduction to Spark Internals
Introduction to Spark Internals
Pietro Michiardi
 
Processing Large Data with Apache Spark -- HasGeek
Processing Large Data with Apache Spark -- HasGeekProcessing Large Data with Apache Spark -- HasGeek
Processing Large Data with Apache Spark -- HasGeek
Venkata Naga Ravi
 
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Databricks
 
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedInMagnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Databricks
 
Hive Bucketing in Apache Spark with Tejas Patil
Hive Bucketing in Apache Spark with Tejas PatilHive Bucketing in Apache Spark with Tejas Patil
Hive Bucketing in Apache Spark with Tejas Patil
Databricks
 
Introduction to Storm
Introduction to Storm Introduction to Storm
Introduction to Storm
Chandler Huang
 
Physical Plans in Spark SQL
Physical Plans in Spark SQLPhysical Plans in Spark SQL
Physical Plans in Spark SQL
Databricks
 
Top 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark ApplicationsTop 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark Applications
Spark Summit
 
How to understand and analyze Apache Hive query execution plan for performanc...
How to understand and analyze Apache Hive query execution plan for performanc...How to understand and analyze Apache Hive query execution plan for performanc...
How to understand and analyze Apache Hive query execution plan for performanc...
DataWorks Summit/Hadoop Summit
 
Apache Spark Core – Practical Optimization
Apache Spark Core – Practical OptimizationApache Spark Core – Practical Optimization
Apache Spark Core – Practical Optimization
Databricks
 
Introduction to Apache Spark
Introduction to Apache SparkIntroduction to Apache Spark
Introduction to Apache Spark
Rahul Jain
 
Apache Spark in Depth: Core Concepts, Architecture & Internals
Apache Spark in Depth: Core Concepts, Architecture & InternalsApache Spark in Depth: Core Concepts, Architecture & Internals
Apache Spark in Depth: Core Concepts, Architecture & Internals
Anton Kirillov
 
Apache Spark At Scale in the Cloud
Apache Spark At Scale in the CloudApache Spark At Scale in the Cloud
Apache Spark At Scale in the Cloud
Databricks
 
Beyond SQL: Speeding up Spark with DataFrames
Beyond SQL: Speeding up Spark with DataFramesBeyond SQL: Speeding up Spark with DataFrames
Beyond SQL: Speeding up Spark with DataFrames
Databricks
 
Hive + Tez: A Performance Deep Dive
Hive + Tez: A Performance Deep DiveHive + Tez: A Performance Deep Dive
Hive + Tez: A Performance Deep Dive
DataWorks Summit
 
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the CloudAmazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Amazon S3 Best Practice and Tuning for Hadoop/Spark in the Cloud
Noritaka Sekiyama
 
Top 5 Mistakes to Avoid When Writing Apache Spark Applications
Top 5 Mistakes to Avoid When Writing Apache Spark ApplicationsTop 5 Mistakes to Avoid When Writing Apache Spark Applications
Top 5 Mistakes to Avoid When Writing Apache Spark Applications
Cloudera, Inc.
 
PySpark dataframe
PySpark dataframePySpark dataframe
PySpark dataframe
Jaemun Jung
 
Introduction to Spark Internals
Introduction to Spark InternalsIntroduction to Spark Internals
Introduction to Spark Internals
Pietro Michiardi
 
Processing Large Data with Apache Spark -- HasGeek
Processing Large Data with Apache Spark -- HasGeekProcessing Large Data with Apache Spark -- HasGeek
Processing Large Data with Apache Spark -- HasGeek
Venkata Naga Ravi
 
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Everyday I'm Shuffling - Tips for Writing Better Spark Programs, Strata San J...
Databricks
 
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedInMagnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Databricks
 
Hive Bucketing in Apache Spark with Tejas Patil
Hive Bucketing in Apache Spark with Tejas PatilHive Bucketing in Apache Spark with Tejas Patil
Hive Bucketing in Apache Spark with Tejas Patil
Databricks
 
Introduction to Storm
Introduction to Storm Introduction to Storm
Introduction to Storm
Chandler Huang
 

Similar to Spark shuffle introduction (20)

Apache Spark
Apache SparkApache Spark
Apache Spark
SugumarSarDurai
 
Spark Overview and Performance Issues
Spark Overview and Performance IssuesSpark Overview and Performance Issues
Spark Overview and Performance Issues
Antonios Katsarakis
 
Large scale computing with mapreduce
Large scale computing with mapreduceLarge scale computing with mapreduce
Large scale computing with mapreduce
hansen3032
 
Dongwon Kim – A Comparative Performance Evaluation of Flink
Dongwon Kim – A Comparative Performance Evaluation of FlinkDongwon Kim – A Comparative Performance Evaluation of Flink
Dongwon Kim – A Comparative Performance Evaluation of Flink
Flink Forward
 
A Comparative Performance Evaluation of Apache Flink
A Comparative Performance Evaluation of Apache FlinkA Comparative Performance Evaluation of Apache Flink
A Comparative Performance Evaluation of Apache Flink
Dongwon Kim
 
Spark architechure.pptx
Spark architechure.pptxSpark architechure.pptx
Spark architechure.pptx
SaiSriMadhuriYatam
 
Improving Apache Spark by Taking Advantage of Disaggregated Architecture
 Improving Apache Spark by Taking Advantage of Disaggregated Architecture Improving Apache Spark by Taking Advantage of Disaggregated Architecture
Improving Apache Spark by Taking Advantage of Disaggregated Architecture
Databricks
 
Apache hadoop, hdfs and map reduce Overview
Apache hadoop, hdfs and map reduce OverviewApache hadoop, hdfs and map reduce Overview
Apache hadoop, hdfs and map reduce Overview
Nisanth Simon
 
Apache Spark Fundamentals
Apache Spark FundamentalsApache Spark Fundamentals
Apache Spark Fundamentals
Zahra Eskandari
 
Hadoop ppt on the basics and architecture
Hadoop ppt on the basics and architectureHadoop ppt on the basics and architecture
Hadoop ppt on the basics and architecture
saipriyacoool
 
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptxCLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
bhuvankumar3877
 
11. From Hadoop to Spark 1:2
11. From Hadoop to Spark 1:211. From Hadoop to Spark 1:2
11. From Hadoop to Spark 1:2
Fabio Fumarola
 
Giraph+Gora in ApacheCon14
Giraph+Gora in ApacheCon14Giraph+Gora in ApacheCon14
Giraph+Gora in ApacheCon14
Renato Javier Marroquín Mogrovejo
 
Apache Spark Tutorial
Apache Spark TutorialApache Spark Tutorial
Apache Spark Tutorial
Ahmet Bulut
 
Optimization of modern web applications
Optimization of modern web applicationsOptimization of modern web applications
Optimization of modern web applications
Eugene Lazutkin
 
Apache Tez -- A modern processing engine
Apache Tez -- A modern processing engineApache Tez -- A modern processing engine
Apache Tez -- A modern processing engine
bigdatagurus_meetup
 
Tez big datacamp-la-bikas_saha
Tez big datacamp-la-bikas_sahaTez big datacamp-la-bikas_saha
Tez big datacamp-la-bikas_saha
Data Con LA
 
What no one tells you about writing a streaming app
What no one tells you about writing a streaming appWhat no one tells you about writing a streaming app
What no one tells you about writing a streaming app
hadooparchbook
 
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
Spark Summit
 
Spark_Intro_Syed_Academy
Spark_Intro_Syed_AcademySpark_Intro_Syed_Academy
Spark_Intro_Syed_Academy
Syed Hadoop
 
Spark Overview and Performance Issues
Spark Overview and Performance IssuesSpark Overview and Performance Issues
Spark Overview and Performance Issues
Antonios Katsarakis
 
Large scale computing with mapreduce
Large scale computing with mapreduceLarge scale computing with mapreduce
Large scale computing with mapreduce
hansen3032
 
Dongwon Kim – A Comparative Performance Evaluation of Flink
Dongwon Kim – A Comparative Performance Evaluation of FlinkDongwon Kim – A Comparative Performance Evaluation of Flink
Dongwon Kim – A Comparative Performance Evaluation of Flink
Flink Forward
 
A Comparative Performance Evaluation of Apache Flink
A Comparative Performance Evaluation of Apache FlinkA Comparative Performance Evaluation of Apache Flink
A Comparative Performance Evaluation of Apache Flink
Dongwon Kim
 
Improving Apache Spark by Taking Advantage of Disaggregated Architecture
 Improving Apache Spark by Taking Advantage of Disaggregated Architecture Improving Apache Spark by Taking Advantage of Disaggregated Architecture
Improving Apache Spark by Taking Advantage of Disaggregated Architecture
Databricks
 
Apache hadoop, hdfs and map reduce Overview
Apache hadoop, hdfs and map reduce OverviewApache hadoop, hdfs and map reduce Overview
Apache hadoop, hdfs and map reduce Overview
Nisanth Simon
 
Apache Spark Fundamentals
Apache Spark FundamentalsApache Spark Fundamentals
Apache Spark Fundamentals
Zahra Eskandari
 
Hadoop ppt on the basics and architecture
Hadoop ppt on the basics and architectureHadoop ppt on the basics and architecture
Hadoop ppt on the basics and architecture
saipriyacoool
 
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptxCLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
CLOUD_COMPUTING_MODULE4_RK_BIG_DATA.pptx
bhuvankumar3877
 
11. From Hadoop to Spark 1:2
11. From Hadoop to Spark 1:211. From Hadoop to Spark 1:2
11. From Hadoop to Spark 1:2
Fabio Fumarola
 
Apache Spark Tutorial
Apache Spark TutorialApache Spark Tutorial
Apache Spark Tutorial
Ahmet Bulut
 
Optimization of modern web applications
Optimization of modern web applicationsOptimization of modern web applications
Optimization of modern web applications
Eugene Lazutkin
 
Apache Tez -- A modern processing engine
Apache Tez -- A modern processing engineApache Tez -- A modern processing engine
Apache Tez -- A modern processing engine
bigdatagurus_meetup
 
Tez big datacamp-la-bikas_saha
Tez big datacamp-la-bikas_sahaTez big datacamp-la-bikas_saha
Tez big datacamp-la-bikas_saha
Data Con LA
 
What no one tells you about writing a streaming app
What no one tells you about writing a streaming appWhat no one tells you about writing a streaming app
What no one tells you about writing a streaming app
hadooparchbook
 
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
What No One Tells You About Writing a Streaming App: Spark Summit East talk b...
Spark Summit
 
Spark_Intro_Syed_Academy
Spark_Intro_Syed_AcademySpark_Intro_Syed_Academy
Spark_Intro_Syed_Academy
Syed Hadoop
 
Ad

Recently uploaded (20)

IAS-slides2-ia-aaaaaaaaaaain-business.pdf
IAS-slides2-ia-aaaaaaaaaaain-business.pdfIAS-slides2-ia-aaaaaaaaaaain-business.pdf
IAS-slides2-ia-aaaaaaaaaaain-business.pdf
mcgardenlevi9
 
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Abodahab
 
LLM finetuning for multiple choice google bert
LLM finetuning for multiple choice google bertLLM finetuning for multiple choice google bert
LLM finetuning for multiple choice google bert
ChadapornK
 
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
Simran112433
 
C++_OOPs_DSA1_Presentation_Template.pptx
C++_OOPs_DSA1_Presentation_Template.pptxC++_OOPs_DSA1_Presentation_Template.pptx
C++_OOPs_DSA1_Presentation_Template.pptx
aquibnoor22079
 
Geometry maths presentation for begginers
Geometry maths presentation for begginersGeometry maths presentation for begginers
Geometry maths presentation for begginers
zrjacob283
 
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.pptJust-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
ssuser5f8f49
 
Digilocker under workingProcess Flow.pptx
Digilocker  under workingProcess Flow.pptxDigilocker  under workingProcess Flow.pptx
Digilocker under workingProcess Flow.pptx
satnamsadguru491
 
Conic Sectionfaggavahabaayhahahahahs.pptx
Conic Sectionfaggavahabaayhahahahahs.pptxConic Sectionfaggavahabaayhahahahahs.pptx
Conic Sectionfaggavahabaayhahahahahs.pptx
taiwanesechetan
 
Data Analytics Overview and its applications
Data Analytics Overview and its applicationsData Analytics Overview and its applications
Data Analytics Overview and its applications
JanmejayaMishra7
 
How iCode cybertech Helped Me Recover My Lost Funds
How iCode cybertech Helped Me Recover My Lost FundsHow iCode cybertech Helped Me Recover My Lost Funds
How iCode cybertech Helped Me Recover My Lost Funds
ireneschmid345
 
03 Daniel 2-notes.ppt seminario escatologia
03 Daniel 2-notes.ppt seminario escatologia03 Daniel 2-notes.ppt seminario escatologia
03 Daniel 2-notes.ppt seminario escatologia
Alexander Romero Arosquipa
 
VKS-Python Basics for Beginners and advance.pptx
VKS-Python Basics for Beginners and advance.pptxVKS-Python Basics for Beginners and advance.pptx
VKS-Python Basics for Beginners and advance.pptx
Vinod Srivastava
 
Simple_AI_Explanation_English somplr.pptx
Simple_AI_Explanation_English somplr.pptxSimple_AI_Explanation_English somplr.pptx
Simple_AI_Explanation_English somplr.pptx
ssuser2aa19f
 
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjks
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjksPpt. Nikhil.pptxnshwuudgcudisisshvehsjks
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjks
panchariyasahil
 
How to join illuminati Agent in uganda call+256776963507/0741506136
How to join illuminati Agent in uganda call+256776963507/0741506136How to join illuminati Agent in uganda call+256776963507/0741506136
How to join illuminati Agent in uganda call+256776963507/0741506136
illuminati Agent uganda call+256776963507/0741506136
 
Deloitte Analytics - Applying Process Mining in an audit context
Deloitte Analytics - Applying Process Mining in an audit contextDeloitte Analytics - Applying Process Mining in an audit context
Deloitte Analytics - Applying Process Mining in an audit context
Process mining Evangelist
 
Minions Want to eat presentacion muy linda
Minions Want to eat presentacion muy lindaMinions Want to eat presentacion muy linda
Minions Want to eat presentacion muy linda
CarlaAndradesSoler1
 
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbbEDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
JessaMaeEvangelista2
 
04302025_CCC TUG_DataVista: The Design Story
04302025_CCC TUG_DataVista: The Design Story04302025_CCC TUG_DataVista: The Design Story
04302025_CCC TUG_DataVista: The Design Story
ccctableauusergroup
 
IAS-slides2-ia-aaaaaaaaaaain-business.pdf
IAS-slides2-ia-aaaaaaaaaaain-business.pdfIAS-slides2-ia-aaaaaaaaaaain-business.pdf
IAS-slides2-ia-aaaaaaaaaaain-business.pdf
mcgardenlevi9
 
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Day 1 - Lab 1 Reconnaissance Scanning with NMAP, Vulnerability Assessment wit...
Abodahab
 
LLM finetuning for multiple choice google bert
LLM finetuning for multiple choice google bertLLM finetuning for multiple choice google bert
LLM finetuning for multiple choice google bert
ChadapornK
 
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
1. Briefing Session_SEED with Hon. Governor Assam - 27.10.pdf
Simran112433
 
C++_OOPs_DSA1_Presentation_Template.pptx
C++_OOPs_DSA1_Presentation_Template.pptxC++_OOPs_DSA1_Presentation_Template.pptx
C++_OOPs_DSA1_Presentation_Template.pptx
aquibnoor22079
 
Geometry maths presentation for begginers
Geometry maths presentation for begginersGeometry maths presentation for begginers
Geometry maths presentation for begginers
zrjacob283
 
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.pptJust-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
Just-In-Timeasdfffffffghhhhhhhhhhj Systems.ppt
ssuser5f8f49
 
Digilocker under workingProcess Flow.pptx
Digilocker  under workingProcess Flow.pptxDigilocker  under workingProcess Flow.pptx
Digilocker under workingProcess Flow.pptx
satnamsadguru491
 
Conic Sectionfaggavahabaayhahahahahs.pptx
Conic Sectionfaggavahabaayhahahahahs.pptxConic Sectionfaggavahabaayhahahahahs.pptx
Conic Sectionfaggavahabaayhahahahahs.pptx
taiwanesechetan
 
Data Analytics Overview and its applications
Data Analytics Overview and its applicationsData Analytics Overview and its applications
Data Analytics Overview and its applications
JanmejayaMishra7
 
How iCode cybertech Helped Me Recover My Lost Funds
How iCode cybertech Helped Me Recover My Lost FundsHow iCode cybertech Helped Me Recover My Lost Funds
How iCode cybertech Helped Me Recover My Lost Funds
ireneschmid345
 
VKS-Python Basics for Beginners and advance.pptx
VKS-Python Basics for Beginners and advance.pptxVKS-Python Basics for Beginners and advance.pptx
VKS-Python Basics for Beginners and advance.pptx
Vinod Srivastava
 
Simple_AI_Explanation_English somplr.pptx
Simple_AI_Explanation_English somplr.pptxSimple_AI_Explanation_English somplr.pptx
Simple_AI_Explanation_English somplr.pptx
ssuser2aa19f
 
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjks
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjksPpt. Nikhil.pptxnshwuudgcudisisshvehsjks
Ppt. Nikhil.pptxnshwuudgcudisisshvehsjks
panchariyasahil
 
Deloitte Analytics - Applying Process Mining in an audit context
Deloitte Analytics - Applying Process Mining in an audit contextDeloitte Analytics - Applying Process Mining in an audit context
Deloitte Analytics - Applying Process Mining in an audit context
Process mining Evangelist
 
Minions Want to eat presentacion muy linda
Minions Want to eat presentacion muy lindaMinions Want to eat presentacion muy linda
Minions Want to eat presentacion muy linda
CarlaAndradesSoler1
 
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbbEDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
EDU533 DEMO.pptxccccvbnjjkoo jhgggggbbbb
JessaMaeEvangelista2
 
04302025_CCC TUG_DataVista: The Design Story
04302025_CCC TUG_DataVista: The Design Story04302025_CCC TUG_DataVista: The Design Story
04302025_CCC TUG_DataVista: The Design Story
ccctableauusergroup
 
Ad

Spark shuffle introduction

  • 2. About Me Spark / Hadoop / Hbase / Phoenix contributor For spark mainly contributes in: •  Yarn •  Shuffle •  BlockManager •  Scala2.10 update •  Standalone HA •  Various other fixes. [email protected] Weibo @冷冻蚂蚁 blog.csdn.net/colorant
  • 3. Why Spark is fast(er) •  Whom do we compare to? •  What do we mean by fast? •  fast to write •  fast to run
  • 4. Why Spark is fast(er) cont. •  But the figure in previous page is some how misleading. •  The key is the flexible programming mode. •  Which lead to more reasonable data flow. •  Which lead to less IO operation. •  Especially for iterative heavy workloads like ML. •  Which potentially cut off a lot of shuffle operations needed. •  But, you won’t always be lucky. •  Many app logic did need to exchange a lot of data. •  In the end, you will still need to deal with shuffle •  And which usually impact performance a lot.
  • 8. How does shuffle come into the picture •  Spark run job stage by stage. •  Stages are build up by DAGScheduler according to RDD’s ShuffleDependency •  e.g. ShuffleRDD / CoGroupedRDD will have a ShuffleDependency •  Many operator will create ShuffleRDD / CoGroupedRDD under the hook. •  Repartition/CombineByKey/GroupBy/ReduceByKey/cogroup •  many other operator will further call into the above operators •  e.g. various join operator will call cogroup. •  Each ShuffleDependency maps to one stage in Spark Job and then will lead to a shuffle.
  • 9. So everyone should have seen this before join   union   groupBy   map   Stage  3   Stage  1   Stage  2   A:   B:   C:   D:   E:   F:   G:  
  • 10. why shuffle is expensive •  When doing shuffle, data no longer stay in memory only •  For spark, shuffle process might involve •  data partition: which might involve very expensive data sorting works etc. •  data ser/deser: to enable data been transfer through network or across processes. •  data compression: to reduce IO bandwidth etc. •  DISK IO: probably multiple times on one single data block •  E.g. Shuffle Spill, Merge combine
  • 11. History •  Spark 0.6-0.7, same code path with RDD’s persistent method, can choose MEMORY_ONLY and DISK_ONLY (default). •  Spark 0.8-0.9: •  separate shuffle code path from BM and create ShuffleBlockManager and BlockObjectWriter only for shuffle, now shuffle data can only be written to disk. •  Shuffle optimization: Consolidate shuffle write. •  Spark 1.0, pluggable shuffle framework. •  Spark 1.1, sort-based shuffle implementation. •  Spark 1.2 netty transfer service reimplementation. sort- based shuffle by default •  Spark 1.2+ on the go: external shuffle service etc.
  • 13. 13 Pluggable Shuffle Framework •  ShuffleManager •  Manage shuffle related components, registered in SparkEnv, configured through SparkConf, default is sort (pre 1.2 is hash), •  ShuffleWriter •  Handle shuffle data output logics. Will return MapStatus to be tracked by MapOutputTracker. •  ShuffleReader •  Fetch shuffle data to be used by e.g. ShuffleRDD •  ShuffleBlockManager •  Manage the mapping relation between abstract bucket and materialized data block.
  • 14. High level data flow BlockManager HashShuffleManager DiskBlockManager FileShuffleBlockManager Local File System SortShuffleManager IndexShuffleBlockManager GetBlockData BlockTransferService GetBlockData Direct mapping or mapping by File Groups Map to One Data File and One Index File per mapId Just do one-one File mapping
  • 15. 15 Hash Based Shuffle - Shuffle Writer •  Basic shuffle writer Map  Task Map  Task Map  Task Map  Task File File File File File File File File File File File File File File File File Aggregator AggregatorAggregatorAggregator Each bucket is mapping to a single file
  • 16. 16 Hash Based Shuffle - Shuffle Writer •  Consolidate Shuffle Writer Each bucket is mapping to a segment of file Aggregator Aggregator Aggregator Aggregator
  • 17. 17 Hash Based Shuffle - Shuffle Writer •  Basic Shuffle Writer •  M * R shuffle spill files •  Concurrent C * R opened shuffle files. •  If shuffle spill enabled, could generate more tmp spill files say N. •  Consolidate Shuffle Writer •  Reduce the total spilled files into C * R if (M >> C) •  Concurrent opened is the same as the basic shuffle writer. •  Memory consumption •  Thus Concurrent C * R + N file handlers. •  Each file handler could take up to 32~100KB+ Memory for various buffers across the writer stream chain.
  • 18. 18 Sort Based Shuffle - Shuffle Writer •  Sort Shuffle Writer Map  Task Map  Task Map  Task Map  Task FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileFile FileFile FileFile FileFile ExternalSorter ExternalSorterExternalSorterExternalSorter
  • 19. 19 Sort Based Shuffle - Shuffle Writer •  Each map task generates 1 shuffle data file + 1 index file •  Utilize ExternalSorter to do the sort works. •  If map-side combine is required, data will be sorted by key and partition for aggregation. Otherwise data will only be sorted by partition. •  If reducer number <= 200 and no need to do aggregation or ordering, data will not be sorted at all. •  Will go with hash way and spill to separate files for each reduce partition, then merge them into one per map for final output.
  • 20. 20 Hash Based Shuffle - Shuffle Reader •  Actually, at present, Sort Based Shuffle also go with HashShuffleReader Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Reduce  TaskReduce  Task Reduce  Task Reduce  Task Reduce  Task Aggregator AggregatorAggregatorAggregator
  • 22. Related conceptions •  BlockTransferService •  Provide a general interface for ShuffleFetcher and working with BlockDataManager to get local data. •  ShuffleClient •  Wrap up the fetching data process for the client side, say setup TransportContext, new TransportClient etc. •  TransportContext •  Context to setup the transport layer •  TransportServer •  low-level streaming service server •  TransportClient •  Client for fetching consecutive chunks TransportServer
  • 24. ShuffleManager Data Flow 24 BlockManager NettyBlockTransferService GetBlockData BlockDataManager TransportClient NettyBlockTransferService TransportServer BlockStoreShuffleFetcher ShuffleBlockFetcherIterator Block Manager Local Blocks Remote Blocks Local Remote HashShuffleReader fetch ShuffleManager ShuffleBlockManager GetBlockData clientHandler TransportChannel HandlerclientHandler TransportChannel Handler Fetch Request Fetch Results
  • 26. External Shuffle Service •  Design goal •  allow for the service to be long-running •  possibly much longer-running than Spark •  support multiple version of Spark simultaneously etc. •  can be integrated into YARN NodeManager, Standalone Worker, or on its own •  The entire service been ported to Java •  do not include Spark's dependencies •  full control over the binary compatibility of the components •  not depend on the Scala runtime or version.
  • 27. External Shuffle Service •  Current Status •  Basic framework seems ready. •  A Network module extracted from the core module •  BlockManager could be configured with executor built-in shuffle service or external standalone shuffle service •  A standaloneWorkerShuffleService could be launched by worker •  Disabled by default. •  How it works •  Shuffle data is still written by the shuffleWriter to local disks. •  The external shuffle service knows how to read these files on disks (executor will registered related info to it, e.g. shuffle manager type, file dir layout etc.), it follow the same rules applied for written these file, so it could serve the data correctly.
  • 28. 28 Sort Merge Shuffle Reader •  Background: •  Current HashShuffleReader does not utilize the sort result within partition in map-side. •  The actual by key sort work is always done at reduce side. •  While the map side will do by-partition sort anyway ( sort shuffle ) •  Change it to a by-key-and-partition sort does not bring many extra overhead. •  Current Status •  [WIP] https://github.com/apache/spark/pull/3438
  • 29. Some shuffle related configs •  spark.shuffle.spill (true) •  spark.shuffle.memoryFraction (0.2) •  spark.shuffle.manager [sort]/hash •  spark.shuffle.sort.bypassMergeThreshold (200) •  spark.shuffle.blockTransferService [netty]/nio •  spark.shuffle.consolidateFiles (false) •  spark.shuffle.service.enabled (false)
  • 30. What’s next? •  Other custom shuffle logic? •  Alternative way to save shuffle data blocks •  E.g. in memory (again) •  Other transport mechanism? •  Break stage barrier? •  To fetch shuffle data when part of the map tasks are done. •  Push mode instead of pull mode?
  • 31. Thanks to Jerry Shao •  Some of this ppt’s material came from Jerry Shao@Intel weibo: @saisai_shao •  Jerry also contributes a lot of essential patches for spark core / spark streaming etc.
  • 32. Join Us ! • 加盟 / 合作 / 讨论 统统欢迎 • 数据平台开发,大数据相关技术,只要够 Cool,我们都玩 • [email protected]