Faster Spark joins
- Shuffle hash join/ reduce-side join
- Misc partitioning tips
- Broadcast join / map-side join
- Even faster map-side joins
- Filter Large RDD
Some of the most useful operations we get with keyed data comes from using it together with other keyed data. Joining data together is probably one of the most common operations on a pair RDD, and we have a full range of options including right and left outer joins, cross joins, and inner joins. In order to join data, spark needs each key-to-be-joined to exist on the same partition. There are several ways to achieve co-location which are documented below
Shuffle hash join/ reduce-side join
By default, when a RDD is read from textfile or HDFS, it follows the partitioning scheme of hadoop filesystem. As seen above, a pre-requisite for join is co-location.This operation will hash all the keys of both datasets, sending elements with the same key hash across the network to the same machine, and then join together the elements with the same key on that machine.
For example, userData and events are shuffled causing heavy network traffic.
This operation is also called reduce-side join because the actual join process happens in the reduce phase. It follows the traditional map-shuffle-reduce flow.
In cases where a largeRDD is used repeatedly, over and over again, shuffle-hash join would cause re-shuffling on each iteration making it costly. To avoid this, largeRDD is partitioned using RangePartitioner/HashPartitioner and spark uses this information to make the smallRDD use the same partitioner to find the partition to which the keys goes to. This way, a huge shuffle is avoided.
- Use of partitionBy before any action is performed on largeRDD doesn’t cost extra, since it is lazily evaluated.
- Persist the RDD just after its been partitioned, if forgot would cause re-evaluation of RDD’s complete lineage
Functions other than join which take partitioning as advantage are cogroup(), groupWith(), leftOuterJoin(), rightOuterJoin(), groupByKey(), reduceByKey(), combineByKey(), and lookup()
Misc partitioning tips
ShuffleHashJoin can be avoided in the below scenarios
Both tables use the same partitioner
Second RDD is a derivative of First RDD -
Assume First RDD is hash partitioned and Second RDD is derived by using mapValues on First RDD. This way, both are cached on the same machine.
If one of the RDDs is already shuffled before -
Many spark operations automatically result in an RDD with known partitioning information and join takes advantage of this information. For example, sortByKey and groupByKey result in a partitioned RDD, with a valid non-default partitioner. This behaves the same way as of
Above, B is not shuffled when joined with F because groupBy is applied on B.
Broadcast join / map-side join
In order to avoid the shuffle-reduce phase, join operation is delegated to map-stage where-in, one of the tables (smaller one) is broadcasted in-memory to each mapper. This works only when one of the tables is relatively small.
Below are the steps to perform broadcast join:
- Create a RDD for both tables on which join is to be performed
- Download Small RDD to the driver, create map and broadcast on each execution node
- Map over each row of Large RDD, retrieve value (from Small RDD) using key from the iterator of Large RDD.
- Broadcast join will be executed concurrently for each partition since each partition has its own copy of the small RDD.
Even faster map-side joins
There are some scaling problems with map-side join. When thousands of mappers read the small join table from the Hadoop Distributed File System (HDFS) into memory at the same time, the join table easily becomes the performance bottleneck, causing the mappers to time out during the read operations.
The basic idea of optimization is to create a new MapReduce local task just before the original join MapReduce task. This new task reads the small table data from HDFS to an in-memory hash table. After reading, it serializes the in-memory hash table into a hashtable file. In the next stage, when the MapReduce task is launching, it uploads this hashtable file to the Hadoop distributed cache, which populates these files to each mapper’s local disk. So all the mappers can load this persistent hashtable file back into memory and do the join work as before.
After optimization, the small table needs to be read just once. Also if multiple mappers are running on the same machine, the distributed cache only needs to push one copy of the hashtable file to this machine.
Filter Large RDD
When joining an extremely large table and a subset of this table, a huge shuffle takes place. Join causes majority of the large table to drop. For example, when you’re joining two RDDs namely worldRDD and indiaRDD, a join would cause majority of worldRDD to drop. An extremely fast (10x speedup) is to filter the worldRDD using the keys of indiaRDD and then performing a join.
This method is faster and causes less data to be shuffled over the network.
- Always explore the distribution of keys before performing a full-blown shufflejoin.
- The efficiency gain here depends on the filter operation that reduces the size of larger RDD. If there are not a lot of entries lost here (e.g., because indiaRDD is some kind of large dimension table), there is nothing to be gained with this strategy