So I am trying to learn Spark using Python (Pyspark). I want to know how the function mapPartitions
work. That is what Input it takes and what Output it gives. I couldn't find any proper example from the internet. Lets say, I have an RDD object containing lists, such as below.
[ [1, 2, 3], [3, 2, 4], [5, 2, 7] ]
And I want to remove element 2 from all the lists, how would I achieve that using mapPartitions
.
MapPartition works on a partition at a time. MapPartition returns after processing all the rows in the partition. MapPartition output is retained in memory, as it can return after processing all the rows in a particular partition.
Return a new RDD by applying a function to each partition of this RDD.
In PySpark, the map (map()) is defined as the RDD transformation that is widely used to apply the transformation function (Lambda) on every element of Resilient Distributed Datasets(RDD) or DataFrame and further returns a new Resilient Distributed Dataset(RDD).
mapPartition
should be thought of as a map operation over partitions and not over the elements of the partition. It's input is the set of current partitions its output will be another set of partitions.
The function you pass to map
operation must take an individual element of your RDD
The function you pass to mapPartition
must take an iterable of your RDD type and return an iterable of some other or the same type.
In your case you probably just want to do something like:
def filter_out_2(line): return [x for x in line if x != 2] filtered_lists = data.map(filterOut2)
If you wanted to use mapPartition
it would be:
def filter_out_2_from_partition(list_of_lists): final_iterator = [] for sub_list in list_of_lists: final_iterator.append( [x for x in sub_list if x != 2]) return iter(final_iterator) filtered_lists = data.mapPartition(filterOut2FromPartion)
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With