Question: Can map/reduce in Pyspark be used to process a massive dataset (that does not fit in a computer)? If so, why? If not, why?
Can map/reduce in Pyspark be used to process a massive dataset (that does not fit in a computer)? If so, why? If not, why?
Step by Step Solution
There are 3 Steps involved in it
Get step-by-step solutions from verified subject matter experts
