×
MindLuster Logo

Apache Spark Tips Tricks

We Appreciate Your Feedback

Be the First One Review This Course

Excellent
0 Reviews
Good
0 Reviews
medium
0 Reviews
Acceptable
0 Reviews
Not Good
0 Reviews
0
0 Reviews

While Spark chooses good reasonable defaults for your data, if your Spark job runs out of memory or runs slowly, bad partitioning could be at fault. If your dataset is large, you can try repartitioning (using the repartition method) to a larger number to allow more parallelism on your job.

Python programming language