Spark has Spark ML which is used for running machine learning programs when the data set is of huge volume. Running that without Spark would take long time . Before Spark, the data processing programs on huge volume of data was run using map reduce programs but what about machine learning programs on huge volume dataset. How was it done on large dataset before Spark ML?
Fire me boy! Cool, soothing, shameless self promotion: