"Small is the number of people who see with their eyes and think with their minds." Einstein
MapReduce
Split the data set in multiple pieces of data sets. Then multiple machines will execute such a piece of data.
Batch gradient descent:
Machine 1: Use
Machine 2: Use
Machine 3: Use
Machine 4: Use
At the end send all these results to a master server and combine:
where
And this last equation is equal to the first equation of Batch Gradient Descent.
Summation
Many learning algorithms can be expressed as computing sums of functions over the trainging set.