Fig. 3From: A parallel and distributed stochastic gradient descent implementation using commodity clustersVisualization of training time with respect to training data size in thousands. Each line represents results from a cluster with N number of nodes. Note that the X and Y scale are logarithmic, and clearly shows the constant overhead of the larger clusters on small datasetsBack to article page