
Efficient Mini-batch Training for Stochastic Optimization
Published on Feb 4, 20252286 Views
Stochastic gradient descent (SGD) is a popular technique for large-scale optimization problems in machine learning. In order to parallelize SGD, minibatch training needs to be employed to reduce the c