TY - GEN
T1 - Evaluating parallel minibatch training for machine learning applications
AU - Dreiseitl, Stephan
PY - 2018
Y1 - 2018
N2 - The amount of data available for analytics applications continues to rise. At the same time, there are some application areas where security and privacy concerns prevent liberal dissemination of data. Both of these factors motivate the hypothesis that machine learning algorithms may benefit from parallelizing the training process (for large amounts of data) and/or distributing the training process (for sensitive data that cannot be shared). We investigate this hypothesis by considering two real-world machine learning tasks (logistic regression and sparse autoencoder), and empirically test how a model’s performance changes when its parameters are set to the arithmetic means of parameters of models trained on minibatches, i.e., horizontally split portions of the data set. We observe that iterating the minibatch training and parameter averaging process for a small number of times results in models with performance only slightly worse that of models trained on the full data sets.
AB - The amount of data available for analytics applications continues to rise. At the same time, there are some application areas where security and privacy concerns prevent liberal dissemination of data. Both of these factors motivate the hypothesis that machine learning algorithms may benefit from parallelizing the training process (for large amounts of data) and/or distributing the training process (for sensitive data that cannot be shared). We investigate this hypothesis by considering two real-world machine learning tasks (logistic regression and sparse autoencoder), and empirically test how a model’s performance changes when its parameters are set to the arithmetic means of parameters of models trained on minibatches, i.e., horizontally split portions of the data set. We observe that iterating the minibatch training and parameter averaging process for a small number of times results in models with performance only slightly worse that of models trained on the full data sets.
KW - Distributed machine learning
KW - Logistic regression
KW - Minibatch training
KW - Sparse autoencoders
UR - http://www.scopus.com/inward/record.url?scp=85041833589&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-74718-7_48
DO - 10.1007/978-3-319-74718-7_48
M3 - Conference contribution
AN - SCOPUS:85041833589
SN - 9783319747170
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 400
EP - 407
BT - Computer Aided Systems Theory – EUROCAST 2017 - 16th International Conference, Revised Selected Papers
A2 - Moreno-Diaz, Roberto
A2 - Quesada-Arencibia, Alexis
A2 - Pichler, Franz
PB - Springer
T2 - 16th International Conference on Computer Aided Systems Theory, EUROCAST 2017
Y2 - 19 February 2017 through 24 February 2017
ER -