Article
Version 1
Preserved in Portico This version is not peer-reviewed
Adaptive Batch Size Selection in Active Learning for Regression
Version 1
: Received: 25 January 2022 / Approved: 28 January 2022 / Online: 28 January 2022 (15:03:10 CET)
How to cite: Faulds, A. Adaptive Batch Size Selection in Active Learning for Regression. Preprints 2022, 2022010441. https://doi.org/10.20944/preprints202201.0441.v1 Faulds, A. Adaptive Batch Size Selection in Active Learning for Regression. Preprints 2022, 2022010441. https://doi.org/10.20944/preprints202201.0441.v1
Abstract
Training supervised machine learning models requires labeled examples. A judicious choice of examples is helpful when there is a significant cost associated with assigning labels. This article improves upon a promising extant method – Batch-mode Expected Model Change Maximization (B-EMCM) method – for selecting examples to be labeled for regression problems. Specifically, it develops and evaluates alternate strategies for adaptively selecting batch size in B-EMCM. By determining the cumulative error that occurs from the estimation of the stochastic gradient descent, a stop criteria for each iteration of the batch can be specified to ensure that selected candidates are the most beneficial to model learning. This new methodology is compared to B-EMCM via mean absolute error and root mean square error over ten iterations benchmarked against machine learning data sets. Using multiple data sets and metrics across all methods, one variation of AB-EMCM, the max bound of the accumulated error (AB-EMCM Max), showed the best results for an adaptive batch approach. It achieved better root mean squared error (RMSE) and mean absolute error (MAE) than the other adaptive and non-adaptive batch methods while reaching the result in nearly the same number of iterations as the non-adaptive batch methods.
Keywords
Active learning (AL); batch mode; expected model change; linear regression; nonlinear regression
Subject
Computer Science and Mathematics, Artificial Intelligence and Machine Learning
Copyright: This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Comments (0)
We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.
Leave a public commentSend a private comment to the author(s)
* All users must log in before leaving a comment