Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Adaptive Batch Size Selection in Active Learning for Regression

Version 1 : Received: 25 January 2022 / Approved: 28 January 2022 / Online: 28 January 2022 (15:03:10 CET)

How to cite: Faulds, A. Adaptive Batch Size Selection in Active Learning for Regression. Preprints 2022, 2022010441. https://doi.org/10.20944/preprints202201.0441.v1 Faulds, A. Adaptive Batch Size Selection in Active Learning for Regression. Preprints 2022, 2022010441. https://doi.org/10.20944/preprints202201.0441.v1

Abstract

Training supervised machine learning models requires labeled examples. A judicious choice of examples is helpful when there is a significant cost associated with assigning labels. This article improves upon a promising extant method – Batch-mode Expected Model Change Maximization (B-EMCM) method – for selecting examples to be labeled for regression problems. Specifically, it develops and evaluates alternate strategies for adaptively selecting batch size in B-EMCM. By determining the cumulative error that occurs from the estimation of the stochastic gradient descent, a stop criteria for each iteration of the batch can be specified to ensure that selected candidates are the most beneficial to model learning. This new methodology is compared to B-EMCM via mean absolute error and root mean square error over ten iterations benchmarked against machine learning data sets. Using multiple data sets and metrics across all methods, one variation of AB-EMCM, the max bound of the accumulated error (AB-EMCM Max), showed the best results for an adaptive batch approach. It achieved better root mean squared error (RMSE) and mean absolute error (MAE) than the other adaptive and non-adaptive batch methods while reaching the result in nearly the same number of iterations as the non-adaptive batch methods.

Keywords

Active learning (AL); batch mode; expected model change; linear regression; nonlinear regression

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.