Preprint
Article

This version is not peer-reviewed.

Minimising Stochastic Complexity with Ridge Regression

Submitted:

12 May 2026

Posted:

12 May 2026

You are already at the latest version

Abstract
We derive a penalty strength criterion for ridge regression using stochastic complexity, which is a refined variant of the minimum description length principle. Since stochastic complexity doesn’t typically account for the effect of regularisation on complexity, despite its ability to simplify models, we are required to make a slight modification to the un- derlying coding scheme. Our scheme makes use of a weighted ensemble of regularised model fits rather than a mixture of maximum likelihood estimates. Under this modification, regularisation is interpreted as reducing model complexity by constraining flexibility. In the case of ridge regression, the complexity penalty term that we derive can be expressed analytically as the log determinant of the residual operator. We demonstrate the effect of this complexity penalty by fitting a linear readout to a reservoir computer.
Keywords: 
;  ;  ;  ;  ;  
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated