Preserved in Portico This version is not peer-reviewed
DAPT: A Package Enabling Distributed Automated Parameter Testing
: Received: 1 March 2021 / Approved: 2 March 2021 / Online: 2 March 2021 (22:21:47 CET)
: Received: 3 May 2021 / Approved: 10 May 2021 / Online: 10 May 2021 (09:47:54 CEST)
A peer-reviewed article of this Preprint also exists.
Journal reference: Gigabyte 2021
Modern agent-based models (ABM) and other simulation models require evaluation and testing of many different parameters. Managing that testing for large scale parameter sweeps (grid searches) as well as storing simulation data requires multiple, potentially customizable steps that may vary across simulations. Furthermore, parameter testing, processing, and analysis are slowed if simulation and processing jobs cannot be shared across teammates or computational resources. While high-performance computing (HPC) has become increasingly available, models can often be tested faster through the use of multiple computers and HPC resources. To address these issues, we created the Distributed Automated Parameter Testing (DAPT) Python package. By hosting parameters in an online (and often free) "database", multiple individuals can run parameter sets simultaneously in a distributed fashion, enabling ad hoc crowdsourcing of computational power. Combining this with a flexible, scriptable tool set, teams can evaluate models and assess their underlying hypotheses quickly. Here we describe DAPT and provide an example demonstrating its use.
DAPT; workflow; agent-based modeling; model exploration; crowdsourcing
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
We encourage comments and feedback from a broad range of readers. See criteria for comments and our diversity statement.