Submitted:
29 September 2025
Posted:
02 October 2025
Read the latest preprint version here
Abstract
Keywords:
Introduction & Motivation
Our Experience
- Loading of the incorrect data files.
- Errors in the calculations of test statistics.
- Indexing errors in vector operations leading to row-shifted data (with subjects being associated with incorrect data).
- Incorrect custom implementations of complex operations (i.e., nested cross-validation).
- Unintended use of default settings in (external) open-source toolboxes. In this particular case, issues in the code were detected after publication. Once the programming mistake was detected, the journal was proactively contacted and informed of the situation, analyses were rerun with the correct settings, and the updated results were published as a Corrigendum [18].
A Proposal
Fundamental
|
Advanced
|
Code Review and Analysis Plans
Benefits, Costs, and a Way Forward
References
- Soergel, D.A.W. Rampant software errors may undermine scientific results. F1000Research 2015, 3, 303. [Google Scholar] [CrossRef]
- McConnell, S. Code complete, Second edition. Redmond, Washington: Microsoft Press, 2004.
- Arvanitou, E.-M.; Ampatzoglou, A.; Chatzigeorgiou, A.; Carver, J.C. Software engineering practices for scientific software development: A systematic mapping study. Journal of Systems and Software 2021, 172, 110848. [Google Scholar] [CrossRef]
- Storer, T. Bridging the Chasm: A Survey of Software Engineering Practice in Scientific Programming. ACM Comput. Surv. 2018, 50, 1–32. [Google Scholar] [CrossRef]
- Woolston, C. Why science needs more research software engineers. Nature 2022. [CrossRef] [PubMed]
- RSECon 2025. RSECon25. Available online: https://rsecon25.society-rse.org/ (accessed on 26 April 2025).
- Nordic RSE. Nordic RSE. 2025.. Available online: https://nordic-rse.
- nl-rse. nl-rse. Available online: https://nl-rse.org/ (accessed on April 2025).
- Merali, Z. Computational science:...Error. Nature 2010, 467, 775–777. [Google Scholar] [CrossRef]
- Heaton, D.; Carver, J.C. Claims about the use of software engineering practices in science: A systematic literature review. Information and Software Technology 2015, 67, 207–219. [Google Scholar] [CrossRef]
- Schmidberger, M.; Brügge, B. Need of Software Engineering Methods for High Performance Computing Applications. in 2012 11th International Symposium on Parallel and Distributed Computing 2012, 40–46. [Google Scholar] [CrossRef]
- Nangia, U.; Katz, D.S. Track 1 Paper: Surveying the U.S. National Postdoctoral Association Regarding Software Use and Training in Research. Bytes 2017, 4361681. [Google Scholar] [CrossRef]
- Wilson, G. Software Carpentry: Getting Scientists to Write Better Code by Making Them More Productive. Computing in Science & Engineering. [CrossRef]
- Reynolds, M.; Science Is Full of Errors. Bounty Hunters Are Here to Find Them. Wired. Available online: https://www.wired.com/story/bounty-hunters-are-here-to-save-academia-bug-bounty/ (accessed on 24 April 2025).
- Knight, W. Sloppy Use of Machine Learning Is Causing a ‘Reproducibility Crisis’ in Science Wired. Available online: https://www.wired.com/story/machine-learning-reproducibility-crisis/ (accessed on 24 April 2025).
- He, X.; et al. Author Correction: Temporal dynamics in viral shedding and transmissibility of COVID-19. Nat Med 2020, 26, 1491–1493. [Google Scholar] [CrossRef]
- Ashcroft, P.; et al. COVID-19 infectivity profile correction. Swiss Medical Weekly 2020, 150. [Google Scholar] [CrossRef]
- Iglesias, S.; et al. Hierarchical Prediction Errors in Midbrain and Basal Forebrain during Sensory Learning. Neuron 2019, 101, 1196–1201. [Google Scholar] [CrossRef] [PubMed]
- Marcus, A.; Doing the right thing: Psychology researchers retract paper three days after learning of coding error. Retraction Watch. Available online: https://retractionwatch.com/2019/08/13/doing-the-right-thing-psychology-researchers-retract-paper-three-days-after-learning-of-coding-error/ (accessed on 4 September 2025).
- Dolk, T.; Freigang, C.; Bogon, J.; Dreisbach, G. RETRACTED: Auditory (dis-)fluency triggers sequential processing adjustments. Acta Psychologica 2018, 191, 69–75. [Google Scholar] [CrossRef] [PubMed]
- Henson, K.E.; Jagsi, R.; Cutter, D.; McGale, P.; Taylor, C.; Darby, S.C. RETRACTED: Inferring the Effects of Cancer Treatment: Divergent Results From Early Breast Cancer Trialists’ Collaborative Group Meta-Analyses of Randomized Trials and Observational Data From SEER Registries. JCO 2016, 34, 803–809. [Google Scholar] [CrossRef] [PubMed]
- ”In Sickness and in Health? Physical Illness as a Risk Factor for Marital Dissolution in Later Life - Amelia Karraker, Kenzie Latham, 2015. Available online: https://journals.sagepub.com/doi/abs/10.1177/0022146515596354 (accessed on 4 September 2025).
- Mandhane, P.J. Notice of Retraction: Hahn LM, et al. Post–COVID-19 Condition in Children. JAMA Pediatrics. 2023;177(11):1226-1228. JAMA Pediatr 2024, 178, 1085–1086. [Google Scholar] [CrossRef]
- Reinhart, C.; Rogoff, K. Errata: ‘Growth in A Time of Debt,’” Harvard University, 2013. Available online: https://carmenreinhart.com/wp-content/uploads/2020/02/36_data.pdf (accessed on 4 September 2025).
- Wiegers, K. Humanizing Peer Reviews.” Available online:. Available online: https://web.archive.org/web/20060315135514/http://www.processimpact.com/articles/humanizing_reviews.html (accessed on 25 April 2025).
- Atwood, J.; Code Reviews: Just Do It. Coding Horror. Available online: https://blog.codinghorror.com/code-reviews-just-do-it/ (accessed on 17 April 2025).
- Miller, G. A Scientist’s Nightmare: Software Problem Leads to Five Retractions. Science 2006, 314, 1856–1857. [Google Scholar] [CrossRef]
- MIT, “The Missing Semester of Your CS Education. Missing Semester. Available online: https://missing.csail.mit.edu/ (accessed on 25 April 2025).
- Balaban, G.; Grytten, I.; Rand, K.D.; Scheffer, L.; Sandve, G.K. Ten simple rules for quick and dirty scientific programming. PLOS Computational Biology 2021, 17, e1008549. [Google Scholar] [CrossRef]
- Green, R. How To Write Unmaintainable Code.” Available online:. Available online: https://www.doc.ic.ac.uk/~susan/475/unmain.html (accessed on 25 2025).
- Motivation: Reproducible research documentation.” Available online:. Available online: https://coderefinery.github.io/reproducible-research/motivation/ (accessed on April 2025).
- Martin, R.C. , Clean Code: A Handbook of Agile Software Craftsmanship, 1st edition. Upper Saddle River, NJ: Pearson, 2008.
- Lynch, M. ; published, “Rules for Writing Software Tutorials.” Available online:. Available online: https://refactoringenglish.com/ (accessed on 26 April 2025).
- Lessons, S.C. 2025. Available online: https://software-carpentry.
- CodeRefinery.” Available online:. Available online: https://coderefinery.org/lessons/ (accessed on April 2025).
- UNIVERSE-HPC, “Byte-sized RSE. UNIVERSE-HPC. Available online: http://www.universe-hpc.ac.uk/events/byte-sized-rse/ (accessed on 26 April 2025).
- Hastings, J.; Haug, K.; Steinbeck, C. Ten recommendations for software engineering in research. GigaSci, 2014. [Google Scholar] [CrossRef]
- Lynch, M. How to Do Code Reviews Like a Human (Part One)” Available online:. Available online: https://mtlynch.io/human-code-reviews-1/ (accessed on 26 April 2025).
- Lynch, M. How to Do Code Reviews Like a Human (Part Two). Available online: https://mtlynch.io/human-code-reviews-2/ (accessed on 26 April 2025).
- Lynch, M. How to Make Your Code Reviewer Fall in Love with You” Available online:. Available online: https://mtlynch.io/code-review-love/ (accessed on 26 April 2025).
- Tatham, S. Code review antipatterns.” Available online:. Available online: https://www.chiark.greenend.org.uk/~sgtatham/quasiblog/code-review-antipatterns/ (accessed on 26 April 2025).
- Ivimey--Cook, E.R.; et al. Implementing code review in the scientific workflow: Insights from ecology and evolutionary biology. Journal of Evolutionary Biology 2023, 36, 1347–1356. [Google Scholar] [CrossRef]
- Rokem, A. Ten simple rules for scientific code review. PLOS Computational Biology 2024, 20, e1012375. [Google Scholar] [CrossRef]
- Vable, A.M.; Diehl, S.F.; Glymour, M.M. Code Review as a Simple Trick to Enhance Reproducibility, Accelerate Learning, and Improve the Quality of Your Team’s Research. American Journal of Epidemiology 2021, 190, 2172–2177. [Google Scholar] [CrossRef]
- Gee, T. , What to Look for in a Code Review. Leanpub, 2015. Available online: https://leanpub.next/whattolookforinacodereview (accessed on 17 April 2025).
- ”What to look for in a code review. eng-practices. Available online: https://google.github.io/eng-practices/review/reviewer/looking-for.html (accessed on 17 April 2025).
- ”Common pitfalls and recommended practices. scikit-learn. Available online: https://scikit-learn.org/stable/common_pitfalls.html (accessed on 17 April 2025).
- Barton, J.; awesome-code-review. GitHub. Available online: https://github.com/joho/awesome-code-review/blob/main/readme.md (accessed on 17 April 2025).
- Nosek, B.A.; Ebersole, C.R.; DeHaven, A.C.; Mellor, D.T. The preregistration revolution. Proceedings of the National Academy of Sciences 2018, 115, 2600–2606. [Google Scholar] [CrossRef]
- Brodeur, A.; Cook, N.M.; Hartley, J.S.; Heyes, A. Do Pre-Registration and Pre-analysis Plans Reduce p-Hacking and Publication Bias? GLO Discussion Paper, Working Paper 1147, 2022. Available online: https://www.econstor.eu/handle/10419/262738 (accessed on 2 September 2025).
- Pornprasit, C.; Tantithamthavorn, C. Fine-tuning and prompt engineering for large language models-based code review automation. Information and Software Technology 2024, 175, 107523. [Google Scholar] [CrossRef]
- Yu, Y.; et al. Fine-Tuning Large Language Models to Improve Accuracy and Comprehensibility of Automated Code Review. ACM Trans. Softw. Eng. Methodol. 2024, 34, 14–1. [Google Scholar] [CrossRef]
- Haider, M.A.; Mostofa, A.B.; Mosaddek, S.S.B.; Iqbal, A.; Ahmed, T. Prompting and Fine-tuning Large Language Models for Automated Code Review Comment Generation. arXiv 2024, arXiv:2411.10129. [Google Scholar] [CrossRef]
- Lu, J.; Yu, L.; Li, X.; Yang, L.; Zuo, C. LLaMA-Reviewer: Advancing Code Review Automation with Large Language Models through Parameter-Efficient Fine-Tuning. In Proceedings of the 2023 IEEE 34th International Symposium on Software Reliability Engineering (ISSRE), 2023 pp. 647–658. [Google Scholar] [CrossRef]
- Kwon, S.; Lee, S.; Kim, T.; Ryu, D.; Baik, J. Exploring LLM-based Automated Repairing of Ansible Script in Edge-Cloud Infrastructures. Journal of Web Engineering 2023, 889–912. [Google Scholar] [CrossRef]
- Li, Z.; et al. Automating code review activities by large-scale pre-training. In Proceedings of the 30th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering, in ESEC/FSE 2022. New York, NY, 2022, USA: Association for Computing Machinery; pp. 1035–1047. [CrossRef]
- Fan, L.; Liu, J.; Liu, Z.; Lo, D.; Xia, X.; Li, S. Exploring the Capabilities of LLMs for Code Change Related Tasks. arXiv 2024, arXiv:2407.02824. [Google Scholar] [CrossRef]
- Martins, G.F.; Firmino, E.C.M.; De Mello, V.P. The Use of Large Language Model in Code Review Automation: An Examination of Enforcing SOLID Principles. in Artificial Intelligence in HCI, H. Degen and S. Ntoa, Eds., Cham: Springer Nature Switzerland, 2024, 86–97. [CrossRef]
- Tang, X.; et al. CodeAgent: Autonomous Communicative Agents for Code Review. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Y. Al-Onaizan, M. Bansal, and Y.-N. Chen, Eds., Miami, Florida, 2024, 11279–11313., USA: Association for Computational Linguistics. [CrossRef]
- Zhao, Z.; Xu, Z.; Zhu, J.; Di, P.; Yao, Y.; Ma, X. The Right Prompts for the Job: Repair Code-Review Defects with Large Language Model. arXiv arXiv:2312.17485, 2023. [CrossRef]
- Gibney, E. AI tools are spotting errors in research papers: inside a growing movement. Nature 2025. [Google Scholar] [CrossRef]
- Krag, C.H.; et al. Large language models for abstract screening in systematic- and scoping reviews: A diagnostic test accuracy study. medRxiv 2024. [CrossRef]
- Nashaat, M.; Miller, J. Towards Efficient Fine-Tuning of Language Models With Organizational Data for Automated Software Review. IEEE Transactions on Software Engineering 2024, 50, 2240–2253. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).