Preprint
Communication

This version is not peer-reviewed.

Ten Quick Tips for Biomedical Federated Learning

Submitted:

26 January 2026

Posted:

27 January 2026

You are already at the latest version

Abstract
Modern statistical and machine-learning techniques are effective at describing, testing hypotheses and making predictions from complex data. This effectiveness is strongly influenced by the volume and heterogeneity of available data. In many fields, including much of biomedicine, large centralized datasets are not available because of cost, privacy, regulatory or other restrictions. In these cases, smaller datasets are distributed across a large number of independent sites. Medical record data is a classic example of this challenge: the total number of patients may be large, but their records are distributed across many health systems and cannot easily be centralized. Federated learning (FL) is a machine learning paradigm that enables training and validation of a shared model in settings of decentralized data. FL can improve model accuracy and generalizability by increasing sample-size, but has trade-offs ranging from operational complexity to data-privacy risks to the potential to introduce unexpected imbalances in model accuracy. We outline ten tips for successfully and sustainably implementing FL for Biomedical applications, ensuring both ethical data governance and improved model performance in sensitive domains.
Keywords: 
;  

Introduction

Modern machine learning (ML) methods can leverage extremely large datasets to build truly insightful models. It is often the case that no single data source provides sufficient training information to optimize model accuracy, eliminate bias, and form generalizability. Centralizing and pooling data from multiple sources is a classic solution to this problem, and is sometimes called a “mega-analysis.” Still, centralizing data can be challenging for several reasons. First, it centralizes the costs of computing and storage. Second, it can require time-consuming execution of a large number of data sharing and collaboration agreements to manage a range of strategic and regulatory issues. Third, it creates a single point-of-failure for security issues. For these and other reasons, training ML models on pooled datasets is often infeasible.
The ability to aggregate hidden datasets is extremely relevant for biomedical research. Health care systems, research centers and controlled access government funded research datasets all sit in different computational facilities and all have different complications around accessing them. The scale of data required learning complex biological systems is extreme, but the individual datasets are tiny in comparison. In the US there were over 2 million new cancer cases in 2025[1], but they are spread over all 50 states, so that even the state with the most cases, California, only sees 10% of those patients. For rare diseases, or complex datasets that require a large number of samples to understand, no single hospital may see enough patients to collect a cohort sufficient for building accurate models.
Federated learning (FL) is a key component to address this problem. FL enables the training of a shared model on decentralized, unshared and uncoordinated datasets. Similar to a meta-analysis, FL enables machine learning/AI at scales that would be otherwise not feasible[2,3,4]. The most common FL implementations involve many sites/devices, each of which trains a model on its own local data. These local models are then combined at a central server and aggregated into a global model. Iteration and convergence strategies can then be applied, along with cross-device validation. In many ways, these types of FL are a form of ensemble learning, where models are trained on device-specific samples of the data. As a result of its decentralized nature, federated learning provides the capability to preserve the privacy of training data. The results of training, such as model weights, are shared while the data, such as patient records, are kept private. As a result, it is possible to train large scale models while respecting governance and legal frameworks like HIPAA. Beyond the immediate benefits of privacy preservation, federated computing also can provide a framework for accumulating data and computational resources required to build foundation models[5,6]. However, FL is not a privacy panacea[7]. It changes the locus of risk and accountability rather than eliminating it[4]. In biomedical settings, FL should be framed as a socio-technical intervention governed by necessity and proportionality: use FL when it is the least intrusive means to achieve a legitimate scientific aim, and pair it with concrete governance duties (joint controllership or controller–processor clarity, auditable access, and public-facing documentation).
FL has been successfully applied in a number of biomedical contexts, including COVID-19 outcome prediction[8] chest x-ray classification[9] and medical image segmentation[10]. Given the widespread interest in clinical data analysis, the appeal of federated learning is undeniably increasing. At the same time its practical initiation often presents significant complexities, a propensity for errors, and considerable technical difficulties[11].
The Bridge2AI program from the NIH’s Common Fund aims to propel biomedical research forward by setting the stage for widespread adoption of artificial intelligence (AI) that tackles complex biomedical challenges. To do so, four data generation projects work towards building a large multimodal and multi-institutional database of biomedical data and have had to develop ethical and regulatory frameworks to protect patient privacy while fuelling discoveries. During the 3 initial years of the project, avenues for data sharing limiting patient re-identification have been extensively studied as potential methods of dissemination for the bridge2ai datasets. In this paper, we have collected 10 tips to help build a FL network. One of the easiest ways to guide yourself through each of these tips is the table of questions that should be asked (Table 1). These tips aren’t about the technical aspects of a FL network, but rather the social issues that need to be navigated based on our experience with the Bridge2AI program. Following these tips should address common issues and maximize the short and long term success of the new FL initiative.

Tip 1: Assess the Benefits of FL Approaches

It is recognized that the performance of learning models, particularly when addressing complex problems, significantly improves with the incorporation of extensive datasets. This is particularly true for highly parameterized learners, like deep learning models (DL). The data security and privacy benefits of FL can encourage entities to make available samples that would otherwise be inaccessible for regulatory, ethical, or competitiveness reasons. Acquiring and harmonizing large datasets is often costly and time consuming. FL offers a pathway to scale data collection, distributing the efforts of hosting, cleaning, and maintaining a robust compute environment.
But FL is not the optimal solution under all scenarios. Harmonized, centralized repositories enjoy a number of advantages; for example, distributed data can lock in systematic differences in data characteristics across sites; this is a version of “batch effects”. These differences can be subtle and difficult to quantify in a distributed framework. Similarly, aggregating multiple “weak” learners can lead to poorer handling of rare cases than models fit on full datasets. In training of DL methods, samples are typically shuffled to ensure optimization occurs on a diverse set of data. The site-specific structure of data in FL can limit data-shuffling, creating order bias, where the optimization process struggles against a moving objective due to local batch effects and a lack of global data diversity[12]. In ML and RL analyses, the “long tail” effect can occur when there are a large number of rare cases, leading to a marginal improvement per sample as the dataset grows. If an event is only seen once or twice at each site, there is no way to aggregate these solitary events across different sites into a larger collective set to mitigate class imbalance. Methods for accounting for long tails can partially mitigate these issues, but are not commonly applied[13]. The limitations of FL can counterbalance the benefits of increasing sample size such that overall model performance may not increase even as federated sites are added.
Thus, the first step in any FL project is to determine whether it will improve the performance of models. A thorough evaluation of this process necessitates a rigorous cost-benefit analysis, weighing the prospective advantages of an augmented sample size against the potential magnitude of confounding batch effects.

Tip 2: Secure Early Engagement from IT & Security Teams

Once the potential benefits of FL are established, the subsequent phase involves close collaboration with both information technology (IT) and compliance departments. Within a complex health IT ecosystem, the deployment of an FL platform often requires the engagement of various IT teams, including those specializing in network infrastructure, compute infrastructure, devOps, security, and regulatory compliance, among others. The restrictions and complexities compound in environments holding protected data, like Health Insurance Portability Assurance Act (HIPAA) identifiable or limited/sensitive datasets[14]. Most FL setups involve multiple organizations, each with their own IT ecosystem, cultures and policies. Bringing these groups together at the outset, with clear support for IT leadership and in many cases an executive sponsor for the FL project, is necessary for success.
One example of this complexity is the inter-institutional communication required in a federated learning (FL) system. A common workflow involves an external-facing endpoint that receives requests, allocates compute and data resources, runs workflows, and finally transmits results to another endpoint in another institute. This often necessitates executing arbitrary or weakly-defined training and analysis code, as well as communicating through specific ports and to specific IP addresses.
For many organizations, such a workflow, in turn, requires exceptions to security rules and buy-in that typically goes far beyond a simple approval. The resultant expansion of an organization’s attack surface requires acceptance and active participation from appropriate security stakeholders to recognize the mission-driven benefits and proactively minimize risks. IT and security teams need to conduct thorough security reviews, document risk factors, and establish mitigation strategies on an ongoing basis. They are engaged in defining and documenting ingress and egress ports, implementing robust monitoring and logging mechanisms. Key considerations include:
  • Data encryption: Ensuring that data is encrypted both in transit and at rest across all participating sites.
  • Authentication and authorization: Implementing secure authentication and authorization protocols to control access to federated learning resources.
  • Audit trails and review: Establishing comprehensive audit trails to track data access and system activity.
  • Vulnerability scanning: Regularly scanning federated learning infrastructure for vulnerabilities and applying necessary patches.
  • Cost of computational time: AI/ML capable machines, namely those with large memory and GPU capabilities can be much more expensive to request, especially in a cloud infrastructure. Without proper moderation, a large number of scheduled jobs could become a form of Denial of Service (DOS) or Economic Denial of Sustainability (EDoS) driving project costs with excessive cloud fees.

Tip 3: Establish Project Governance

One motivation for building an FL network is the ability to learn from data without explicitly sharing it. This ability does not obviate the need for careful project governance. Despite the lowered risk of inadvertent sharing of controlled data, participation in an FL network, even one with just two sites, requires several issues to be agreed upon up-front. Typically legal agreements are needed to allow ready transfer of code, model structures, model weights, and even example datasets. These agreements should include shared understanding about intellectual property rights, publication expectations, expectations about intended use of the service and other issues. For instance, a key consideration is whether research focused on identifying individuals within a federated learning network would be deemed appropriate and acceptable to all parties. If a potential breach or privacy violation is identified, how should that be reported? Are there limitations on the expectations of how intensively sites employ one another’s compute and network resources? How should notifications about changes to a dataset be made (e.g., withdrawal of consent from a subset of study participants)? What about changes to the underlying FL code? Is there a process to expand the FL network to additional sites? If a site elects to withdraw, what rights and obligations survive their departure? A clear governance framework is needed to address these questions at the outset.
To manage Biomedical data, one should treat governance as the core deliverable[15]. This may start with creating a written Federation Agreement that covers: (a) role allocation under data-protection law (joint controllership vs. controller–processor), (b) a registered/controlled/open access tiering for code, metadata, and model artifacts, (c) a Data Access Committee Office (DACO) style workflow with turnaround service level agreement, and (d) explicit conditions for consent withdrawal, institutional withdrawal, and model incident reporting. One strategy would be to map a Responsible, Accountable, Consulted, Informed (RACI) matrix to each FL artifact (data schemas, local training code, update logs, global model checkpoints, audit dashboards) so accountability survives staff churn. Another good strategy would be to adopt something similar to a GA4GH-aligned “registered access” to enable rapid onboarding without eroding safeguards [https://www.ga4gh.org/]. Plan to document a breach/poisoning response plan, including model rollback and federated unlearning triggers. Academic and hospital Institutional Review Boards (IRB) and General Councils may still be quite unfamiliar with this new technology, so bringing supporting documentation to explain the technology and ethical risks to the governing entities of the institution will be crucial to proper implementation. In cases of studies involving prospective biomedical data collection including a consent process, the FL technology should be explained to participants using lay language describing the risks and benefits of this technology within the informed consent form (ICF) to gain their trust.

Tip 4: Define the First Version of Federation Platform

The landscape of FL platforms is currently fragmented, with numerous independent solutions available and under development (Table 2). This lack of standardization presents a significant challenge to interoperability and model sharing. The initial and often foundational decision in establishing a federated learning network entails a strategic selection of a platform and its associated compute resources. This decision is far more than a technical implementation: it has long-term implications for security, compliance, interoperability, and financial sustainability.
Compute resources also need to be specified proactively. Elements such as memory capacity, CPU manufacturer/model and speed, and storage capabilities are all critical factors. GPU availability is perhaps the most critical aspect of resource management for contemporary modeling approaches. GPU systems are usually architected with separated RAM allocations, bound to the GPU over a high-speed bus. The GPU-RAM capacity is smaller than system memory, with current top of the line cards having 80 GB of RAM, while computer servers will now typically come with over a 1TB of system RAM. The GPU-RAM capacity may have significant effects on the training performance for some classes of problems. The storage of 1 billion 32-bit floating point numbers requires almost 4GB of RAM, with additional space required for optimization and computation. Even “baseline” LLM models designed to run on consumer hardware can have 20 billion parameters, causing strain on systems with lower specifications.
Decisions about resource allocation are critical, not only for feasibility but because of the fiscal implications of a federated learning topology. External groups are sending requests to the data host, and those requests are carried out using the host computing facilities. This means the host will be the default billed entity for any costs incurred as part of the computation. These billing issues may be more discrete for on-premises facilities, but they manifest as electric bills and system queue availability. In cloud-based systems, this could translate into larger billed time on some of the most costly virtual machines (VMs) offered by the cloud provider. It may be prudent to identify usage limits, or develop agreements around reciprocal usage or charge-back policies. If not managed correctly, a federated endpoint could become a monetary “denial of service” endpoint allowing outside entities to impose unwanted financial costs on the host.

Tip 5: Start With an Example Dataset At Two Sites

Establishing and verifying IT protocols and governance is time-consuming, with many external dependencies. To create momentum, teams should plan on developing test scenarios around safe example datasets that involve minimally encumbered data (simulated federated learning). Public data is ideal for initial test scenarios, but any dataset that has minimal security and legal risk, and allows inspection of a unified dataset from both sides can be useful in this stage.
In some cases, development of this example dataset can be difficult, especially when scale is valued. In such cases, synthetic datasets may be ideal. For example, to test electronic health records (EHRs), there is Synthea[16]; SyntheVAEiser[17] for transcriptomic data; and other methods for H7E image generation[18]. Other reviews have been written on the topic[19] and can help guide researchers to tools that can be used to create new datasets for testing.
It is also useful if the example dataset contains actual signals and supports multiple question types. Statistical tests should be built into the example dataset, such as batch connected alterations to the data. These can include shifts in the population distributions, introducing missingness, and embedding noise into different variables at different sites and including “N/A” (not available) or “NaN” (not a number) values in only one of the datasets.
By having more than one institution work on this problem, it provides an opportunity to start validating data standards and testing ideas around the problem space. But this phase should also be treated as a “warm up,” including a large number of institutions early on where the datasets are meaningful. This aspect requires both an understanding of why the data is acquired, and how it is curated. Endeavors representative of the University of California Data Health Warehouse and the NIH Bridge2AI Data Generation Projects[20] are notable here, appreciating how specific efforts are able to collect data within specific scopes.
This rule also be extended to keep the FL system as simple as possible. If a certain aspect is not going to be needed by at least two sites, it can be scheduled for a later time. This principle keeps feature creep at control - which is especially critical in the current rapidly changing environment.

Tip 6: Stress Test with Models of Escalating Complexity

Example datasets are intended for “beta testing” before full security review and governance protocols are established. Such a process allows them to be used as a technical exercise, allowing the team to flesh out issues surrounding the governance and technical platform. But their simplicity also means they are less useful for scientific discovery. In a synthetic dataset, any signal needs to be designed and baked into the data. So if there is no design done to specifically code signals into data, machine learning models trained on the data will be unable to detect and learn any useful patterns. Developing complex and realistic signals in a synthetic dataset can be an increasing time sink, and the project team should determine the amount of time that should be invested in improvising more complex synthetic datasets.
As data and model size and complexity increase, many of the nuanced technical challenges inherent in machine learning become more tactically significant. To facilitate preliminary analysis and identify such potential issues early on, it is prudent to begin with simpler models. Starting with easily interpretable models like simple linear models and decision trees, which allow for a faster iteration cycle and helps pinpoint problems with the platform, data generation, data cleaning, and data formatting stages. As initial tests are successful, move to slightly more complex models like random forests and single-layer neural networks, before progressing to large scale DL. While extremely large models may ultimately be needed to achieve top-tier performance, they also present increased risk of instability, heightened sensitivity to parameter tuning, and potential for unexpected failures. Once the foundational testing loop has been thoroughly validated with simpler models, the project can then progressively introduce models of increasing complexity.
Modern AI/ML systems, particularly in federated learning environments, face significant challenges with robustness and stability. This is largely due to their distributed and sectioned nature, where artifacts like data, models, and software are updated independently. To address this, it’s crucial to implement a continuous integration (CI) testing philosophy. All tests, once developed, should be isolated to reduce compute costs and integrated into a testing suite that is run frequently. This approach ensures all stakeholders have transparent access to the results, allowing for proactive identification and resolution of issues.

Tip 7: Pressure Test Security & Governance

Even as the example dataset is identified and consideration is given to the kinds of analyses to be supported, it is critical to also recognize how assumptions made around security and governance may fail. While many existing platforms incorporate security measures, these are often designed to defend against external, out-of-band attacks. A more rigorous evaluation is required, encompassing “white hat” intrusion testing and vulnerability assessments to identify and address potential weaknesses before broader researcher access is granted.
Federated learning presents unique security challenges. A critical concern is the potential for protected information to be inadvertently leaked within model weights[21,22]. Efforts should be made to reconstruct or infer protected fields as a form of ‘white hat’ hacking experiment[23,24]. Surveys on the strategies for securing federated learning systems have been performed[24] and outline strategies to minimize the risks of sensitive data leakage.
The system must be safeguarded against “model poisoning”[25], where malicious or even unintentionally biased participants contribute optimized weights that skew the aggregate model. This results in models that lose confidence on calls, become less generalizable, and have worse overall performance. Batch effects result from genuine data incongruities or unforeseen batch effects, highlighting the importance of robust validation and anomaly detection mechanisms. While not the result of malicious intent, batch effects can still have similar effects as model poisoning. Similar strategies of monitoring overall performance changes and correlating to dataset providers and help identify the origins of spurious training intervals. By way of illustration, different types of batch effects can become apparent as the real datasets become available per site, including differences in data distributions (e.g., Site X has 65% of cases, Site Y has 25%, etc. each with different case mixes) that propagate imbalanced model learning due to underlying site-specific biases; questions of fairness in FL thus arise in different ways that should be identified earlier in the process[26,27,28,29,30].
As projects and cohorts grow, governance should draft plans for fairness and address the possibilities of withdrawal and unlearning. Security tests should be paired with fairness and reversibility tests. Another strategy is to build a federated fairness audit that (i) computes per-site and per-subgroup performance with confidence intervals, (ii) flags contribution–benefit gaps (sites contributing rare cases but receiving poor local performance), and (iii) halts aggregation when fairness deltas exceed pre-set bounds. This report can be published with the audit summaries to all nodes. To ensure compliance with potential issues with IRBs there should be a plan to operationalize federated unlearning: this would involve defining triggers (participant withdrawal, consent revocation, data error), select a method (client-level or update-level unlearning), and test rollback to pre-event checkpoints. Finally, calibrate differential privacy: document the privacy budget and expected utility loss at node-level so privacy protections do not disproportionately degrade models for under-resourced sites.
Security and governance testing should not be treated as a one time event at the beginning of the project. These elements should be revisited at common intervals, including period security inspections, project review and software updates to ensure security related patches as incorporated into the federation platform. Logging from all requested jobs can be reviewed to ensure that usage patterns follow expectations laid out by project descriptions.

Tip 8: Lock Data & Model Standards

Software development is continual, there are always tweaks and fixes to be made. But for the purposes of federated learning it is important to synchronize standards, model code, and datasets before proceeding with large scale learning processes. Locking these elements before starting a cycle of federated learning is incredibly important for both debugging and provenance. Debugging issues can be an incredibly difficult process even in the best case scenarios, and multi-site code deployment only complicates these issues, especially if the different sites are running different versions of code. Model provenance is also only really possible if the model and data have been frozen for the duration of the training cycle. If a training regiment repeatedly visits a site to adjust weights, but the datasets shift between training epochs, reproducing the final results will become difficult if not impossible.
To combat these issues and coordinate research efforts, a federated environment should establish a versioning system for model code and datasets. Options include using semantic versioning[31], tags generated from checksumming file and code manifests (e.g., Secure Hash Algorithm), or a date-based system. When a model is trained, the model code and dataset version numbers should be recorded as part of the training process’s metadata to enable reproducibility in the future. If external pre-trained models are used they should also be version controlled. These efforts will help to identify issues of changing model performance over time, and help identify if updates to datasets have had unintended side effects.

Tip 9: Perform Descriptive Analyses & Verify Offline

As the project progresses from freely accessible example datasets to controlled access, real-world datasets, visibility into training results will inevitably decrease. This transition will expose true batch effects and data biases not apparent with the initial testing data. Critically, access control rules may now restrict individual access to both component datasets, further complicating debugging and result evaluation. To proactively address these challenges, project teams must establish robust descriptive analysis methods and offline verification processes. These methods should prioritize anonymized data exploration and summarization, allowing teams to identify and investigate trends in the training data without compromising the privacy and security of the underlying samples. Specifically, teams should explore techniques for generating aggregate statistics, visualizations, and summary reports that can be shared and discussed collaboratively. Additionally, the experience of the team members of each site should be collected periodically: 1) to understand “pain points” from the effort, addressing the challenges in deployment; and 2) to evaluate the impact of the learned model given their site’s input into helping to create it (and thus ongoing participation in the endeavor). Not only are such insights important for future FL-based projects, but to validate the cost and utility of the shared effort.

Tip 10: Document Document Document

A federated learning framework is a consortium: no one person can drive the entirety of the project. The “bus factor,” a term from software engineering, quantifies the systemic risk to a project should key individuals become unavailable, representing the number of people whose loss would cause the project to fail due to critical knowledge or skill gaps. Even for complex projects, this number can be surprisingly low. For FL, loss of skill sets or domain knowledge at any of the sites could remove it from the federation and affect the entire group. One of the most effective ways to combat this systemic risk is to ensure the details of the project can be proper documentation. A partial list of items that should be documented include:
  • Initial rationale for deploying FL
  • Compliance and security buy-in and evaluations
  • Governance documents
  • Procedure for amending governance documents
  • Procedure for adding or removing a data site from the federation
  • Procedure for adding or removing an analysis site from the federation
  • Code of conduct for sharing of compute resources
  • Code and system documentation standards
  • Emergency procedures for breaches or other unexpected events
  • Security audits
  • Backup protocols
  • Code samples to scaffold new analyses

Conclusions

The vast majority of FL literature focuses on its mathematical foundations and algorithmic optimization; practical application of FL requires navigating complex social, administrative and governance landscapes. Similarly, many examples of FL system deployment focus on software; in the real-world, the data cleaning, formatting, and validation is often rate-limiting. The intrinsic value, privacy, size, and legal considerations surrounding healthcare-derived data make FL highly appealing. To avoid bottlenecks that stall or terminate efforts, a holistic approach that proactively jointly addresses technical and non-technical issues is needed.

Acknowledgments

The authors gratefully acknowledge the extensive insights from our long-standing conversations with colleagues in the Bridge2AI and GA4GH consortia. This work was supported by the NIH through awards U54HG012517, U24CA248265, U2CCA271894, P30CA016042. S.M. was supported by a grant of the Ministry of Research, Innovation and Digitization under Romania’s National Recovery and Resilience Plan - Funded by EU - NextGenerationEU” program, project “Artificial intelligence-powered personalized health and genomics libraries for the analysis of long-term effects in COVID-19 patients (AI-PHGL-COVID)” number 760073/23.05.2023, code 285/30.11.2022, within Pillar III, Component C9, Investment 81.

References

  1. Siegel, R.L.; Kratzer, T.B.; Giaquinto, A.N.; Sung, H.; Jemal, A. Cancer statistics, 2025. CA Cancer J Clin. 2025, 75(1), 10–45. [Google Scholar] [CrossRef] [PubMed]
  2. Yurdem, B.; Kuzlu, M.; Gullu, M.K.; Catak, F.O.; Tabassum, M. Federated learning: Overview, strategies, applications, tools and future directions. Heliyon 2024, 10(19), e38137. [Google Scholar] [CrossRef]
  3. Teo, Z.L.; Jin, L.; Li, S.; Miao, D.; Zhang, X.; Ng, W.Y.; et al. Federated machine learning in healthcare: A systematic review on clinical applications and technical architecture. Cell Rep Med. 2024, 5(2), 101419. [Google Scholar] [CrossRef]
  4. Rieke, N.; Hancox, J.; Li, W.; Milletarì, F.; Roth, H.R.; Albarqouni, S.; et al. The future of digital health with federated learning. NPJ Digit Med 2020, 3(1), 119. [Google Scholar] [CrossRef]
  5. Zhuang, W.; Chen, C.; Li, J.; Chen, C.; Jin, Y.; Lyu, L. When Foundation Model meets Federated Learning: Motivations, challenges, and future directions [Internet]. arXiv [cs.LG]. 2023. Available online: http://arxiv.org/abs/2306.15546.
  6. Chen, J.; Guo, S.; Qi, Q.; Hao, J.; Guo, S.; Qiu, X. Enabling foundation models: A distributed collaboration framework based on graph federated learning. IEEE Trans Serv Comput., 2024; pp. 1–14. [Google Scholar]
  7. Brauneck, A.; Schmalhorst, L.; Kazemi Majdabadi, M.M.; Bakhtiari, M.; Völker, U.; Baumbach, J.; et al. Federated machine learning, privacy-enhancing technologies, and data protection laws in medical research: Scoping review. J Med Internet Res. 2023, 25(1), e41588. [Google Scholar] [CrossRef]
  8. Dayan, I.; Roth, H.R.; Zhong, A.; Harouni, A.; Gentili, A.; Abidin, A.Z.; et al. Federated learning for predicting clinical outcomes in patients with COVID-19. Nat Med. 2021, 27(10), 1735–1743. [Google Scholar] [CrossRef] [PubMed]
  9. Pan, Z.; Wang, H.; Wan, J.; Zhang, L.; Huang, J.; Shen, Y. Efficient federated learning for pediatric pneumonia on chest X-ray classification. Sci Rep. 2024, 14(1), 23272. [Google Scholar] [CrossRef]
  10. Schutte, P.; Corbetta, V.; Beets-Tan, R.; Silva, W. FedGS: Federated gradient scaling for heterogeneous medical image segmentation. In Lecture Notes in Computer Science; Springer Nature Switzerland: Cham; Lecture notes in computer science, 2025; pp. 246–255. [Google Scholar]
  11. Wen, J.; Zhang, Z.; Lan, Y.; Cui, Z.; Cai, J.; Zhang, W. A survey on federated learning: challenges and applications. Int J Mach Learn Cybern. 2023, 14(2), 513–535. [Google Scholar] [CrossRef]
  12. Zhang, T.; Chen, M.; Bui, A.A.T. Diagnostic prediction with sequence-of-sets representation learning for clinical events. Artif Intell Med Conf Artif Intell Med 2020, 12299, 348–358. [Google Scholar]
  13. Li, K.; Li, Y.; Zhang, J.; Liu, X.; Ma, Z. Federated deep long-tailed learning: A survey. Neurocomputing 2024, 595(127906), 127906. [Google Scholar] [CrossRef]
  14. Pati, S.; Kumar, S.; Varma, A.; Edwards, B.; Lu, C.; Qu, L.; et al. Privacy preservation for federated learning in health care. Patterns (N Y) 2024, 5(7), 100974. [Google Scholar] [CrossRef]
  15. Eden, R.; Chukwudi, I.; Bain, C.; Barbieri, S.; Callaway, L.; de Jersey, S.; et al. A scoping review of the governance of federated learning in healthcare. NPJ Digit Med. 2025, 8(1), 427. [Google Scholar] [CrossRef]
  16. Walonoski, J.; Kramer, M.; Nichols, J.; Quina, A.; Moesel, C.; Hall, D.; et al. Synthea: An approach, method, and software mechanism for generating synthetic patients and the synthetic electronic health care record. J Am Med Inform Assoc. 2018, 25(3), 230–238. [Google Scholar] [CrossRef]
  17. Karlberg, B.; Kirchgaessner, R.; Lee, J.; Peterkort, M.; Beckman, L.; Goecks, J.; et al. SyntheVAEiser: augmenting traditional machine learning methods with VAE-based gene expression sample generation for improved cancer subtype predictions. Genome Biol. 2024, 25(1), 309. [Google Scholar] [CrossRef]
  18. Pozzi, M.; Noei, S.; Robbi, E.; Cima, L.; Moroni, M.; Munari, E.; et al. Generating and evaluating synthetic data in digital pathology through diffusion models. Sci Rep. 2024, 14(1), 28435. [Google Scholar] [CrossRef] [PubMed]
  19. Pezoulas, V.C.; Zaridis, D.I.; Mylona, E.; Androutsos, C.; Apostolidis, K.; Tachos, N.S.; et al. Synthetic data generation methods in healthcare: A review on open-source tools and methods. Comput Struct Biotechnol J 2024, 23, 2892–2910. [Google Scholar] [CrossRef]
  20. Rincon, J.; Pelletier, A.R.; Gilliland, D.; Wang, W.; Wang, D.; Sankar, B.S.; et al. Bridge2AI: Building A cross-disciplinary curriculum towards AI-enhanced biomedical and clinical care [Internet]. arXiv [cs.CY]. 2025. Available online: http://arxiv.org/abs/2505.14757.
  21. Carlini, N.; Hayes, J.; Nasr, M.; Jagielski, M.; Sehwag, V.; Tramèr, F.; et al. Extracting training data from diffusion models [Internet]. arXiv [cs.CR]. 2023. Available online: http://arxiv.org/abs/2301.13188.
  22. Haim, N.; Vardi, G.; Yehudai, G.; Shamir, O.; Irani, M. Reconstructing training data from trained neural networks. In Neural Inf Process Syst.; Jun 15;abs/2206.07758:22911–24; Koyejo, S, Mohamed, S, Agarwal, A, Belgrave, D, Cho, K, Oh, A, Eds.; 15 Jun 2022. [Google Scholar]
  23. Dalziel, H. US Cybersecurity Jobs. In How to Become an Ethical Hacker and Penetration Tester; Elsevier, 2015; pp. 1–8. [Google Scholar]
  24. Hu, K.; Gong, S.; Zhang, Q.; Seng, C.; Xia, M.; Jiang, S. An overview of implementing security and privacy in federated learning. Artif Intell Rev [Internet] Available from. 2024, 57(8). [Google Scholar] [CrossRef]
  25. Xia, G.; Chen, J.; Yu, C.; Ma, J. Poisoning attacks in federated learning: A survey. IEEE Access. 2023, 11, 10708–22. [Google Scholar] [CrossRef]
  26. Zhang, F.; Shuai, Z.; Kuang, K.; Wu, F.; Zhuang, Y.; Xiao, J. Unified fair federated learning for digital healthcare. Patterns (N Y) 2024, 5(1), 100907. [Google Scholar] [CrossRef]
  27. Li, S.; Wu, Q.; Zhou, D.; Li, X.; Miao, D.; Hong, C.; et al. FairFML: fair federated machine learning with a case study on reducing gender disparities in cardiac arrest outcome prediction. Npj Health Syst. 2025, 2(1), 1–8. [Google Scholar] [CrossRef]
  28. Xing, H.; Sun, R.; Ren, J.; Wei, J.; Feng, C.M.; Ding, X.; et al. Achieving flexible fairness metrics in federated medical imaging. Nat Commun. 2025, 16(1), 3342. [Google Scholar] [CrossRef] [PubMed]
  29. Li, X.; Zhao, S.; Chen, C.; Zheng, Z. Heterogeneity-aware fair federated learning. Inf Sci (Ny) 2023, 619, 968–86. [Google Scholar] [CrossRef]
  30. Zhou, Z.; Chu, L.; Liu, C.; Wang, L.; Pei, J.; Zhang, Y. Towards fair federated learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining [Internet]; Available from; ACM: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  31. Preston-Werner, T. Semantic Versioning. [cited 2025 Oct 13]. Semantic Versioning 2.0.0. Available online: https://semver.org/.
Table 1. Questions to ask.
Table 1. Questions to ask.
Tip 1: Assess the Benefits of FL Approaches
  • Will federated learning help build a more accurate model?
  • Is there extreme class imbalance in the datasets, and can it be addressed?
Tip 2: Secure Early Engagement from IT & Security Teams
  • Does data in the training set fall under HIPAA or other protection?
  • What systems will be used for computation?
  • What security audits will need to be done to be in compliance with institutional standards?
  • Who will be responsible for which aspects of security and compliance?
  • Is there a level of security that all participant sites agree to that meets a common acceptable baseline?
Tip 3: Establish Project Governance
  • Who are the data owners and what rules are already established regarding its usage?
  • How will issues in governance be reported, including data leakage?
  • How will requests to remove data or add sites be handled?
  • How will intellectual property (IP) from the model be managed?
Tip 4: Define the First Version of Federation Platform
  • Which platform will be used?
  • How will billing for compute/storage usage be handled?
Tip 5: Start With an Example Dataset At Two Sites
  • Is there an existing synthetic or open access dataset that can be used for prototyping?
Tip 6: Stress Test with Models of Escalating Complexity
  • What is the largest model our current deployment could accommodate?
Tip 7: Pressure Test Security & Governance
  • If a bad actor was able to send models to the system, is there a way they could game the system to retrieve identifiable data?
Tip 8: Lock Data & Model Standards
  • What would the version 1.0 of the system and datasets look like? What would version 2.0 be?
Tip 9: Perform Descriptive Analyses & Verify Offline
  • Do the results match expectations?
  • Do the stakeholders perceive any gain from the effort required for setting up the FL framework?
Tip 10: Document Document Document
  • If a new individual had to use or maintain the system, how would they learn?
Table 2. Federated Learning Frameworks Review.
Table 2. Federated Learning Frameworks Review.
Name Framework Developer Homepage
OpenFL Tensorflow, PyTorch Intel https://openfl.readthedocs.io
TensorFlow Federated TensorFlow Google https://www.tensorflow.org/federated
APPFL PyTorch DoE https://appfl.readthedocs.io/
NVFlare Tensorflow, PyTorch NVidia https://nvflare.readthedocs.io
Flower Tensorflow, PyTorch Startup https://flower.dev
FATE Tencent https://fate.fedai.org/
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated