Preprint
Article

This version is not peer-reviewed.

FAFedZO: Faster Zero-order Adaptive Federated Learning Algorithm

A peer-reviewed article of this preprint also exists.

Submitted:

22 January 2025

Posted:

22 January 2025

You are already at the latest version

Abstract
Federated Learning represents a newly emerging methodology in the field of machine learning, it enables distributed agents to collaboratively learn a centralized model without sharing their raw data. Some scholars have already proposed many first-order algorithms and second-order algorithms for federated learning to reduce communication costs and speed up convergence. However, these algorithms generally rely on gradient or Hessian information, and we find it difficult to solve such federated optimization problems when the analytical expression of the loss function is not available, that is, when gradient information is not available. Therefore, we employed derivative-free federated zeroth-order optimization in this paper which not rely on specific gradient information, but instead utilizes the changes in function values or model outputs to estimate the optimization direction. Furthermore, to enhance the performance of derivative-free zeroth-order optimization, we propose an effective adaptive algorithm that can dynamically adjust the learning rate and other hyperparameters based on the performance during the optimization process, aiming to accelerate convergence. We rigorously analyze the convergence of our approach, and the experimental findings demonstrate our method indeed can achieve faster convergence speed on the MNIST and CIFAR-10 datasets in cases where gradient information is not available.
Keywords: 
;  ;  ;  

1. Introduction

With the rapid development of big data and artificial intelligence technologies, machine learning has become one of the key technologies in various fields. However, in practical applications, issues of data privacy and security have become a non-negligible challenge. In traditional machine learning, data usually needs to be sent to data centers for model training, but in the process of transmission, it faces the problems of data leakage and privacy protection. To address this issue, federated learning(FL) [1], an emerging distributed machine learning technology, has emerged. It achieves collaborative learning among multiple data sources, and while protecting data privacy by exchanging model parameters instead of raw data. Currently, FL is utilized across diverse domains such as autonomous driving [2], personalized recommendation systems [3], and medical informatics [4], among others.
However, existing federated learning methods also have some obvious shortcomings. For instance, some federated learning methods are unable to effectively prevent privacy leaks when faced with complex attack techniques, and the convergence speed of current federated learning algorithms remains a bottleneck. Additionally, designing more efficient optimization algorithms tailored to different types of models and tasks is also an urgent issue that needs to be addressed. Researching how to enhance the privacy protection capabilities of federated learning is of great significance in an environment where data security and privacy protection are increasingly critical. This not only safeguards users’ personal information but also promotes the widespread application of federated learning technology across various domains. In practical applications, federated learning needs to meet the demands of various scenarios. By investigating the issue of slow convergence speed, it can better satisfy these practical application needs and enhance the practicality and value of federated learning.
In recent years, the federated optimization for model training has increasingly captured the interest of both academic researchers and industry professionals. To achieve fast convergence rates and reduce communication loads, a wide range of algorithms have been proposed, including first-order methods (such as FedAvg [1], FedFOR [5], FedFomo [6]) and second-order methods (such as Fed-Sophia [7]). Most of these algorithms rely on gradient or Hessian information, and these methods are often only applicable to differentiable functions. However, when the objective function is difficult to compute gradients or the gradient computation cost is excessively high, these algorithms will not be able to solve such problems, such as in the case of black-box functions and in federated hyperparameter tuning [8], so we must look for other solutions. To alleviate the reliance on gradient or Hessian information, the federated zeroth-order optimization (FedZO) algorithm is proposed in [9] through the integration of zeroth-order optimization into the federated learning. The characteristic of derivative-free zeroth-order optimization is that it only queries the objective function values to tackle the problems of federated optimization. It approximates the full gradient and stochastic gradient of the function using function values, without using gradient and Hessian information. In each round of communication, several local updates are performed utilizes a two-point stochastic gradient estimator. Some recent studies such as [10] and [11] have also employed the zeroth-order idea. [10] proposed FedDisco algorithm, leveraging zeroth-order optimization techniques reduces communication overhead significantly. [11] designed a federated zeroth-order algorithm FedZeN, it focused on convex optimization and estimate the curvature of the global objective.
However, existing zeroth-order methods rely solely on function value information, lacking the utilization of gradient and other relevant information. So in complex federated learning scenarios, this method may result in the model failing to fully leverage the feature information of the data, leading to the need for more iterations to achieve convergence during the optimization process. This significantly increases computational time and affects model performance and convergence speed. If the convergence speed of zeroth-order optimization can be accelerated, it can play a better role in scenarios with limited resources such as large-scale distributed optimization and edge computing, improving resource utilization efficiency and thereby promoting the widespread deployment of optimization technologies in resource-sensitive applications.
The emergence of adaptive methods has effectively alleviated the problem of slow convergence. In deep learning, an excessively large learning rate may lead to unstable model training, resulting in overfitting or underfitting. Besides, traditional fixed learning rate methods often struggle to adapt to complex and diverse training data and model structures, potentially leading to slow convergence speed of the model. The adaptive method can dynamically adjust the learning rate based on feedback from the training process, thereby accelerating the convergence of the model. In addition, compared with stochastic gradient descent, this method can also escape saddle points more quickly [12]. Therefore, introducing it in practice represents a crucial approach to enhancing the performance of federated learning algorithms.
However, improper design of adaptive federated learning methods may lead to convergence issues [13]. Reddi et al. [14] first proposed a federated version of the adaptive optimizer, among them are FedAdagrad and FedYogi. However, their analysis is only valid under the condition that β 1 = 0 (where β 1 is the decay parameter), which cannot leverage the advantages of momentum. In the algorithm FedCAMS [15], it provides a complete proof, but it does not improve the convergence speed. Moreover, they usually require a global learning rate for initialization and adjustment, and the selection of this global learning rate may affect the performance of the algorithm. Existing adaptive methods still have shortcomings such as lagging response to complex environmental changes and high consumption of computational resources. Our goal is to design adaptive algorithms tailored to various real-world scenarios, characterized by efficiency, precision, and flexibility, so as to rapidly adapt to environmental changes, optimize resource allocation, and provide robust support for development across various fields.
Based on the above, we combine the gradient-free optimization with adaptive methods, aiming to leverage the advantages of both to solve the problem of unavailable gradient information of the objective function in a finite-sum optimization problem, while achieving fast convergence and effectively improving the training efficiency and performance of the model.
Contributions The contributions of this paper are summarized as follows:
  • By combining the zero-order optimization and adaptive gradient method, we have proposed a novel faster zero-order adaptive federated learning algorithm, called FAFedZO, it can eliminate the reliance on gradient information and accelerate convergence at the same time.
  • We conducted a theoretical analysis of the proposed zeroth-order adaptive algorithm and provided a convergence analysis framework under some mild assumptions, demonstrating its convergence. Additionally, we have analyzed the computational complexity and convergence rate of the algorithm.
  • Extensive comparative experimental results under both IID and non-IID settings have confirmed the efficacy of our proposed FAFedZO algorithm on the MNIST and CIFAR-10 datasets. when gradient information is not available, it can achieve faster convergence speed.
The structure of the rest of this paper is outlined below. Section 2 provides a summary of the related work. In Section 3, we present the formulation of the federated optimization problem and the algorithm framework of FAFedZO. Section 4 provides the convergence analysis of FAFedZO. Section 5 presents the outcomes of our experiments. Section 6 discusses the beneficial applications of our method in the real world. The paper concludes with Section 7.

2. Related Work

Federated Learning
There have been numerous studies on federated learning, and the pioneering work on federated learning began with [1], where proposed the algorithm called FedAvg. After FedAvg, numerous additional first-order schemes emerged, such as FedNova [16], FedProx [17], SCAFFOLD [18], FedSplit [19], and FedPD [20]. Among them, SCAFFOLD employs control variates to rectify "client drift" while maintaining the same sampling and communication complexity as FedAvg. FedProx introduced a penalty-based approach, which can reduced communication complexity to O ( ϵ 2 ) . To further minimize communication costs, various second-order optimization methods have been introduced, including GIANT [21] and FedDANE [22]. Additionally, there are also some FL algorithms based on momentum, including [23,24,25]. The work presented in [23] proposes a momentum fusion approach for synchronizing the server and local momentum buffers, however, it does not aim to decrease complexity. [24] proposed a momentum-based global update algorithm, Fed-GLOMO, which reduces variance on the server side using variance reduction techniques. [25] proposed the STEM algorithm, it employs momentum-assisted stochastic gradient directions for updates at both worker nodes and the central server.
Recent research [26,27] has also focused on the application of federated learning in satellite–terrestrial integrated networks. [26] proposes a federated split learning framework, which aims to leverage the computing capabilities of satellite and terrestrial networks for collaborative processing of sequential data while protecting data privacy. [27] presents an intrusion detection method that combines federated learning with conditional generative adversarial networks to address the challenges of data privacy and imbalanced datasets. However, there are still numerous federated optimization problems in reality that are challenging to solve, for instance when gradient information is unavailable or costly to acquire. Therefore, the study of gradient-free zeroth-order optimization is essential.
Zero-order Optimization
Early literatures that began to use the zero-order idea for estimation include [28,29,30]. Specifically, In [28], the authors developed a distributed zero-order algorithm utilizing gradient tracking techniques. In [29], the author provides the first generalization error analysis for black-box learning via derivative-free optimization, and demonstrates that under the assumption of Lipschitz and smoothing (unknown) losses, the ZoSS method attains a comparable generalization error boundary to Stochastic Gradient Descent (SGD). [30] proposed and analyzed zero-order stochastic approximation algorithms for non-convex and convex objective function, focusing on solving the problems of constrained optimization and high-dimensional settings. Recent key research achievements have also applied zero-order methods to various fields, such as [9,10,11] and [31]. In [9], a derivative-free algorithm FedZO is proposed. It is proven that this algorithm achieves a linear increase in speed in relation to the number of devices involved and local iteration times under non-convex settings. Under the framework of cross-device federated learning, [31] introduces a dual-communication zeroth-order method, which is the first technique to incorporate wireless channels into the algorithm. It employs a single-point gradient estimator, replacing long vectors with scalar values for communication while leveraging the characteristics of wireless channels. This represents a significant new achievement in the field of zeroth-order optimization, achieving a convergence rate of O ( 1 K 3 ) in non-convex settings. However, due to the nature of zeroth-order optimization, it can lead to slower convergence speeds. Therefore, an adaptive method is introduced below.
Adaptive Methods
Early proposed adaptive algorithms include [32,33]. The Adam method in [32] introduces decay coefficients and combines momentum optimization to better adapt to non-stationary data and large-scale datasets, thereby accelerating model convergence. The AdaGrad method in [33] can adaptively adjust the learning rate of each parameter that can makes sparse features to obtain larger learning rates, while frequently occurring features obtain smaller learning rates. Then researchers have extended these algorithms to the context of federated learning, like recent studies such as [32-35]. In [34], the authors designed an Adaptive Local Iteration Differential Privacy Federated Learning algorithm (ALI-DPFL) and demonstrated its superiority in resource-constrained scenarios. The adaptive methods in [35] and [36] effectively mitigate the issue of non-IID data, achieving critical progress in enabling better performance when dealing with non-IID data. In [35] proposed an adaptive fairness federated aggregation algorithm Ada-FFL, which dynamically adjusts fairness coefficients based on local model updates to ensure both convergence performance of the global model and fairness among clients. [36] introduced a novel framework called FedARF, aimed at enhancing federated learning performance through adaptively reconstructing local features during the training process. It adopts an adaptive feature fusion strategy, enabling the model to better adapt to the data distribution of each client, thereby accelerating the convergence speed on non-IID data. [37] introduced a novel federated learning algorithm named AFedAvg, which significantly reduced communication costs and accelerated convergence speed by combining adaptive communication frequency and gradient sparsity techniques. This algorithm dynamically selects the communication frequency for the next round based on the number of sparse parameters. The results show that it can achieve a communication compression ratio ranging from 2.4 times to 23.1 times, which marks a significant step forward in mitigating the obstacles posed by high communication volume to artificial intelligence. However, none of these studies have investigated the integration of adaptive methods within the context of zero-order optimization. Hence, our article addresses this by combining zero-order optimization with adaptive approaches to conduct our discussion.

3. Problem Formulation and Algorithm Design

In this part, we will introduce the federated optimization problem and the design of the zero-order adaptive federated optimization algorithm FAFedZO.

3.1. Federated Optimization Problem Formulation

We consider a federated learning task involving a central server and Q edge devices with an index of { 1 , 2 , . . . Q } . The central server aims to facilitate collaboration among these devices to address a specific optimization problem
min Ξ R d f ( Ξ ) 1 Q i = 1 Q f i ( Ξ ) ,
and
f i ( Ξ ) E ϑ i D i [ F i ( Ξ , ϑ i ) ] ,
here, Ξ R d represents a d-dimensional model parameter. For each edge device i, f i ( Ξ ) denotes its local loss function, while f ( Ξ ) stands for the global loss function. In formula (1), f i ( Ξ ) evaluates the anticipated risk on the data distribution D i on the edge device i, which is presented in formula (2). ϑ i D i denotes the random variable ϑ i is uniformly drawn from the distribution D i , and F i ( Ξ , ϑ i ) denotes the loss function of ϑ i at the parameter Ξ .

3.2. Algorithm Design of FAFedZO

We explored how to design a method that combines adaptive gradient with zero-order optimization in federated learning.
Firstly, we expand our algorithm description within the context of the FedAvg framework. We are focused here on solving problem (1) through zeroth-order optimization methods and propose a novel zero-order fast adaptive FL method (FAFedZO) employs a shared adaptive learning rate to address the issues. In particular, Algorithm 1 outlines the specifics of our FAFedZO method.
At the beginning, input the parameters and perform initialize. For all i, compute Ξ 1 , i = Ξ 0 , i η 0 n 0 , i , which represents the first model update based on the initial parameters and the initial gradient estimates.
Then, for each edge device i, perform the following steps: First, extract a mini-batch B t , i of size b from the local dataset D i . Next, compute the stochastic gradient estimates g ^ t , i and g ^ t 1 , i for the current model parameters Ξ t , i based on the mini-batch B t , i . Here, we elaborate on the specific method for estimating the gradients.
To address the issue of unavailable gradient information and reduce the frequency of model exchange, we achieve this by using gradient estimators and performing stochastic zero-order updates in each communication round. In particular, at the t-th round, edge device i calculates a two-point stochastic gradient estimator [28] as outlined below
˜ v μ F i ( Ξ t , i , ϑ t , i ) = d v t μ t F i ( Ξ t , i + μ t v t , ϑ t , i ) F i ( Ξ t , i , ϑ t , i ) ,
where Ξ t , i denote the local model of edge device i, while ϑ t , i signifies a random variable drawn by edge device i according to its local data distribution D i during the t-th round. v t represents a randomly chosen d-dimensional direction, uniformly sampled from the unit sphere S d , while μ t stands for a positive step size.
When computing gradient estimates, we do not directly use precise gradient information but instead approximate the gradient through random sampling and estimation. This zero-order optimization approach does not require the exact calculation of function derivatives, rather, it optimizes based on sampling and estimation of function values.
Afterward, we update local models according to the following stochastic zeroth-order update method
Ξ t + 1 , i = Ξ t , i η t ˜ v μ F i ( Ξ t , i , ϑ t , i ) , t = 1 , 2 , , T
where η t denotes the learning rate.
In step 9 of Algorithm 1, we use n t , i to update the model, where n t , i is the momentum-based variance reduced gradient estimator, is the weighted sum of the current gradient estimate and previous gradient estimates, with the weights determined by χ t . Its definition is as follows
n t , i = ˜ v μ F i ( Ξ t , i ; ϑ t , i ) + ( 1 χ t ) ( n t 1 , i ˜ v μ F i ( Ξ t 1 , i ; ϑ t , i ) ) ,
where the hyperparameter χ t ( 0 , 1 ) , representing the decay factor.
In step 10 of Algorithm 1, we incorporate the coordinate-wise adaptive learning rate approach, akin to that utilized in Adam [32], which is defined as follows
ι t , i = ϱ ι t 1 , i + ( 1 ϱ ) ( ˜ v μ F i ( Ξ t , i ; ϑ t , i ) ) 2 ,
where the hyperparameter ϱ ( 0 , 1 ) , representing the decay factor.
Then, when the current iteration number t is an integer multiple of the local update number p (i.e., mod ( t , p ) = 0 ), we perform step 12.
In step 12 of Algorithm 1, for ι t , i , we perform averaging and aggregation step and resulting in ι ¯ t . Subsequently, we utilize the ι ¯ t to create an adaptive matrix K t = d i a g ( ι ¯ t + ρ ) (where d i a g denotes diagonal matrix), and where ρ > 0 ( ρ represent tuning parameter).
At the step 13 in algorithm 1, perform averaging and aggregation on all nodes of n t , i .
In step 14, the global model is updated based on the obtained K t and n ¯ t .
Otherwise, proceed to steps 15 and 16 of the algorithm. Keep K t as its previous value K t 1 and update the global model in the same manner as before.
In step 16 of algorithm 1, the identical K t is employed for the local updates across various edge devices.
Here, model parameters, n t , i , ι t , i are aggregated by the global server every interval of p steps.
At the step 17 in algorithm 1, all edge devices calculate the model difference of their local models in the current round, that is, Δ t , i = Ξ T , i Ξ 0 , i ( i [ Q ] , [ Q ] represents all Q edge devices from 1 to Q), uploaded to the central server.
At the step 18, 19 in Algorithm 1, after receiving the local model difference, central server summarizes it as Δ t = 1 Q i [ Q ] Δ t , i , and proceeds to update the global model, that is Ξ t + 1 = Ξ t + Δ t .
Finally, after all iterations are completed, the algorithm outputs a model Ξ ¯ , which is uniformly randomly selected from all the global models { Ξ ¯ t } t = 1 T obtained during the iterations, as the final result.
In the algorithm, K t is a diagonal matrix whose diagonal elements are related to ι ¯ t , which is computed based on certain statistics from the local model update process. Specifically, ι t , i is related to the square of the gradient estimates, and ι ¯ t is obtained by averaging ι t , i across all devices. Subsequently, K t is updated accordingly. This approach allows K t to adaptively adjust based on the variations in gradient information during the local model update process.
During the model training process, the data distribution on different devices may vary, leading to differing updates in model parameters. By adaptively updating K t , it is possible to influence the aggregation and update methods of the global model. For instance, when the gradient variation on a particular device is significant (i.e., ι ¯ t is large), the value of K t will also change accordingly, thereby assigning different weights to the amount of updates from this device when aggregating local updates. This enables the algorithm to better adapt to the data distribution and model training situations on different devices, accelerating the convergence speed of the model and enhancing the final model performance.
Preprints 146881 i001

4. Convergence Analysis of FAFedZO Method

For this part, the convergence of FAFedZO method will be discussed. To facilitate the theoretical analysis of the proposed algorithm, we need to make some assumptions as follows.
Assumption A1. 
The global loss function f ( Ξ ) in (1) has a lower bound, that is, there is a fixed value f * that exists f ( Ξ ) f * > .
The Assumption 1 means that regardless of the value of the optimization variable Ξ , the global loss function f ( Ξ ) will not decrease indefinitely, and there exists a minimum value f * , ensuring that the optimization problem is meaningful.
Assumption A2. 
We suppose that function F i ( Ξ , ϑ i ) , f i ( Ξ ) , f ( Ξ ) are all L-smooth, that is, for all Ξ 1 R d , Ξ 2 R d , we can conclude that
f i ( Ξ 1 ) f i ( Ξ 2 ) L Ξ 1 Ξ 2 , f ( Ξ 1 ) f ( Ξ 2 ) + < f ( Ξ 2 ) , Ξ 1 Ξ 2 > + L 2 Ξ 1 Ξ 2 2 .
The Assumption 2 is common in the theoretical analysis of non-convex optimization, indicating that the gradient changes of f i ( Ξ ) , f ( Ξ ) are smooth and do not change suddenly, which is crucial for proving the convergence of algorithms.
Assumption A3. 
F i ( Ξ , ϑ i ) serves as a precise approximation of f i ( Ξ ) without bias, namely
E ϑ i [ F i ( Ξ , ϑ i ) ] = f i ( Ξ ) , Ξ R d .
The Assumption 3 allows us to use sample gradients to approximate the true gradient.
Assumption A4. 
The variance of the stochastic gradient is limited within a certain range, that is, there have a fixed value σ g can satisfies
E ϑ i F i ( Ξ , ϑ i ) f i ( Ξ ) 2 σ g 2 , Ξ R d .
The Assumption 4 indicates that the variance of stochastic gradients will not be infinite, which is important for controlling noise and uncertainty in the optimization process.
Assumption A5. 
The difference between each local loss function and the global loss function remains within a certain limit, that is, there have a fixed value σ h can satisfies
f ( Ξ ) f i ( Ξ ) 2 σ h 2 , Ξ R d .
The Assumption 5 ensures that local optimization does not lead to a significant decrease in global performance.
Assumption A6. 
The inter-node variance is bounded, namely
f i ( Ξ ) f j ( Ξ ) 2 ζ 2 , Ξ R d .
The Assumption 6 indicates that the differences between loss functions on different nodes are limited, ensuring that the optimization processes on different nodes do not diverge too much.
Assumption A7. 
In our algorithms, for all t 1 , the adaptive matrices K t satisfies the condition that
λ m i n ( K t ) ρ > 0 ,
there ρ represents an appropriate positive value.
The Assumption 7 guarantees the step size in the optimization process does not become too small, thereby guaranteeing the convergence of the algorithm.
Assumption A8. 
The gradients of function f i ( Ξ ) is G-bounded, that is, for all i, Ξ R d , we have
f i ( Ξ ) G .
The Assumption 8 means that the gradients of the loss functions on each node are not infinite, ensuring stability and controllability in the optimization process.
We define the ϵ -stationary point as follows:
Definition 1. 
A point Ξ is called ϵ-stationary point if f ( Ξ ) ϵ . Generally, a stochastic algorithm is defined to achieve an ϵ-stationary point in T iterations if E f ( Ξ T ) ϵ .
Then, we investigate the convergence characteristics of our novel method based on Assumptions 1-8. The Appendix A contains the comprehensive proofs of our results.
About the convergence of the FAFedZO algorithm, we present the main results of this paper as follows:
Theorem 1. 
Assuming that the sequence { Ξ ¯ t } t = 1 T is produced by Algorithm 1. Based on the aforementioned Assumptions 1-8, given that t 0 , χ t + 1 = c · η t 2 , c = 1 12 L p h ¯ 3 ρ 2 + 30 L 2 ρ 2 60 L 2 ρ 2 , w t = max ( 3 2 , 1728 L 3 p 3 h ¯ 3 t ) , h ¯ = 1 L and set
η t = ρ h ¯ w t + t 1 / 3
we can conclude that
1 T t = 1 T E f ( Ξ ¯ t ) P 1 T t = 0 T 1 E [ Q t ] P [ 12 L p ρ 2 T + L ρ 2 T 2 / 3 E f Ξ ¯ 0 f * + 6 L 2 μ 2 p 2 ρ 2 Q T + 12 2 × 75 p ρ 2 T + 900 ρ 2 T 2 / 3 5 L 2 μ 2 3 + 3 ζ 2 ( ln T + 1 ) + L 2 μ 2 p 2 ρ 2 Q T 2 / 3 ] 1 / 2
where P = 5 2 d ( G 2 + σ g 2 ) + 2 ρ 2 + 1 2 d 2 L 2 μ 2
Remark 1. 
We utilize ζ as an indicator of data heterogeneity. The final results demonstrate that an increase in ζ (indicating greater data heterogeneity) leads to a slowdown in the training process. In addition, as the parameters L and μ increase, the boundary in the theorem conclusion will also increase, which will also result in a decrease in the training speed.
Remark 2. 
A suitable value of ρ ensures a balanced incorporation of adaptive information in the learning rate. In practice, we commonly select ρ to be within the order of O ( 1 ) , steering clear of excessively small or large values.
Analysis of computational complexity and convergence speed
(computational complexity) For Q t = 1 12 η t 2 Ξ ¯ t + 1 Ξ ¯ t 2 + 1 4 ρ 2 f Ξ ¯ t n ¯ t 2 , we have
1 T t = 0 T 1 E [ Q t ] = 1 T t = 0 T 1 E 1 12 η t 2 Ξ ¯ t + 1 Ξ ¯ t 2 + 1 4 ρ 2 f Ξ ¯ t n ¯ t 2 12 L p ρ 2 T + L ρ 2 T 2 / 3 E f Ξ ¯ 0 f * + 6 L 2 μ 2 p 2 ρ 2 Q T + L 2 μ 2 p 2 ρ 2 Q T 2 / 3 + 12 2 × 75 p ρ 2 T + 900 ρ 2 T 2 / 3 5 L 2 μ 2 3 + 3 ζ 2 ( ln T + 1 )
without loss of generality, we let p Q = O ( 1 ) , and choose p = T 1 3 . To make the right side of the inequality less than ϵ 2 , we can get T = O ( ϵ 3 ) and T p = T 2 3 = ϵ 2 . Therefore, to satisfy the definition of an ϵ -stationary point, which is E f ( Ξ T ) ϵ and E [ Q t ] ϵ 2 , we obtain the total sample cost as O ( ϵ 3 ) and the communication round as O ( ϵ 2 ) .
(convergence speed) In the FedZO algorithm, when full devices participate, it can satisfy
min t [ T ] E f x t 2 4 f x 0 f * H T η + η 24 d L N ( σ g 2 + 3 σ h 2 ) + d L 2 μ 2 12 + 4 L 2 μ 2 ,
when partial devices participate, it can satisfy
min t [ T ] E f x t 2 4 f x 0 f * H T η + η 32 d L M ( σ g 2 + 3 σ h 2 ) + η 36 H L σ h 2 M + 24 η H L 3 μ 2 + d L 2 μ 2 12 + 4 L 2 μ 2 .
Regardless of the situation, we can observe that the FedZO algorithm achieves the convergence rate of O ( 1 T ) . Furthermore, as indicated by Theorem 1, our algorithm can achieve the convergence rate of O ( 1 T 2 / 3 ) . Therefore, our FAFedZO method theoretically possesses a faster convergence rate compared to general zero-order methods. This further validates the advantages of our approach.

5. Experimental Results

For this part, we are conducting comparative experiments on the MNIST and CIFAR-10 datasets, present some experimental outcomes to assess the performance of the proposed FAFedZO method in federated black-box attacks, thereby confirming the advantages of this algorithm.
Experimental environment and datasets
Experimental environment configuration: This study utilizes the FedZO framework as the experimental baseline to conduct research on federated learning algorithms. This framework is renowned for its exceptional scalability and operational convenience, supporting numerous prominent federated learning algorithms. For this study, Python version 3.10.13 is adopted as the programming language, and PyTorch serves as the development platform. All tests are conducted on a Windows 10 platform equipped with an NV GTX1650 GPU and CUDA version 12.4.99.
Dataset: By conducting comparative experimental studies on the MNIST and CIFAR-10 datasets respectively, we have validated the practicality of the FAFedZO algorithm while significantly enhancing the performance of the model.
MNIST (Modified National Institute of Standards and Technology) is a widely used benchmark dataset for evaluating and comparing the performance of handwritten digit recognition algorithms. It was adapted by Professor Yann LeCun and his colleagues at New York University in 1998 from an original dataset of the National Institute of Standards and Technology (NIST). This dataset comprises a collection of grayscale images of handwritten digits (0-9) with 60,000 training samples and 10,000 test samples, each image measuring 28 × 28 pixels. Each image has undergone a preprocessing pipeline, including centering and normalization, aimed at enhancing the accuracy of classification algorithms. The MNIST dataset is extensively utilized for training and testing image processing and machine learning algorithms, particularly as a classic dataset for introductory deep learning and neural network research. An example of this dataset is shown in Figure 1.
CIFAR-10 (Canadian Institute For Advanced Research-10) is a commonly used dataset in computer vision, particularly for image classification tasks. It is consist of 60,000 32x32 color images across 10 categories and each category contains 6,000 images, and the dataset is divided into 50,000 training and 10,000 test images. Due to its smaller image size and fewer categories, it is often used for rapid verification, prototype development, as well as learning and understanding the basics of various computer vision algorithms. It is also employed as a benchmark dataset for deep learning models to assess their performance and generalization ability on image classification tasks.
Experimental result analysis
Here, we present the results of simulation experiments to assess the performance of the FAFedZO method in the context of black-box attack strategies.
Given the characteristics of black-box scenarios, optimizing black-box attacks falls into the realm of zero-order optimization. We investigate black-box attacks on a trained deep neural network(DNN) classification model. The purpose of our experiment is to train an interference image with the same size as the image in the dataset, which makes it difficult for the human eyes to recognize the difference between the original image and the adversarial image after adding the interference image, but it can induce the classification model to make a wrong judgment. We want to achieve a higher attack success rate with as little disturbance as possible, so we consider the loss function as shown below:
ψ ( Ξ , ϑ ) = max φ y ϑ 1 2 t a n h t a n h 1 2 ϑ + Ξ max j y ϑ φ j 1 2 t a n h t a n h 1 2 ϑ + Ξ , 0 I + c 1 2 t a n h t a n h 1 2 ϑ + Ξ ϑ 2 I I
Here Ξ represents the interference image to be trained, ϑ represents the original image in the dataset, and y ϑ represents the label corresponding to the image ϑ (for example, the label of the image "deer" is 4) , φ j ( ϑ ) represents the confidence that the classification model recognizes image ϑ as label j. I in (8) formula measures the probability of attack failure (marked as attack loss), I I represents the image distortion caused by disturbance, and c is the balance coefficient. In this way our goal can be achieved by minimizing ψ ( Ξ , ϑ ) . Next we use the (8) formula to construct the local loss function.
We divide the samples in the dataset into Q groups randomly and unevenly without repetition, and then distribute them to each edge device, where Q is the total count of edge devices we preset. Then for all edge devices we define its local loss function as:
f i ( Ξ ) E ϑ i D i ψ ( Ξ , ϑ i ) , i { 1 , 2 , . . . , Q }
Where D i represents the dataset at the ith edge device. In this way, the federal black-box attack problem on the DNN classification model can be formulated as a federated optimization problem: f ( Ξ ) 1 Q i = 1 Q f i ( Ξ ) . Next, we use the FAFedZO algorithm proposed in this paper to solve this problem.
We select balance parameter c = 1 . And for the remaining parameters, we select them as ( b 1 , b 2 , η , μ ) = ( 25 , 20 , 0.1 , 0.001 ) .
In Figure 2, we demonstrate the change in the accuracy of the federated black-box attack as the number of communication rounds increases. Figure (a) illustrates the impact of the account of local updates E { 3 , 15 , 60 } on attack accuracy of the proposed FAFedZO algorithm when the total number of edge devices Q = 50 and the number of participating edge devices M = 30 . We also compare it with the FedZO algorithm. It can be observed that the larger the E value, the higher the attack accuracy of the algorithm. Furthermore, the attack accuracy of FAFedZO is significantly higher than that of FedZO. When E = 60 , they achieve comparable accuracy, proving that the algorithm proposed in this paper outperforms the original FedZO algorithm. Figure (b) shows how the convergence performance of the FAFedZO method is affected by the number of participating edge devices M { 10 , 50 , 100 } when Q = 100 and E = 60 . It can be seen that by adjusting the M value, the FAFedZO algorithm can effectively enhance the attack accuracy.
In Figure 3, we present the relationships between attack loss and communication rounds as well as attack accuracy and communication rounds, respectively. The influence of the number of local updates E { 1 , 5 , 20 } on the convergence performance of the algorithm is studied under the condition of Q = 50 and M = 50 , i.e., all devices participate. It is evident that the FAFedZO scheme is capable of significantly decreasing the attack loss and improve attack accuracy compared to the FedZO algorithm under different E values. Moreover, as the value of E increases, the convergence speed of the FAFedZO algorithm also accelerates, with lower attack loss and higher attack accuracy, both of which tend to stabilize faster.
Figure 4 also shows the relationship between attack loss, attack accuracy, and communication rounds. However, we study how the number of participating edge devices M { 10 , 20 , 50 } influences the convergence performance of the scheme under the condition of Q = 50 and E = 1 . It is easy to see that the FAFedZO algorithm is far superior to the FedZO algorithm in terms of both attack loss and accuracy. Even when the FedZO algorithm takes the optimal value of M among 10, 20, and 50, its performance is still weaker than the worst value of the FAFedZO algorithm. In addition, as M increases, the attack loss value decreases, and the accuracy increases, which is in line with our expectations.
The aforementioned experiments were all conducted on the MNIST dataset, and similar conclusions were also drawn from experiments performed on the CIFAR-10 dataset. As shown in Figure 5, under the conditions of Q = 50 and M = 50 , the impact curves of the number of local updates E { 1 , 5 , 20 } on attack loss and accuracy are presented. It is evident that by adjusting the E value, the performance of the FAFedZO algorithm can also be significantly enhanced.
In addition to investigating the impact of the number of local updates and the number of participating edge devices on the algorithm’s performance, we also studied the influence of varying the number of random directions on the proposed algorithm. As shown in Figure 6, on the MNIST dataset, under the conditions of Q = 50 and M = 10 , we present the curves illustrating the impact of the number of directions H { 3 , 15 , 60 } on attack loss and attack accuracy. It can be observed that we have similar conclusions to the above figures, which further confirms the superiority of the FAFedZO algorithm.
In addition, we also discuss the performance of our algorithm under the setting of non-independent and identically distributed (non-iid) data.
Figure 7 presents the impact of the number of local updates on the attack loss and accuracy of the algorithm under the non-iid setting. It can be observed that, compared with Figure 2, the attack accuracy in this case is lower than that under the iid setting, which is attributed to the influence of the non-iid setting and aligns with our expectations. Furthermore, the results also indicate that under the non-iid setting, the performance of the FAFedZO algorithm remains superior to that of the FedZO algorithm, and both achieve comparable performance levels when E = 60 .
The conclusions we previously mentioned are also supported by Figure 8 and Figure 9. Therefore, in summary, the FAFedZO algorithm demonstrates superior performance in both iid and non-iid environments.

6. Discussion

Our approach combines zeroth-order optimization with adaptive techniques, yielding significant benefits in various real-world applications. For instance, in healthcare, medical data contains sensitive patient information and varies across different regions and hospitals. Utilizing this method allows the model to automatically adapt to these variations during training. Additionally, zeroth-order optimization can be leveraged to explore better solutions, enhancing the model’s generalization capability across diverse medical scenarios and improving its assistance in disease diagnosis. In the financial sector, financial institutions need to accurately identify risks such as fraudulent transactions. By employing this method, risk characteristics can be adaptively captured from vast amounts of financial transaction data, and zeroth-order optimization can be used for efficient optimization, thereby improving the accuracy of risk identification and preventing financial risks. In the realm of IoT devices, where resources are limited, zeroth-order optimization eliminates the need for gradient computation, reducing computational load. Adaptive methods can adjust training strategies based on device resources and data characteristics, enabling federated learning to operate efficiently on IoT devices and enhancing their data processing capabilities.

7. Conclusion

In this paper, we proposed FAFedZO, a zero-order federated optimization method based on an adaptive approach with a shared adaptive learning rate. In the non-convex setting, when gradient information is not available, FAFedZO can estimates the first-order gradient information through function queries to optimize the objective function. We have combined the advantages of zero-order optimization and adaptability. Zero-order estimation can solve the problem that the objective function is not available, thus the gradient information is not usable. Adaptability can accelerate the performance of the algorithm. Furthermore, our theoretical analysis of the algorithm demonstrates the convergence of FAFedZO. Finally, we conducted extensive experiments to verify the effectiveness of the proposed algorithm again on the MNIST and CIFAR-10 datasets, respectively.

Author Contributions

All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Key Technologies Research and Development Program of Henan Province under Grant No. 242102210102.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare there is no conflicts of interest.

Appendix A

Here, we present a thorough examination of our algorithm’s convergence properties. To facilitate the analysis, we introduce the notation g t , i that g t , i = f i ( Ξ t , i ) in the subsequent sections. Define s t as s t = t / p p and use ⊗ as Kronecker product symbol. Below are some basic lemmas required for our analysis.
At first, for the zeroth-order gradient ˜ v μ F i ( Ξ t , i , ϑ t , i ) , based on the characteristics of the gradient estimator outlined in [38], Lemma 4.2], we can deduce that
E ˜ v μ F i ( Ξ t , i , ϑ t , i ) = f i μ Ξ t , i
Then, we can get
E ˜ v μ F i ( Ξ t , i , ϑ t , i ) f i μ Ξ t , i 2 E ˜ v μ F i ( Ξ t , i , ϑ t , i ) 2 2 d E F i ( Ξ t , i , ϑ t , i ) 2 + 1 2 d 2 L 2 μ 2 = 2 d E F i ( Ξ t , i , ϑ t , i ) f i ( Ξ t , i ) + f i ( Ξ t , i ) 2 + 1 2 d 2 L 2 μ 2 = 2 d E F i ( Ξ t , i , ϑ t , i ) f i ( Ξ t , i ) 2 + 2 d E f i ( Ξ t , i ) 2 + 1 2 d 2 L 2 μ 2 2 d σ g 2 + 2 d E f i ( Ξ t , i ) 2 + 1 2 d 2 L 2 μ 2 2 d ( G 2 + σ g 2 ) + 1 2 d 2 L 2 μ 2
where the second equality is due to Assumption 3, the establishment of these four inequalities is because E z E z 2 E z 2 , [38], Lemma 4.1] and Assumption 4, Assumption 8.
Thus, we can get
E ˜ v μ F i ( Ξ t , i , ϑ t , i ) f i μ Ξ t , i 2 2 d ( G 2 + σ g 2 ) + 1 2 d 2 L 2 μ 2
At the same time, we can also obtain
E ˜ v μ F i ( Ξ t , i , ϑ t , i ) 2 2 d ( G 2 + σ g 2 ) + 1 2 d 2 L 2 μ 2
The following Lemma A1 is a small conclusion throughout the proof.
Lemma A1 
([39]).
Given Ξ t R Q d and Ξ ¯ t R d , there 1 R Q represents the vector composed entirely of ones, we can get conclusion that
Ξ 1 Ξ ¯ 2 Ξ 2
Then we continue to delve into the upper bound of the adaptive matrices K t .
Lemma A2. 
Suppose the adaptive matrices sequence { K t } t = 1 T is derived from Algorithm. On the Basis of Assumptions 1 to 8, we can conclude that
E K t 2 2 d ( G 2 + σ g 2 ) + 2 ρ 2 + 1 2 d 2 L 2 μ 2
Proof. Firstly, we know that
E K t 2 = E K s t 2 = E ι ¯ s t + ρ 2 2 E ι ¯ s t + 2 ρ 2
there K t = diag ( ι ¯ t + ρ ) is a diagnoal matrix. And following the definition of ι s t , we can deduce that
E ι ¯ s t = E ϱ ι ¯ s t 1 + ( 1 ϱ ) 1 Q i = 1 Q [ ˜ v μ F i ( Ξ s t 1 , i , ϑ s t 1 , i ) ] 2 ϱ E ι ¯ s t 1 + ( 1 ϱ ) 1 Q i = 1 Q E [ ˜ v μ F i ( Ξ s t 1 , i , ϑ s t 1 , i ) ] 2 ϱ E ι ¯ s t 1 + ( 1 ϱ ) 1 Q i = 1 Q E ˜ v μ F i ( Ξ s t 1 , i , ϑ s t 1 , i ) 2 2 ϱ E ι ¯ s t 1 + ( 1 ϱ ) 1 Q i = 1 Q [ 2 d ( G 2 + σ g 2 ) + 1 2 d 2 L 2 μ 2 ] = ϱ E ι ¯ s t 1 + ( 1 ϱ ) [ 2 d ( G 2 + σ g 2 ) + 1 2 d 2 L 2 μ 2 ] ϱ E ι ¯ s t 1 + ( 1 ϱ ) C
Therefore, because of ϱ ( 0 , 1 ) , taking the recursive expansion we can get
E ι ¯ s t ( 1 ϱ ) C + ϱ ( 1 ϱ ) C + ϱ 2 ( 1 ϱ ) C + + ϱ s t 1 ( 1 ϱ ) C = 1 ϱ s t 1 ϱ ( 1 ϱ ) C 1 2 C
So we finally got
E K t 2 2 E ι ¯ s t + 2 ρ 2 2 ρ 2 + C = 2 d ( G 2 + σ g 2 ) + 2 ρ 2 + 1 2 d 2 L 2 μ 2
Lemma A2 is proved. □
The following Lemma A3 measures the boundary of variance between gradients.
Lemma A3. 
For i [ Q ] , [ Q ] represents all Q edge devices from 1 to Q, we can conclude that
E ˜ v μ F i ( Ξ t , i ; B t , i ) g t , i 2 L 2 μ 2
E g t 1 g ¯ t 2 6 L 2 E Ξ t 1 Ξ ¯ t 2 + 3 Q ζ 2
Proof. For inequality (A9), we have
E ˜ v μ F i ( Ξ t , i ; B t , i ) g t , i 2 = E 1 b ϑ t , i B t , i ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) 2 = 1 b 2 E ϑ t , i B t , i ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) 2 1 b ϑ t , i B t , i E f i μ ( Ξ t , i ; ϑ t , i ) g t , i 2 L 2 μ 2
where the first inequality is because i = 1 q a i 2 q i = 1 q a i 2 , the last inequality follows by [28], Lemma 5.2].
For inequality (A10), we have
E g t 1 g ¯ t 2 = i = 1 Q E g t , i g ¯ t 2 3 i = 1 Q E [ g t , i f i ( Ξ ¯ t ) 2 + 1 Q j = 1 Q f j ( Ξ ¯ t ) g t , j 2 + 1 Q j = 1 Q f i ( Ξ ¯ t ) f j ( Ξ ¯ t ) 2 ] = 3 i = 1 Q E g t , i f i ( Ξ ¯ t ) 2 + 3 i = 1 Q 1 Q j = 1 Q E f j ( Ξ ¯ t ) g t , j 2 + 3 i = 1 Q 1 Q j = 1 Q E f i ( Ξ ¯ t ) f j ( Ξ ¯ t ) 2 3 L 2 i = 1 Q E Ξ t , i Ξ ¯ t 2 + 3 L 2 j = 1 Q E Ξ ¯ t Ξ t , j 2 + 3 i = 1 Q 1 Q j = 1 Q E f i ( Ξ ¯ t ) f j ( Ξ ¯ t ) 2 6 L 2 E Ξ t 1 Ξ ¯ t 2 + 3 Q ζ 2
the last two inequalities here are attributed to Assumption 2, 6.
Thus lemma A3 is proved. □
Below, we study the bounds under different values of i = 1 Q Ξ t , i Ξ ¯ t 2 , which is important for the proof of the subsequent theorems.
Lemma A4. 
Given that t [ t / p p , t / p ( p + 1 ) ] , Ξ t is derived from the Algorithm, we can conclude that
(1) if t = s t = t / p p , then we can get
i = 1 Q Ξ s t , i Ξ ¯ s t 2 = 0
(2) if t t / p p , then we can get
i = 1 Q Ξ t , i Ξ ¯ t 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q K s 1 ( n s , i n ¯ s ) 2
Proof.
(1) t = s t = t / p p , then we can conclude that mod ( t , p ) = 0 , so there is Ξ t + 1 , i = Ξ ¯ t + 1 , thus we have obtained the final result.
(2) we have
Ξ t , i = Ξ s t , i s = s t t 1 η s K s 1 n s , i Ξ ¯ t = Ξ ¯ s t s = s t t 1 η s K s 1 n ¯ s
thus
i = 1 Q Ξ t , i Ξ ¯ t 2 = i = 1 Q Ξ s t , i Ξ ¯ s t s = s t t 1 η s K s 1 n s , i s = s t t 1 η s K s 1 n ¯ s 2 = i = 1 Q s = s t t 1 η s K s 1 n s , i n ¯ s 2 ( t s t ) s = s t t 1 η s 2 i = 1 Q K s 1 ( n s , i n ¯ s ) 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q K s 1 ( n s , i n ¯ s ) 2
Lemma A4 is proved. □
The following lemmas A5-A7 are crucial conclusions that will be used in proving the final theorem.
Lemma A5. 
Assuming the sequence { Ξ t } 0 T is generated by the Algorithm, then we can get
E f ( Ξ ¯ t + 1 ) E f ( Ξ ¯ t ) ( 3 ρ 4 η t L 2 ) E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 + 5 η t 2 ρ E g ¯ t n ¯ t 2 + 5 η t L 2 2 ρ Q E Ξ t 1 Ξ ¯ t 2
Proof.
f ( Ξ ¯ t + 1 ) f ( Ξ ¯ t ) + f ( Ξ ¯ t ) , Ξ ¯ t + 1 Ξ ¯ t + L 2 Ξ ¯ t + 1 Ξ ¯ t 2 = f ( Ξ ¯ t ) + f ( Ξ ¯ t ) n ¯ t , Ξ ¯ t + 1 Ξ ¯ t ( 1 ) + n ¯ t , Ξ ¯ t + 1 Ξ ¯ t ( 2 ) + L 2 Ξ ¯ t + 1 Ξ ¯ t 2
Regarding (1), we can obtain
( 1 ) = f ( Ξ ¯ t ) n ¯ t , Ξ ¯ t + 1 Ξ ¯ t f Ξ ¯ t n ¯ t Ξ ¯ t + 1 Ξ ¯ t η t ρ f ( Ξ ¯ t ) n ¯ t 2 + ρ 4 η t Ξ ¯ t + 1 Ξ ¯ t 2
Regarding term (2), K t = diag ( ι ¯ s t + ρ ) , Ξ ¯ t + 1 = Ξ ¯ t η t K t 1 n ¯ t . We can obtain based on the given definition of K t and assumption 7 that
n ¯ t , 1 η t ( Ξ ¯ t Ξ ¯ t + 1 ) = K t 1 η t ( Ξ ¯ t Ξ ¯ t + 1 ) , 1 η t ( Ξ ¯ t Ξ ¯ t + 1 ) ρ 1 η t ( Ξ ¯ t Ξ ¯ t + 1 ) 2
So
( 2 ) = n ¯ t , Ξ ¯ t + 1 Ξ ¯ t η t ρ 1 η t ( Ξ ¯ t Ξ ¯ t + 1 ) 2 = ρ η t Ξ ¯ t + 1 Ξ ¯ t 2
Bring (A20), (A18) into (A17), we can get
f ( Ξ ¯ t + 1 ) f ( Ξ ¯ t ) + η t ρ f ( Ξ ¯ t ) n ¯ t 2 + ρ 4 η t Ξ ¯ t + 1 Ξ ¯ t 2 ρ η t Ξ ¯ t + 1 Ξ ¯ t 2 + L 2 Ξ ¯ t + 1 Ξ ¯ t 2 f ( Ξ ¯ t ) η t 4 ρ f ( Ξ ¯ t ) n ¯ t 2 + 5 η t 4 ρ f ( Ξ ¯ t ) n ¯ t 2 ( 3 ρ 4 η t L 2 ) Ξ ¯ t + 1 Ξ ¯ t 2 f ( Ξ ¯ t ) ( 3 ρ 4 η t L 2 ) Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ f ( Ξ ¯ t ) n ¯ t 2 + 5 η t 2 ρ g ¯ t n ¯ t 2 + 5 η t 2 ρ f ( Ξ ¯ t ) g ¯ t 2
We considering the last term in (A21), taking expectation on both sides have
E f ( Ξ ¯ t ) g ¯ t 2 = E 1 Q i = 1 Q ( f i ( Ξ ¯ t ) g t , i ) 2 1 Q 2 Q i = 1 Q E f i ( Ξ ¯ t ) g t , i 2 L 2 Q i = 1 Q E Ξ t , i Ξ ¯ t 2 = L 2 Q E Ξ t 1 Ξ ¯ t 2
By substituting it into (A21), then taking the expectation of both sides, we can get this conclusion.
Lemma A5 is proved. □
Lemma A6. 
Suppose that n t are produced by the Algorithm, we subsequently obtain
E n ¯ t g ¯ t 2 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 4 ( 1 χ t ) 2 L 2 Q E Ξ t Ξ t 1 2 + 4 χ t 2 L 2 μ 2
Proof. We know that
n ¯ t = 1 Q i = 1 Q [ ˜ v μ F i ( Ξ t , i ; B t , i ) + ( 1 χ t ) ( n ¯ t 1 ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ) ]
So we can get
E n ¯ t g ¯ t 2 = E 1 Q i = 1 Q [ ˜ v μ F i ( Ξ t , i ; B t , i ) + ( 1 χ t ) ( n ¯ t 1 ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ) ] g ¯ t 2 = E 1 Q i = 1 Q [ ( ˜ v μ F i ( Ξ t , i ; B t , i ) g t , i ) ( 1 χ t ) ( ˜ v μ F i ( Ξ t 1 , i ; B t , i ) g t 1 , i ) ] + ( 1 χ t ) ( n ¯ t 1 g ¯ t 1 ) 2 = E 1 b ϑ t , i B t , i 1 Q i = 1 Q [ ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) ( 1 χ t ) ( f i μ ( Ξ t 1 , i ; ϑ t , i ) g t 1 , i ) ] + ( 1 χ t ) ( n ¯ t 1 g ¯ t 1 ) 2 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 2 b 2 Q 2 b Q ϑ t , i B t , i i = 1 Q E ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) ( 1 χ t ) ( f i μ ( Ξ t 1 , i ; ϑ t , i ) g t 1 , i ) 2 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 2 b Q ϑ t , i B t , i i = 1 Q E ( 1 χ t ) [ ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) ( f i μ ( Ξ t 1 , i ; ϑ t , i ) g t 1 , i ) ] + χ t ( f i μ ( Ξ t , i ; ϑ t , i ) g t , i ) 2 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 4 ( 1 χ t ) 2 b Q ϑ t , i B t , i i = 1 Q E f i μ ( Ξ t , i ; ϑ t , i ) f i μ ( Ξ t 1 , i ; ϑ t , i ) 2 + 4 χ t 2 b Q ϑ t , i B t , i i = 1 Q E f i μ ( Ξ t , i ; ϑ t , i ) g t , i 2 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 4 ( 1 χ t ) 2 L 2 Q i = 1 Q E Ξ t , i Ξ t 1 , i 2 + 4 χ t 2 L 2 μ 2 = 2 ( 1 χ t ) 2 E n ¯ t 1 g ¯ t 1 2 + 4 ( 1 χ t ) 2 L 2 Q E Ξ t Ξ t 1 2 + 4 χ t 2 L 2 μ 2
where the last inequality is because the L-smoothness and Lemma A3.
Lemma A6 is proved. □
Lemma A7. 
Suppose that n t are produced by the Algorithm, we subsequently obtain
30 ρ 72 Q t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 ρ 4 t = s t s ¯ 1 η t E Ξ ¯ t + 1 Ξ ¯ t 2 + ρ μ 2 c 2 4 Q + 3 ρ ζ 2 c 2 4 L 2 t = s t s ¯ η t 3
Proof.
i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 i = 1 Q E K t 1 [ [ ˜ v μ F i ( Ξ t , i ; B t , i ) + ( 1 χ t ) ( n t 1 , i ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ) ] 1 Q i = 1 Q [ ˜ v μ F i ( Ξ t , i ; B t , i ) + ( 1 χ t ) ( n t 1 , i ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ) ] ] 2 = i = 1 Q E K t 1 [ ( 1 χ t ) ( n t 1 , i n ¯ t 1 ) + [ ˜ v μ F i ( Ξ t , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t , i ; B t , i ) ( 1 χ t ) ( ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ) ] ] 2 ( 1 + γ ) ( 1 χ t ) 2 i = 1 Q E K t 1 ( n t 1 , i n ¯ t 1 ) 2 + ( 1 + 1 γ ) 1 ρ 2 i = 1 Q E [ ˜ v μ F i ( Ξ t , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t , i ; B t , i ) ] ( 1 χ t ) [ ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ] 2
there (A26) arises is because the fact that K t ρ I d . Continuing to process the latter term in (A26) yields
i = 1 Q E ˜ v μ F i ( Ξ t , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t , i ; B t , i ) ( 1 χ t ) [ ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ] 2 2 i = 1 Q E [ ˜ v μ F i ( Ξ t , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t , i ; B t , i ) ] [ ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) ] 2 + 2 χ t 2 i = 1 Q E ˜ F i v μ ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 2 2 i = 1 Q E ˜ v μ F i ( Ξ t , i ; B t , i ) ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 2 + 2 χ t 2 i = 1 Q E ˜ F i v μ ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 2 2 i = 1 Q L 2 E Ξ t , i Ξ t 1 , i 2 + 2 χ t 2 i = 1 Q E ˜ F i v μ ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 2 = 2 L 2 E Ξ t Ξ t 1 2 + 2 χ t 2 i = 1 Q E ˜ F i v μ ( Ξ t 1 , i ; B t , i ) 1 Q i = 1 Q ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 2
the last two inequalities here are attributed to Lemma A1 and the L-smoothness. Regarding the last term, we can get
i = 1 Q E ˜ v μ F i ( Ξ t 1 , i ; B t , i ) 1 Q j = 1 Q ˜ v μ F j ( Ξ t 1 , j ; B t , j ) 2 = i = 1 Q E [ ˜ v μ F i ( Ξ t 1 , i ; B t , i ) g t 1 , i ] 1 Q j = 1 Q [ ˜ v μ F j ( Ξ t 1 , j ; B t , j ) g t 1 , j ] + [ g t 1 , i g ¯ t 1 ] 2 2 i = 1 Q E ˜ v μ F i ( Ξ t 1 , i ; B t , i ) g t 1 , i 2 + 2 i = 1 Q E g t 1 , i g ¯ t 1 2 2 Q L 2 μ 2 + 6 Q ζ 2 + 12 L 2 E Ξ t 1 1 Ξ ¯ t 1 2
the last two inequalities here are attributed to Lemma A1 and Lemma A3. Then, by integrating the aforementioned inequalities (A28), (A27), (A26) with the definition of K t , we can deduce that when mod ( t , p ) 0 , the following formula can be reached.
i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 ( 1 χ t ) 2 ( 1 + γ ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 2 L 2 ρ 2 ( 1 + 1 γ ) E Ξ t Ξ t 1 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 E Ξ t 1 1 Ξ ¯ t 1 2 ( 1 χ t ) 2 ( 1 + γ ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 2 L 2 ρ 2 ( 1 + 1 γ ) i = 1 Q E Ξ t , i Ξ t 1 , i 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 i = 1 Q E Ξ t 1 , i Ξ ¯ t 1 2 ( 1 χ t ) 2 ( 1 + γ ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 2 L 2 ρ 2 ( 1 + 1 γ ) i = 1 Q E η t 1 K t 1 1 n t 1 , i 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 ( 1 χ t ) 2 ( 1 + γ ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 + 4 L 2 ρ 2 ( 1 + 1 γ ) i = 1 Q E η t 1 K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + η t 1 K t 1 1 n ¯ t 1 2
where we utilize the Lemma A4. Then combine like terms we have
i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 [ ( 1 χ t ) 2 ( 1 + γ ) + 4 L 2 ρ 2 ( 1 + 1 γ ) η t 1 2 ] i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 4 Q L 2 ρ 2 ( 1 + 1 γ ) η t 1 2 E K t 1 1 n ¯ t 1 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2
We select γ = 1 p , η t ρ 12 L p and we know that χ t ( 0 , 1 ) , then
( 1 χ t ) 2 ( 1 + γ ) + 4 L 2 ρ 2 ( 1 + 1 γ ) η t 1 2 1 + 1 p + 4 L 2 ρ 2 ( 1 + p ) η t 1 2 1 + 1 p + p + 1 36 p 2 1 + 19 18 p
Substitute (A31) into (A30), then we can get
i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 ( 1 + 19 18 p ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 4 Q L 2 ρ 2 ( 1 + 1 γ ) η t 1 2 E K t 1 1 n ¯ t 1 2 + 4 L 2 μ 2 ρ 2 ( 1 + 1 γ ) χ t 2 + 12 Q ρ 2 ζ 2 ( 1 + 1 γ ) χ t 2 + 24 L 2 ( 1 + 1 γ ) χ t 2 ρ 2 ( p 1 ) s = s t t 1 η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 ( 1 + 19 18 p ) i = 1 Q E K t 1 1 ( n t 1 , i n ¯ t 1 ) 2 + 2 Q L 3 ρ η t 1 E K t 1 1 n ¯ t 1 2 + 2 L μ 2 c 2 3 ρ η t 1 3 + 2 Q ζ 2 c 2 L ρ η t 1 3 + 48 L 2 p 2 c 2 η t 1 4 ρ 2 s = s t t 1 η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2
where we considering χ t = c η t 1 2 and 1 + p p + p = 2 p .
On the other hand, if mod ( t , p ) = 0 that is t = s t , it follows that i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 = 0 . So taking the recursive expansion to (A32), we have
i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 2 Q L 3 ρ s = s t t 1 ( 1 + 19 18 p ) t 1 s η s E K s 1 n ¯ s 2 + 2 L μ 2 c 2 3 ρ + 2 Q ζ 2 c 2 L ρ s = s t t 1 ( 1 + 19 18 p ) t 1 s η s 3 + 48 L 2 p 2 c 2 ρ 2 s = s t t 1 ( 1 + 19 18 p ) t 1 s η s 4 s ¯ = s t s η s ¯ 2 i = 1 Q E K s ¯ 1 ( n s ¯ , i n ¯ s ¯ ) 2 2 Q L 3 ρ s = s t t 1 ( 1 + 19 18 p ) p η s E K s 1 n ¯ s 2 + 2 L μ 2 c 2 3 ρ + 2 Q ζ 2 c 2 L ρ s = s t t 1 ( 1 + 19 18 p ) p η s 3 + 48 L 2 p 3 c 2 ρ 2 ( ρ 12 L p ) 5 ( 1 + 19 18 p ) p s = s t t η s i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 2 Q L ρ s = s t t η s E K s 1 n ¯ s 2 + 2 L μ 2 c 2 ρ + 6 Q ζ 2 c 2 L ρ s = s t t η s 3 + 144 L 2 p 3 c 2 ρ 2 ( ρ 12 L p ) 5 s = s t t η s i = 1 Q E K s 1 ( n s , i n ¯ s ) 2
where we utilize that t 1 s p 1 < p and ( 1 + 19 18 p ) p e 19 18 3 . Then by multiplying both sides by t = s t s ¯ η t we can get
t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 2 Q L ρ t = s t s ¯ η t s = s t t η s E K s 1 n ¯ s 2 + 2 L μ 2 c 2 ρ + 6 Q ζ 2 c 2 L ρ t = s t s ¯ η t s = s t t η s 3 + 144 L 2 p 3 c 2 ρ 2 ( ρ 12 L p ) 5 t = s t s ¯ η t s = s t t η s i = 1 Q E K s 1 ( n s , i n ¯ s ) 2
Finally,
t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 2 Q L ρ ( t = s t s ¯ η t ) t = s t s ¯ η t E K t 1 n ¯ t 2 + 2 L μ 2 c 2 ρ + 6 Q ζ 2 c 2 L ρ ( t = s t s ¯ η t ) t = s t s ¯ η t 3 + 144 L 2 p 3 c 2 ρ 2 ( ρ 12 L p ) 5 ( t = s t s ¯ η t ) t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 Q 6 t = s t s ¯ η t E K t 1 n ¯ t 2 + μ 2 c 2 6 + Q ζ 2 c 2 2 L 2 t = s t s ¯ η t 3 + 144 L 2 p 4 c 2 ρ 2 ( ρ 12 L p ) 6 t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2
where the last inequality is because η t ρ 12 L p so that t = s t s ¯ η t ( s ¯ s t ) ρ 12 L p p ρ 12 L p = ρ 12 L .
Therefore,
[ 1 144 L 2 p 4 c 2 ρ 2 ( ρ 12 L p ) 6 ] t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 Q 6 t = s t s ¯ η t E K t 1 n ¯ t 2 + μ 2 c 2 6 + Q ζ 2 c 2 2 L 2 t = s t s ¯ η t 3
Given that c 60 L 2 ρ 2 and 1 144 L 2 p 4 c 2 ρ 2 ( ρ 12 L p ) 6 20 72 . By multiply 3 ρ 2 Q on both side, we can get
30 ρ 72 Q t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 ρ 4 t = s t s ¯ η t E K t 1 n ¯ t 2 + ρ μ 2 c 2 4 Q + 3 ρ ζ 2 c 2 4 L 2 t = s t s ¯ η t 3 = ρ 4 t = s t s ¯ 1 η t E Ξ ¯ t + 1 Ξ ¯ t 2 + ρ μ 2 c 2 4 Q + 3 ρ ζ 2 c 2 4 L 2 t = s t s ¯ η t 3
Lemma A7 is proved. □
Proof of Theorem
Now let’s prove the final theorem.
Proof. We set η t = ρ h ¯ w t + t 1 / 3 , χ t + 1 = c · η t 2 , c = 1 12 L p h ¯ 3 ρ 2 + 30 L 2 ρ 2 60 L 2 ρ 2 , h ¯ = 1 L and w t = max ( 3 2 , 1728 L 3 p 3 h ¯ 3 t ) . So we can infer that
η t = min { ρ 1 L 3 2 + t 1 3 , ρ 1 L 1728 L 3 p 3 1 L 3 t + t 1 3 } = min { ρ L 3 2 + t 1 3 , ρ 12 L p }
It is clear that η t ρ 12 L p . And
2 η t 1 η t 1 1 = 2 ( w t + t ) 1 / 3 ρ h ¯ ( w t 1 + t 1 ) 1 / 3 ρ h ¯ 2 3 ρ h ¯ ( w t + ( t 1 ) ) 2 / 3 2 3 ρ h ¯ ( w t / 3 + t / 3 ) 2 / 3 = 2 · 3 2 / 3 3 ρ h ¯ ( w t + t ) 2 / 3 = 2 · 3 2 / 3 3 ρ 3 h ¯ 3 · ρ 2 h ¯ 2 ( w t + t ) 2 / 3 = 2 · 3 2 / 3 3 ρ 3 h ¯ 3 η t 2 η t 6 ρ 2 h ¯ 3 L p
where we leverage the concavity of f ( Ξ ) = Ξ 1 / 3 , that is ( Ξ + y ) 1 / 3 Ξ 1 / 3 + y 3 Ξ 2 / 3 , the second inequality is valid because of w t 3 2 and the last inequality is obtained based on η t ρ 12 L p .
E n ¯ t + 1 g ¯ t + 1 2 η t E n ¯ t g ¯ t 2 η t 1 2 ( 1 χ t + 1 ) 2 η t 1 η t 1 E n ¯ t g ¯ t 2 + 4 ( 1 χ t + 1 ) 2 L 2 Q η t E Ξ t + 1 Ξ t 2 + 4 χ t + 1 2 L 2 μ 2 η t [ 2 η t 1 η t 1 1 2 c η t ] E n ¯ t g ¯ t 2 + 8 ( 1 χ t + 1 ) 2 L 2 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + 8 ( 1 χ t + 1 ) 2 L 2 η t E K t 1 n ¯ t 2 + 4 χ t + 1 2 L 2 μ 2 η t 60 L 2 ρ 2 η t E n ¯ t g ¯ t 2 + 8 L 2 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + 8 L 2 η t E K t 1 n ¯ t 2 + 4 c 2 η t 3 L 2 μ 2 .
where the second inequality holds true because ( a + b ) 2 2 a 2 + 2 b 2 and ( 1 χ t + 1 ) 2 1 χ t + 1 , χ t + 1 = c · η t 2 and the last inequality is obtained based on (A38). Therefore, we have
ρ 24 L 2 [ E n ¯ t + 1 g ¯ t + 1 2 η t E n ¯ t g ¯ t 2 η t 1 ] 5 2 ρ η t E n ¯ t g ¯ t 2 + ρ 3 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + ρ 3 η t E K t 1 n ¯ t 2 + ρ c 2 η t 3 μ 2 6
Subsequently, we set
Γ t = f ( Ξ ¯ t ) + ρ 24 L 2 n ¯ t g ¯ t 2 η t 1
E [ Γ t + 1 Γ t ] = E [ f ( Ξ ¯ t + 1 ) f ( Ξ ¯ t ) + ρ 24 L 2 ( n ¯ t + 1 g ¯ t + 1 2 η t n ¯ t g ¯ t 2 η t 1 ) ] ( 3 ρ 4 η t L 2 ) E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 + 5 η t L 2 ( p 1 ) 2 ρ Q s = s t t η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 + ρ 3 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + ρ 3 η t E K t 1 n ¯ t 2 + ρ c 2 η t 3 μ 2 6 = ( 3 ρ 4 η t L 2 ) E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f Ξ ¯ t n ¯ t 2 + 5 η t L 2 ( p 1 ) 2 ρ Q s = s t t η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 + ρ 3 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 + ρ c 2 η t 3 μ 2 6 ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 + 5 η t L 2 ( p 1 ) 2 ρ Q s = s t t η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 + ρ 3 Q η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + ρ c 2 η t 3 μ 2 6
where we utilize that Lemma A4, Lemma A5 and L 2 ρ 24 η t p ρ 24 η t . By summing the results from t = s t to s ¯ , where s ¯ [ t / p p , ( t / p + 1 ) p ] , then can obtain
E [ Γ s ¯ + 1 Γ s t ] t = s t s ¯ [ ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 ] + t = s t s ¯ ρ c 2 η t 3 μ 2 6 + t = s t s ¯ 5 η t L 2 ( p 1 ) 2 ρ Q s = s t t η s 2 i = 1 Q E K s 1 ( n s , i n ¯ s ) 2 + ρ 3 Q t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 t = s t s ¯ [ ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 ] + t = s t s ¯ ρ c 2 η t 3 μ 2 6 + 5 L 2 ( p 1 ) 2 ρ Q ( t = s t s ¯ η t ) t = s t s ¯ η t 2 i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 + ρ 3 Q t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 t = s t s ¯ [ ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 ] + ρ 3 Q t = s t s ¯ η t i = 1 Q E K t 1 n t , i n ¯ t 2 + 5 L 2 ( p 1 ) 2 ρ Q p × ρ 12 L p × ρ 12 L p t = s t s ¯ η t i = 1 Q E K t 1 n t , i n ¯ t 2 + t = s t s ¯ ρ c 2 η t 3 μ 2 6 t = s t s ¯ [ ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 ] + t = s t s ¯ ρ c 2 η t 3 μ 2 6 + 26 ρ 72 Q t = s t s ¯ η t i = 1 Q E K t 1 ( n t , i n ¯ t ) 2 t = s t s ¯ [ ρ 3 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f ( Ξ ¯ t ) n ¯ t 2 ] + t = s t s ¯ ρ c 2 η t 3 μ 2 6 + ρ 4 t = s t s ¯ 1 η t E Ξ ¯ t + 1 Ξ ¯ t 2 + ρ μ 2 c 2 4 Q + 3 ρ ζ 2 c 2 4 L 2 t = s t s ¯ η t 3
here, by utilizing Lemma A7 and the fact that 26 72 < 30 72 , we can derive the last inequality. Subsequently, summing the terms from the start can obtain
E Γ T Γ 0 t = 0 T 1 ρ 12 η t E Ξ ¯ t + 1 Ξ ¯ t 2 η t 4 ρ E f Ξ ¯ t n ¯ t 2 + t = 0 T 1 ρ c 2 η t 3 μ 2 6 + ρ μ 2 c 2 4 Q t = 0 T 1 η t 3 + 3 ρ ζ 2 c 2 4 L 2 t = 0 T 1 η t 3
Furthermore, we can obtain
t = 0 T 1 E ρ 12 η t Ξ ¯ t + 1 Ξ ¯ t 2 + η t 4 ρ f Ξ ¯ t n ¯ t 2 E Γ 0 Γ T + 5 ρ μ 2 c 2 12 t = 0 T 1 η t 3 + 3 ρ ζ 2 c 2 4 L 2 t = 0 T 1 η t 3 E f ( Ξ ¯ 0 ) f * + ρ 24 L 2 n ¯ 0 g ¯ 0 2 η 0 + 5 ρ μ 2 c 2 12 t = 0 T 1 η t 3 + 3 ρ ζ 2 c 2 4 L 2 t = 0 T 1 η t 3
Then consider that t = 0 T 1 η t 3 = t = 0 T 1 ρ 3 h ¯ 3 w t + t t = 0 T 1 ρ 3 h ¯ 3 1 + t ρ 3 h ¯ 3 ( ln T + 1 ) , since w t 3 2 > 1 . Taking Lemma A3 and dividing both sides of the above result by ρ η T T can get
1 T t = 0 T 1 E 1 12 η t 2 Ξ ¯ t + 1 Ξ ¯ t 2 + 1 4 ρ 2 f Ξ ¯ t n ¯ t 2 E f ( Ξ ¯ 0 ) f * η T T ρ + μ 2 η T T 24 Q η 0 + ρ 3 η T T L 2 5 L 2 μ 2 12 + 3 ζ 2 4 c 2 h ¯ 3 ( ln T + 1 )
Regarding the first term in (A45),
1 η T T = w T + T 1 / 3 ρ h ¯ T w T 1 / 3 ρ h ¯ T + 1 ρ h ¯ T 2 / 3 12 L q ρ T + L ρ T 2 / 3
For the middle term, we have
μ 2 η T T 24 Q η 0 12 L p ρ T + L ρ T 2 / 3 × μ 2 24 Q × w 0 1 / 3 ρ h ¯ 12 L p ρ T + L ρ T 2 / 3 × μ 2 24 Q × 12 L p ρ 6 L 2 μ 2 p 2 ρ 2 Q T + L 2 μ 2 p 2 ρ 2 Q T 2 / 3
For the third term
ρ 3 c 2 h ¯ 3 4 η T T L 2 12 L p ρ T + L ρ T 2 / 3 × 60 L 2 ρ 2 2 × ρ 3 h ¯ 3 4 L 2 = 12 L p ρ T + L ρ T 2 / 3 × 3600 L 4 ρ 4 × ρ 3 1 L 3 4 L 2 = 12 L p ρ T + L ρ T 2 / 3 × 900 ρ L = 12 2 × 75 p ρ 2 T + 900 ρ 2 T 2 / 3
We let Q t = 1 12 η t 2 Ξ ¯ t + 1 Ξ ¯ t 2 + 1 4 ρ 2 f Ξ ¯ t n ¯ t 2
1 T t = 0 T 1 E [ Q t ] = 1 T t = 0 T 1 E 1 12 η t 2 Ξ ¯ t + 1 Ξ ¯ t 2 + 1 4 ρ 2 f Ξ ¯ t n ¯ t 2 12 L p ρ 2 T + L ρ 2 T 2 / 3 E f Ξ ¯ 0 f * + 6 L 2 μ 2 p 2 ρ 2 Q T + L 2 μ 2 p 2 ρ 2 Q T 2 / 3 + 12 2 × 75 p ρ 2 T + 900 ρ 2 T 2 / 3 5 L 2 μ 2 3 + 3 ζ 2 ( ln T + 1 )
and if we choose p = T Q 2 1 3 , then p T = 1 ( Q T ) 2 3 , p 2 Q T = 1 Q 7 3 T 1 3 , p Q T 2 3 = 1 Q 5 3 T 1 3 . So we can infer that it is convergent.
Then with Jensen’s inequality and K t ρ can get
1 η t Ξ ¯ t Ξ ¯ t + 1 + 1 ρ f Ξ ¯ t n ¯ t = K t 1 n ¯ t + 1 ρ f ( Ξ ¯ t ) n ¯ t = 1 K t K t K t 1 n ¯ t + 1 ρ f Ξ ¯ t n ¯ t 1 K t n ¯ t + 1 K t f Ξ ¯ t n ¯ t 1 K t f Ξ ¯ t
Finally,
1 T t = 1 T E f ( Ξ ¯ t ) 1 T t = 1 T E K t 1 η t Ξ ¯ t Ξ ¯ t + 1 + 1 ρ f Ξ ¯ t n ¯ t 1 T t = 1 T E λ 2 K t 2 + 1 2 λ 1 η t Ξ ¯ t Ξ ¯ t + 1 + 1 ρ f Ξ ¯ t n ¯ t 2 = λ 2 1 T t = 1 T E K t 2 + 1 2 λ 1 T t = 1 T E 1 η t Ξ ¯ t Ξ ¯ t + 1 + 1 ρ f Ξ ¯ t n ¯ t 2 1 T t = 1 T E K t 2 1 T t = 1 T E 2 η t 2 Ξ ¯ t Ξ ¯ t + 1 2 + 2 ρ 2 f Ξ ¯ t n ¯ t 2 5 2 d ( G 2 + σ g 2 ) + 2 ρ 2 + 1 2 d 2 L 2 μ 2 1 T t = 0 T 1 E [ Q t ]
where we utilize Young’s inequality and ( a + b ) 2 2 a 2 + 2 b 2 ,
and λ = 1 T t = 1 T E 1 η t Ξ ¯ t Ξ ¯ t + 1 + 1 ρ f Ξ ¯ t n ¯ t 2 / 1 T t = 1 T E K t 2 .
Since 1 T t = 0 T 1 E [ Q t ] is convergent as mentioned before, it can be known that 1 T t = 1 T E f ( Ξ ¯ t ) is also convergent, thus the theorem is proved. □

References

  1. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial intelligence and statistics; 2017; pp. 1273–1282. [Google Scholar]
  2. Shi, Y.; Yang, K.; Yang, Z.; Zhou, Y. Mobile edge artificial intelligence: Opportunities and challenges 2021.
  3. Yang, L.; Tan, B.; Zheng, V.W.; Chen, K.; Yang, Q. Federated recommendation systems. Federated Learning: Privacy and Incentive, 2020, pp. 225–239.
  4. Yang, K.; Shi, Y.; Zhou, Y.; Yang, Z.; Fu, L.; Chen, W. Federated machine learning for intelligent IoT via reconfigurable intelligent surface. IEEE network 2020, 34, 16–22. [Google Scholar] [CrossRef]
  5. Tian, J.; Smith, J.S.; Kira, Z. Fedfor: Stateless heterogeneous federated learning with first-order regularization. arXiv.2209.10537 2022. [Google Scholar]
  6. Zhang, M.; Sapra, K.; Fidler, S.; Yeung, S.; Alvarez, J.M. Personalized federated learning with first order model optimization 2021.
  7. Elbakary, A.; Issaid, C.B.; Shehab, M.; Seddik, K.G.; ElBatt, T.A.; Bennis, M. Fed-Sophia: A Communication-Efficient Second-Order Federated Learning Algorithm. arXiv.2406.06655 2024. [Google Scholar]
  8. Dai, Z.; Low, B.K.H.; Jaillet, P. Federated Bayesian optimization via Thompson sampling. Advances in Neural Information Processing Systems 2020, 33, 9687–9699. [Google Scholar]
  9. Fang, W.; Yu, Z.; Jiang, Y.; Shi, Y.; Jones, C.N.; Zhou, Y. Communication-efficient stochastic zeroth-order optimization for federated learning. IEEE Transactions on Signal Processing 2022, 70, 5058–5073. [Google Scholar] [CrossRef]
  10. Li, Z.; Ying, B.; Liu, Z.; Yang, H. Achieving Dimension-Free Communication in Federated Learning via Zeroth-Order Optimization. arXiv.2405.15861 2024. [Google Scholar]
  11. Maritan, A.; Dey, S.; Schenato, L. FedZeN: Quadratic convergence in zeroth-order federated learning via incremental Hessian estimation. In Proceedings of the 2024 European Control Conference; 2024; pp. 2320–2327. [Google Scholar]
  12. Staib, M.; Reddi, S.; Kale, S.; Kumar, S.; Sra, S. Escaping saddle points with adaptive gradient methods. In Proceedings of the International Conference on Machine Learning; 2019; pp. 5956–5965. [Google Scholar]
  13. Chen, X.; Li, X.; Li, P. Toward communication efficient adaptive gradient method. In Proceedings of the Proceedings of the 2020 ACM-IMS on Foundations of Data Science Conference, 2020, pp. 119–128.
  14. Reddi, S.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečnỳ, J.; Kumar, S.; McMahan, H.B. Adaptive federated optimization. arXiv:2003.00295, 2020. [Google Scholar]
  15. Wang, Y.; Lin, L.; Chen, J. Communication-efficient adaptive federated learning. In Proceedings of the International Conference on Machine Learning; 2022; pp. 22802–22838. [Google Scholar]
  16. Wang, J.; Liu, Q.; Liang, H.; Joshi, G.; Poor, H.V. A novel framework for the analysis and design of heterogeneous federated learning. IEEE Transactions on Signal Processing 2021, 69, 5234–5249. [Google Scholar] [CrossRef]
  17. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proceedings of Machine learning and systems 2020, 2, 429–450. [Google Scholar]
  18. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. In Proceedings of the International conference on machine learning; 2020; pp. 5132–5143. [Google Scholar]
  19. Pathak, R.; Wainwright, M.J. FedSplit: An algorithmic framework for fast federated optimization. Advances in neural information processing systems 2020, 33, 7057–7066. [Google Scholar]
  20. Zhang, X.; Hong, M.; Dhople, S.; Yin, W.; Liu, Y. Fedpd: A federated learning framework with adaptivity to non-iid data. IEEE Transactions on Signal Processing 2021, 69, 6055–6070. [Google Scholar] [CrossRef]
  21. Wang, S.; Roosta, F.; Xu, P.; Mahoney, M.W. Giant: Globally improved approximate newton method for distributed optimization. Advances in Neural Information Processing Systems 2018, 31. [Google Scholar]
  22. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smithy, V. Feddane: A federated newton-type method. In Proceedings of the 2019 53rd Asilomar Conference on Signals, Systems, and Computers; 2019; pp. 1227–1231. [Google Scholar]
  23. Xu, A.; Huang, H. Coordinating momenta for cross-silo federated learning. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2022, Vol. 36, pp. 8735–8743.
  24. Das, R.; Acharya, A.; Hashemi, A.; Sanghavi, S.; Dhillon, I.S.; Topcu, U. Faster non-convex federated learning via global and local momentum. In Proceedings of the Uncertainty in Artificial Intelligence; 2022; pp. 496–506. [Google Scholar]
  25. Khanduri, P.; Sharma, P.; Yang, H.; Hong, M.; Liu, J.; Rajawat, K.; Varshney, P. Stem: A stochastic two-sided momentum algorithm achieving near-optimal sample and communication complexities for federated learning. Advances in Neural Information Processing Systems 2021, 34, 6050–6061. [Google Scholar]
  26. Jiang, W.; Han, H.; Zhang, Y.; Mu, J. Federated split learning for sequential data in satellite-terrestrial integrated networks. Information Fusion 2024, 103, 102141. [Google Scholar] [CrossRef]
  27. Jiang, W.; Han, H.; Zhang, Y.; Mu, J.; Shankar, A. Intrusion Detection with Federated Learning and Conditional Generative Adversarial Network in Satellite-Terrestrial Integrated Networks. Mobile Networks and Applications, 2024, pp. 1–14.
  28. Tang, Y.; Zhang, J.; Li, N. Distributed zero-order algorithms for nonconvex multiagent optimization. IEEE Transactions on Control of Network Systems 2020, 8, 269–281. [Google Scholar] [CrossRef]
  29. Nikolakakis, K.; Haddadpour, F.; Kalogerias, D.; Karbasi, A. Black-box generalization: Stability of zeroth-order learning. Advances in neural information processing systems 2022, 35, 31525–31541. [Google Scholar]
  30. Balasubramanian, K.; Ghadimi, S. Zeroth-order (non)-convex stochastic optimization via conditional gradient and gradient updates. Advances in Neural Information Processing Systems 2018, 31. [Google Scholar]
  31. Mhanna, E.; Assaad, M. Rendering Wireless Environments Useful for Gradient Estimators: A Zero-Order Stochastic Federated Learning Method. arXiv.2401.17460 2024. [Google Scholar]
  32. Diederik, P.K. Adam: A method for stochastic optimization 2014.
  33. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research 2011, 12. [Google Scholar]
  34. Ling, X.; Fu, J.; Wang, K.; Liu, H.; Chen, Z. Ali-dpfl: Differentially private federated learning with adaptive local iterations. In Proceedings of the 2024 IEEE 25th International Symposium on a World of Wireless, Mobile and Multimedia Networks; 2024; pp. 349–358. [Google Scholar]
  35. Cong, Y.; Qiu, J.; Zhang, K.; Fang, Z.; Gao, C.; Su, S.; Tian, Z. Ada-FFL: Adaptive computing fairness federated learning. CAAI Transactions on Intelligence Technology 2024, 9, 573–584. [Google Scholar] [CrossRef]
  36. Huang, Y.; Zhu, S.; Chen, W.; Huang, Z. FedAFR: Enhancing Federated Learning with adaptive feature reconstruction. Computer Communications 2024, 214, 215–222. [Google Scholar] [CrossRef]
  37. Li, Y.; He, Z.; Gu, X.; Xu, H.; Ren, S. AFedAvg: Communication-efficient federated learning aggregation with adaptive communication frequency and gradient sparse. Journal of Experimental & Theoretical Artificial Intelligence 2024, 36, 47–69. [Google Scholar]
  38. Gao, X.; Jiang, B.; Zhang, S. On the information-adaptive variants of the ADMM: an iteration complexity perspective. Journal of Scientific Computing 2018, 76, 327–363. [Google Scholar] [CrossRef]
  39. Wu, X.; Huang, F.; Hu, Z.; Huang, H. Faster adaptive federated learning. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2023, Vol. 37, pp. 10379–10387.
Figure 1. Example of the MNIST dataset.
Figure 1. Example of the MNIST dataset.
Preprints 146881 g001
Figure 2. Federated black-box attack accuracy against the number of communication rounds.
Figure 2. Federated black-box attack accuracy against the number of communication rounds.
Preprints 146881 g002
Figure 3. Influence of the number of local updates that select Q = 50 and M = 50 .
Figure 3. Influence of the number of local updates that select Q = 50 and M = 50 .
Preprints 146881 g003
Figure 4. Influence of the number of participating edge devices that select Q = 50 and E = 1 .
Figure 4. Influence of the number of participating edge devices that select Q = 50 and E = 1 .
Preprints 146881 g004
Figure 5. Influence of the number of local updates that select Q = 50 and M = 50 .
Figure 5. Influence of the number of local updates that select Q = 50 and M = 50 .
Preprints 146881 g005
Figure 6. Influence of the number of directions that select Q = 50 and M = 10 .
Figure 6. Influence of the number of directions that select Q = 50 and M = 10 .
Preprints 146881 g006
Figure 7. The impact of different numbers of local updates when selecting Q = 50 and M = 30 in a non-IID setting.
Figure 7. The impact of different numbers of local updates when selecting Q = 50 and M = 30 in a non-IID setting.
Preprints 146881 g007
Figure 8. The impact of the number of participating edge devices when selecting Q = 50 and E = 1 under the non-IID setting.
Figure 8. The impact of the number of participating edge devices when selecting Q = 50 and E = 1 under the non-IID setting.
Preprints 146881 g008
Figure 9. The impact of different numbers of local updates when selecting Q = 50 and M = 50 under the non-IID setting.
Figure 9. The impact of different numbers of local updates when selecting Q = 50 and M = 50 under the non-IID setting.
Preprints 146881 g009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated