Author
Listed:
- Ping Xie
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Xiangrui Gao
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Fan Li
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Ling Xing
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Yu Zhang
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Hanxiao Sun
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
Abstract
Federated learning has become a prevalent distributed training paradigm, in which local devices collaboratively train learning models without exchanging local data. One of the most dominant frameworks of federated learning (FL) is FedAvg, since it is efficient and simple to implement; here, the first-order information is generally utilized to train the parameters of learning models. In practice, however, the gradient information may be unavailable or infeasible in some applications, such as federated black-box optimization problems. To solve the issue, we propose an innovative zeroth-order adaptive federated learning algorithm without using the gradient information, referred to as ZO-AdaFL, which integrates the zeroth-order optimization algorithm into the adaptive gradient method. Moreover, we also rigorously analyze the convergence behavior of ZO-AdaFL in a non-convex setting, i.e., where ZO-AdaFL achieves convergence to a region close to a stationary point at a speed of O ( 1 / T ) ( T represents the total iteration number). Finally, to verify the performance of ZO-AdaFL, simulation experiments are performed using the MNIST and FMNIST datasets. Our experimental findings demonstrate that ZO-AdaFL outperforms other state-of-the-art zeroth-order FL approaches in terms of both effectiveness and efficiency.
Suggested Citation
Ping Xie & Xiangrui Gao & Fan Li & Ling Xing & Yu Zhang & Hanxiao Sun, 2024.
"Communication-Efficient Zeroth-Order Adaptive Optimization for Federated Learning,"
Mathematics, MDPI, vol. 12(8), pages 1-21, April.
Handle:
RePEc:gam:jmathe:v:12:y:2024:i:8:p:1148-:d:1373834
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:12:y:2024:i:8:p:1148-:d:1373834. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.