Author
Listed:
- Zhihao Li
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Qingtao Wu
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China
Intelligent System Science and Technology Innovation Center, Longmen Laboratory, Luoyang 471023, China)
- Moli Zhang
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Lin Wang
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China
Intelligent System Science and Technology Innovation Center, Longmen Laboratory, Luoyang 471023, China)
- Youming Ge
(School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)
- Guoyong Wang
(School of Computer and Information Engineering, Luoyang Institute of Science and Technology, Luoyang 471023, China)
Abstract
Multi-objective optimization (MOO) has become an important method in machine learning, which involves solving multiple competing objective problems simultaneously. Nowadays, many MOO algorithms assume that gradient information is easily available and use this information to optimize functions. However, when encountering situations where gradients are not available, such as black-box functions or non-differentiable functions, these algorithms become ineffective. In this paper, we propose a zeroth-order MOO algorithm named SZMG (stochastic zeroth-order multi-gradient algorithm), which approximates the gradient of functions by finite difference methods. Meanwhile, to avoid conflicting gradients between functions and reduce stochastic multi-gradient direction bias caused by stochastic gradients, an SGD-type method is adopted to acquire weight parameters. Under the non-convex setting and mild assumptions, the convergence rate is established for the SZMG algorithm. Simulation results demonstrate the effectiveness of the SZMG algorithm.
Suggested Citation
Zhihao Li & Qingtao Wu & Moli Zhang & Lin Wang & Youming Ge & Guoyong Wang, 2025.
"Stochastic Zeroth-Order Multi-Gradient Algorithm for Multi-Objective Optimization,"
Mathematics, MDPI, vol. 13(4), pages 1-31, February.
Handle:
RePEc:gam:jmathe:v:13:y:2025:i:4:p:627-:d:1591536
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:13:y:2025:i:4:p:627-:d:1591536. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.