IDEAS home Printed from https://ideas.repec.org/a/spr/jglopt/v76y2020i4d10.1007_s10898-019-00842-6.html
   My bibliography  Save this article

A class of exact penalty functions and penalty algorithms for nonsmooth constrained optimization problems

Author

Listed:
  • Qian Liu

    (Shandong Normal University)

  • Yuqing Xu

    (Shandong Normal University)

  • Yang Zhou

    (Shandong Normal University)

Abstract

In this paper, a class of smoothing penalty functions is proposed for optimization problems with equality, inequality and bound constraints. It is proved exact, under the condition of weakly generalized Mangasarian–Fromovitz constraint qualification, in the sense that each local optimizer of the penalty function corresponds to a local optimizer of the original problem. Furthermore, necessary and sufficient conditions are discussed for the inverse proposition of exact penalization. Based on the theoretical results in this paper, a class of smoothing penalty algorithms with feasibility verification is presented. Theories on the penalty exactness, feasibility verification and global convergence of the proposed algorithm are presented. Numerical results show that this algorithm is effective for nonsmooth nonconvex constrained optimization problems.

Suggested Citation

  • Qian Liu & Yuqing Xu & Yang Zhou, 2020. "A class of exact penalty functions and penalty algorithms for nonsmooth constrained optimization problems," Journal of Global Optimization, Springer, vol. 76(4), pages 745-768, April.
  • Handle: RePEc:spr:jglopt:v:76:y:2020:i:4:d:10.1007_s10898-019-00842-6
    DOI: 10.1007/s10898-019-00842-6
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10898-019-00842-6
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10898-019-00842-6?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Changyu Wang & Cheng Ma & Jinchuan Zhou, 2014. "A new class of exact penalty functions and penalty algorithms," Journal of Global Optimization, Springer, vol. 58(1), pages 51-73, January.
    2. Mengwei Xu & Jane Ye & Liwei Zhang, 2015. "Smoothing augmented Lagrangian method for nonsmooth constrained optimization problems," Journal of Global Optimization, Springer, vol. 62(4), pages 675-694, August.
    3. Changyu Wang & Qian Liu & Biao Qu, 2017. "Global saddle points of nonlinear augmented Lagrangian functions," Journal of Global Optimization, Springer, vol. 68(1), pages 125-146, May.
    4. NESTEROV, Yu., 2005. "Smooth minimization of non-smooth functions," LIDAM Reprints CORE 1819, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Suxia Ma & Yuelin Gao & Bo Zhang & Wenlu Zuo, 2022. "A New Nonparametric Filled Function Method for Integer Programming Problems with Constraints," Mathematics, MDPI, vol. 10(5), pages 1-16, February.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Lei Zhang & Yang Liu & Jianneng Chen & Heng Zhou & Yunsheng Jiang & Junhua Tong & Lianlian Wu, 2023. "Trajectory Synthesis and Optimization Design of an Unmanned Five-Bar Vegetable Factory Packing Machine Based on NSGA-II and Grey Relation Analysis," Agriculture, MDPI, vol. 13(7), pages 1-21, July.
    2. Dirk Lorenz & Marc Pfetsch & Andreas Tillmann, 2014. "An infeasible-point subgradient method using adaptive approximate projections," Computational Optimization and Applications, Springer, vol. 57(2), pages 271-306, March.
    3. Jueyou Li & Zhiyou Wu & Changzhi Wu & Qiang Long & Xiangyu Wang, 2016. "An Inexact Dual Fast Gradient-Projection Method for Separable Convex Optimization with Linear Coupled Constraints," Journal of Optimization Theory and Applications, Springer, vol. 168(1), pages 153-171, January.
    4. Guoyin Li & Alfred Ma & Ting Pong, 2014. "Robust least square semidefinite programming with applications," Computational Optimization and Applications, Springer, vol. 58(2), pages 347-379, June.
    5. Masaru Ito, 2016. "New results on subgradient methods for strongly convex optimization problems with a unified analysis," Computational Optimization and Applications, Springer, vol. 65(1), pages 127-172, September.
    6. TAYLOR, Adrien B. & HENDRICKX, Julien M. & François GLINEUR, 2016. "Exact worst-case performance of first-order methods for composite convex optimization," LIDAM Discussion Papers CORE 2016052, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    7. Dimitris Bertsimas & Nishanth Mundru, 2021. "Sparse Convex Regression," INFORMS Journal on Computing, INFORMS, vol. 33(1), pages 262-279, January.
    8. Amir Beck & Shoham Sabach, 2015. "Weiszfeld’s Method: Old and New Results," Journal of Optimization Theory and Applications, Springer, vol. 164(1), pages 1-40, January.
    9. Donghwan Kim & Jeffrey A. Fessler, 2021. "Optimizing the Efficiency of First-Order Methods for Decreasing the Gradient of Smooth Convex Functions," Journal of Optimization Theory and Applications, Springer, vol. 188(1), pages 192-219, January.
    10. Nguyen Thai An & Nguyen Mau Nam & Xiaolong Qin, 2020. "Solving k-center problems involving sets based on optimization techniques," Journal of Global Optimization, Springer, vol. 76(1), pages 189-209, January.
    11. Manlio Gaudioso & Giovanni Giallombardo & Giovanna Miglionico, 2020. "Essentials of numerical nonsmooth optimization," 4OR, Springer, vol. 18(1), pages 1-47, March.
    12. Masoud Ahookhosh, 2019. "Accelerated first-order methods for large-scale convex optimization: nearly optimal complexity under strong convexity," Mathematical Methods of Operations Research, Springer;Gesellschaft für Operations Research (GOR);Nederlands Genootschap voor Besliskunde (NGB), vol. 89(3), pages 319-353, June.
    13. Alexandre Belloni & Victor Chernozhukov & Lie Wang, 2013. "Pivotal estimation via square-root lasso in nonparametric regression," CeMMAP working papers CWP62/13, Centre for Microdata Methods and Practice, Institute for Fiscal Studies.
    14. Chao, Shih-Kang & Härdle, Wolfgang K. & Yuan, Ming, 2021. "Factorisable Multitask Quantile Regression," Econometric Theory, Cambridge University Press, vol. 37(4), pages 794-816, August.
    15. M. V. Dolgopolik, 2018. "Augmented Lagrangian functions for cone constrained optimization: the existence of global saddle points and exact penalty property," Journal of Global Optimization, Springer, vol. 71(2), pages 237-296, June.
    16. DEVOLDER, Olivier & GLINEUR, François & NESTEROV, Yurii, 2013. "First-order methods with inexact oracle: the strongly convex case," LIDAM Discussion Papers CORE 2013016, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    17. Lingxue Zhang & Seyoung Kim, 2014. "Learning Gene Networks under SNP Perturbations Using eQTL Datasets," PLOS Computational Biology, Public Library of Science, vol. 10(2), pages 1-20, February.
    18. David Degras, 2021. "Sparse group fused lasso for model segmentation: a hybrid approach," Advances in Data Analysis and Classification, Springer;German Classification Society - Gesellschaft für Klassifikation (GfKl);Japanese Classification Society (JCS);Classification and Data Analysis Group of the Italian Statistical Society (CLADAG);International Federation of Classification Societies (IFCS), vol. 15(3), pages 625-671, September.
    19. Yunmei Chen & Xiaojing Ye & Wei Zhang, 2020. "Acceleration techniques for level bundle methods in weakly smooth convex constrained optimization," Computational Optimization and Applications, Springer, vol. 77(2), pages 411-432, November.
    20. Silvia Villa & Lorenzo Rosasco & Sofia Mosci & Alessandro Verri, 2014. "Proximal methods for the latent group lasso penalty," Computational Optimization and Applications, Springer, vol. 58(2), pages 381-407, June.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:jglopt:v:76:y:2020:i:4:d:10.1007_s10898-019-00842-6. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.