IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v8y2020i10p1821-d430501.html
   My bibliography  Save this article

Binary Whale Optimization Algorithm for Dimensionality Reduction

Author

Listed:
  • Abdelazim G. Hussien

    (Faculty of Science, Fayoum University, Faiyum 63514, Egypt)

  • Diego Oliva

    (IN3-Computer Science Department, Universitat Oberta de Catalunya, 08018 Barcelona, Spain
    Departamento de Ciencias Computacionales, Universidad de Guadalajara, CUCEI, Guadalajara 44430, Mexico)

  • Essam H. Houssein

    (Faculty of Computers and Information, Minia University, Minia 61519, Egypt)

  • Angel A. Juan

    (IN3-Computer Science Department, Universitat Oberta de Catalunya, 08018 Barcelona, Spain)

  • Xu Yu

    (School of Information Science and Technology, Qingdao University of Science and Technology, Qingdao 266061, China)

Abstract

Feature selection (FS) was regarded as a global combinatorial optimization problem. FS is used to simplify and enhance the quality of high-dimensional datasets by selecting prominent features and removing irrelevant and redundant data to provide good classification results. FS aims to reduce the dimensionality and improve the classification accuracy that is generally utilized with great importance in different fields such as pattern classification, data analysis, and data mining applications. The main problem is to find the best subset that contains the representative information of all the data. In order to overcome this problem, two binary variants of the whale optimization algorithm (WOA) are proposed, called bWOA-S and bWOA-V. They are used to decrease the complexity and increase the performance of a system by selecting significant features for classification purposes. The first bWOA-S version uses the Sigmoid transfer function to convert WOA values to binary ones, whereas the second bWOA-V version uses a hyperbolic tangent transfer function. Furthermore, the two binary variants introduced here were compared with three famous and well-known optimization algorithms in this domain, such as Particle Swarm Optimizer (PSO), three variants of binary ant lion (bALO1, bALO2, and bALO3), binary Dragonfly Algorithm (bDA) as well as the original WOA, over 24 benchmark datasets from the UCI repository. Eventually, a non-parametric test called Wilcoxon’s rank-sum was carried out at 5% significance to prove the powerfulness and effectiveness of the two proposed algorithms when compared with other algorithms statistically. The qualitative and quantitative results showed that the two introduced variants in the FS domain are able to minimize the selected feature number as well as maximize the accuracy of the classification within an appropriate time.

Suggested Citation

  • Abdelazim G. Hussien & Diego Oliva & Essam H. Houssein & Angel A. Juan & Xu Yu, 2020. "Binary Whale Optimization Algorithm for Dimensionality Reduction," Mathematics, MDPI, vol. 8(10), pages 1-24, October.
  • Handle: RePEc:gam:jmathe:v:8:y:2020:i:10:p:1821-:d:430501
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/8/10/1821/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/8/10/1821/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Mohammed M. Ahmed & Essam H. Houssein & Aboul Ella Hassanien & Ayman Taha & Ehab Hassanien, 2019. "Maximizing lifetime of large-scale wireless sensor networks using multi-objective whale optimization algorithm," Telecommunication Systems: Modelling, Analysis, Design and Management, Springer, vol. 72(2), pages 243-259, October.
    2. Piramuthu, Selwyn, 2004. "Evaluating feature selection methods for learning in data mining applications," European Journal of Operational Research, Elsevier, vol. 156(2), pages 483-494, July.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Shuang Wang & Abdelazim G. Hussien & Heming Jia & Laith Abualigah & Rong Zheng, 2022. "Enhanced Remora Optimization Algorithm for Solving Constrained Engineering Optimization Problems," Mathematics, MDPI, vol. 10(10), pages 1-32, May.
    2. Rong Zheng & Abdelazim G. Hussien & He-Ming Jia & Laith Abualigah & Shuang Wang & Di Wu, 2022. "An Improved Wild Horse Optimizer for Solving Optimization Problems," Mathematics, MDPI, vol. 10(8), pages 1-30, April.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Cheng-Yu Ho & Ke-Sheng Cheng & Chi-Hang Ang, 2023. "Utilizing the Random Forest Method for Short-Term Wind Speed Forecasting in the Coastal Area of Central Taiwan," Energies, MDPI, vol. 16(3), pages 1-18, January.
    2. Crone, Sven F. & Lessmann, Stefan & Stahlbock, Robert, 2006. "The impact of preprocessing on data mining: An evaluation of classifier sensitivity in direct marketing," European Journal of Operational Research, Elsevier, vol. 173(3), pages 781-800, September.
    3. Lee, In Gyu & Yoon, Sang Won & Won, Daehan, 2022. "A Mixed Integer Linear Programming Support Vector Machine for Cost-Effective Group Feature Selection: Branch-Cut-and-Price Approach," European Journal of Operational Research, Elsevier, vol. 299(3), pages 1055-1068.
    4. Gambella, Claudio & Ghaddar, Bissan & Naoum-Sawaya, Joe, 2021. "Optimization problems for machine learning: A survey," European Journal of Operational Research, Elsevier, vol. 290(3), pages 807-828.
    5. Juheng Zhang & Selwyn Piramuthu, 2018. "Product recommendation with latent review topics," Information Systems Frontiers, Springer, vol. 20(3), pages 617-625, June.
    6. Du, Wen Sheng & Hu, Bao Qing, 2018. "A fast heuristic attribute reduction approach to ordered decision systems," European Journal of Operational Research, Elsevier, vol. 264(2), pages 440-452.
    7. Hashim, Fatma A. & Houssein, Essam H. & Hussain, Kashif & Mabrouk, Mai S. & Al-Atabany, Walid, 2022. "Honey Badger Algorithm: New metaheuristic algorithm for solving optimization problems," Mathematics and Computers in Simulation (MATCOM), Elsevier, vol. 192(C), pages 84-110.
    8. Anzanello, Michel J. & Albin, Susan L. & Chaovalitwongse, Wanpracha A., 2012. "Multicriteria variable selection for classification of production batches," European Journal of Operational Research, Elsevier, vol. 218(1), pages 97-105.
    9. Ding‐Wen Tan & William Yeoh & Yee Ling Boo & Soung‐Yue Liew, 2013. "The Impact Of Feature Selection: A Data‐Mining Application In Direct Marketing," Intelligent Systems in Accounting, Finance and Management, John Wiley & Sons, Ltd., vol. 20(1), pages 23-38, January.
    10. Piccialli, Veronica & Romero Morales, Dolores & Salvatore, Cecilia, 2024. "Supervised feature compression based on counterfactual analysis," European Journal of Operational Research, Elsevier, vol. 317(2), pages 273-285.
    11. Juheng Zhang & Selwyn Piramuthu, 0. "Product recommendation with latent review topics," Information Systems Frontiers, Springer, vol. 0, pages 1-9.
    12. Fouskakis, D., 2012. "Bayesian variable selection in generalized linear models using a combination of stochastic optimization methods," European Journal of Operational Research, Elsevier, vol. 220(2), pages 414-422.
    13. Aytug, Haldun, 2015. "Feature selection for support vector machines using Generalized Benders Decomposition," European Journal of Operational Research, Elsevier, vol. 244(1), pages 210-218.
    14. Bodendorf, Frank & Xie, Qiao & Merkl, Philipp & Franke, Jörg, 2022. "A multi-perspective approach to support collaborative cost management in supplier-buyer dyads," International Journal of Production Economics, Elsevier, vol. 245(C).
    15. Verbeke, Wouter & Dejaeger, Karel & Martens, David & Hur, Joon & Baesens, Bart, 2012. "New insights into churn prediction in the telecommunication sector: A profit driven data mining approach," European Journal of Operational Research, Elsevier, vol. 218(1), pages 211-229.
    16. Bertolazzi, P. & Felici, G. & Festa, P. & Fiscon, G. & Weitschek, E., 2016. "Integer programming models for feature selection: New extensions and a randomized solution algorithm," European Journal of Operational Research, Elsevier, vol. 250(2), pages 389-399.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:8:y:2020:i:10:p:1821-:d:430501. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.