IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v12y2024i20p3245-d1500265.html
   My bibliography  Save this article

Action Recognition in Videos through a Transfer-Learning-Based Technique

Author

Listed:
  • Elizabeth López-Lozada

    (Centro de Investigación en Computación, Instituto Politécnico Nacional, Ciudad de México 07738, Mexico)

  • Humberto Sossa

    (Centro de Investigación en Computación, Instituto Politécnico Nacional, Ciudad de México 07738, Mexico)

  • Elsa Rubio-Espino

    (Centro de Investigación en Computación, Instituto Politécnico Nacional, Ciudad de México 07738, Mexico)

  • Jesús Yaljá Montiel-Pérez

    (Centro de Investigación en Computación, Instituto Politécnico Nacional, Ciudad de México 07738, Mexico)

Abstract

In computer vision, human action recognition is a hot topic, popularized by the development of deep learning. Deep learning models typically accept video input without prior processing and train them to achieve recognition. However, conducting preliminary motion analysis can be beneficial in directing the model training to prioritize the motion of individuals with less priority for the environment in which the action occurs. This paper puts forth a novel methodology for human action recognition based on motion information that employs transfer-learning techniques. The proposed method comprises four stages: (1) human detection and tracking, (2) motion estimation, (3) feature extraction, and (4) action recognition using a two-stream model. In order to develop this work, a customized dataset was utilized, comprising videos of diverse actions (e.g., walking, running, cycling, drinking, and falling) extracted from multiple public sources and websites, including Pexels and MixKit. This realistic and diverse dataset allowed for a comprehensive evaluation of the proposed method, demonstrating its effectiveness in different scenarios and conditions. Furthermore, the performance of seven pre-trained models for feature extraction was evaluated. The models analyzed were Inception-v3, MobileNet-v2, MobileNet-v3-L, VGG-16, VGG-19, Xception, and ConvNeXt-L. The results demonstrated that the ConvNeXt-L model yielded the most optimal outcomes. Furthermore, using pre-trained models for feature extraction facilitated the training process on a personal computer with a single graphics processing unit, achieving an accuracy of 94.9%. The experimental findings and outcomes suggest that integrating motion information enhances action recognition performance.

Suggested Citation

  • Elizabeth López-Lozada & Humberto Sossa & Elsa Rubio-Espino & Jesús Yaljá Montiel-Pérez, 2024. "Action Recognition in Videos through a Transfer-Learning-Based Technique," Mathematics, MDPI, vol. 12(20), pages 1-17, October.
  • Handle: RePEc:gam:jmathe:v:12:y:2024:i:20:p:3245-:d:1500265
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/12/20/3245/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/12/20/3245/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:12:y:2024:i:20:p:3245-:d:1500265. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.