IDEAS home Printed from https://ideas.repec.org/h/spr/prochp/978-3-031-56576-2_12.html
   My bibliography  Save this book chapter

Determining Emotion Intensities from Audio Data Using a Convolutional Neural Network

In: Artificial Intelligence Tools and Applications in Embedded and Mobile Systems

Author

Listed:
  • Simon Kipyatich Kiptoo

    (Jomo Kenyatta University of Agriculture and Technology)

  • Kennedy Ogada

    (Jomo Kenyatta University of Agriculture and Technology)

  • Tobias Mwalili

    (Jomo Kenyatta University of Agriculture and Technology)

Abstract

Human beings communicate their feelings in the form of emotions. The feelings are expressed via speech, facial expressions, gestures, or other non-verbal signs. An emotion is a complex severe disturbance of an individual’s mental state that involves a subjective experience coupled with physiological, behavioral, and expressive responses. It denotes the mental state of the human mind and thought processes that represent a recognizable pattern. The emotions can be expressed with normal intensity or strong intensity depending on the incident being communicated. Detecting and classifying these emotions encompasses three fundamental machine learning processes; Feature Extraction, Feature Selection, and Feature Classification. Machine learning is the science of making computers learn and act like humans, while enhancing the learning with time in an independent manner, by giving them data and information in the form of observations and real-world interactions. A multi-modal approach comprising several machine learning algorithms is required to map out the intensities contained in the emotion classes. Mel Frequency Cepstral Coefficients are a set of about 10–20 features obtained from a speech signal describing the overall shape of a spectral envelope. The MFCC algorithm is designed to mimic the human hearing, thus it is ideal for this experiment.

Suggested Citation

  • Simon Kipyatich Kiptoo & Kennedy Ogada & Tobias Mwalili, 2024. "Determining Emotion Intensities from Audio Data Using a Convolutional Neural Network," Progress in IS, in: Jorge Marx Gómez & Anael Elikana Sam & Devotha Godfrey Nyambo (ed.), Artificial Intelligence Tools and Applications in Embedded and Mobile Systems, pages 125-138, Springer.
  • Handle: RePEc:spr:prochp:978-3-031-56576-2_12
    DOI: 10.1007/978-3-031-56576-2_12
    as

    Download full text from publisher

    To our knowledge, this item is not available for download. To find whether it is available, there are three options:
    1. Check below whether another version of this item is available online.
    2. Check on the provider's web page whether it is in fact available.
    3. Perform a search for a similarly titled item that would be available.

    More about this item

    Keywords

    Emotion; Emotion intensity; CNN; MFCC;
    All these keywords.

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:prochp:978-3-031-56576-2_12. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.