Improving Mental Health Through Multimodal Emotion Detection from Speech and Text Data Using Long-Short Term Memory

  • Dhritesh Bhagat
  • , Aritra Ray
  • , Adarsh Sarda
  • , Nilanjana Dutta Roy*
  • , Mufti Mahmud
  • , Debashis De
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

9 Scopus citations

Abstract

In today’s world of cut-throat competition, where everyone is running an invisible race, we often find ourselves alone amongst the crowd. The advancements in technology are making our lives easier, yet man being a social animal is losing touch with society. As a result, today a huge part of the population is suffering from psychological disorders. Inferiority complex, inability to fulfil dreams, loneliness, etc., are considered to be the common reasons to disturb mental stability, which may further lead to disorders like depression. In extreme cases, depression causes loss of precious lives when an individual decides to commit suicide. Assessing an individual’s mental health in an interactive way with the core help of machine learning is the primary focus of this work. To realize this objective, we have used the most suitable long-short term memory (LSTM) architecture. It is an artificial recurrent neural network (RNN) in the field of deep learning on Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) and FastText datasets to get 86% accuracy when fed with model-patient conversational data. Further, we discussed the scope of enhancing cognitive control capabilities over the psychiatric disorders, which may even lead to severe level of depression and suicidal attacks. Here, the proposed system will help to determine the severity level of depression in a person and will help with the recovery process. The system comprises of a wrist-band to measure some biological parameters, a headband to analyse the mental health and a user-friendly website and mobile application which has an in-built chatbot. AI-based chatbot will talk to the patients and help them reveal their thoughts, which they are otherwise not able to communicate to their peers. A person can chat via text message, which is to be stored in the database for further analysis. The novelty of this work is in the sentiment analysis of voice chat, which therefore creates a comfortable environment for the user.

Original languageEnglish
Title of host publicationFrontiers of ICT in Healthcare - Proceedings of EAIT 2022
EditorsJyotsna Kumar Mandal, Debashis De
PublisherSpringer Science and Business Media Deutschland GmbH
Pages13-23
Number of pages11
ISBN (Print)9789811951909
DOIs
StatePublished - 2023
Externally publishedYes
Event7th International Conference on Emerging Applications of Information Technology, EAIT 2022 - kolkata, India
Duration: 27 Mar 202228 Mar 2022

Publication series

NameLecture Notes in Networks and Systems
Volume519 LNNS
ISSN (Print)2367-3370
ISSN (Electronic)2367-3389

Conference

Conference7th International Conference on Emerging Applications of Information Technology, EAIT 2022
Country/TerritoryIndia
Citykolkata
Period27/03/2228/03/22

Bibliographical note

Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

Keywords

  • AI-supported chatbot
  • Depression prediction
  • Emotion analysis
  • Mental health
  • Neural network
  • Sentiment

ASJC Scopus subject areas

  • Control and Systems Engineering
  • Signal Processing
  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'Improving Mental Health Through Multimodal Emotion Detection from Speech and Text Data Using Long-Short Term Memory'. Together they form a unique fingerprint.

Cite this