Repository logo
  • English
  • العربية
  • বাংলা
  • Català
  • Čeština
  • Deutsch
  • Ελληνικά
  • Español
  • Suomi
  • Français
  • Gàidhlig
  • हिंदी
  • Magyar
  • Italiano
  • Қазақ
  • Latviešu
  • Nederlands
  • Polski
  • Português
  • Português do Brasil
  • Srpski (lat)
  • Српски
  • Svenska
  • Türkçe
  • Yкраї́нська
  • Tiếng Việt
Log In
New user? Click here to register.Have you forgotten your password?
  1. Home
  2. Browse by Subject

Browsing by Subject "1D Cnn"

Filter results by typing the first few letters
Now showing 1 - 1 of 1
  • Results Per Page
  • Sort Options
  • Loading...
    Thumbnail Image
    Some of the metrics are blocked by your 
    consent settings
    Publication
    Enhancing Speech Recognition in Noisy Environments Through Multiscale-Multichannel Feature Learning and Adaptive Noise Resilience
    (Proceedings of the 2024 3rd Edition of IEEE Delhi Section Flagship Conference, Delcon 2024, 2024)
    Sengottaiyan, N  
    ;
    Sathishkumar, S
    ;
    Rajesh Sharma, R  
    ;
    Sungheetha, Akey  
    ;
    Ghantasala, G S Pradeep
    ;
    Karthika, K
    Speech Emotion Recognition (SER) has seen much research done recently, but little is being done to minimize the effect of environmental noise on the predictions. Existing SER models primarily aim to learn the best feature representations of speech from clean datasets while neglecting the pragmatic constraint when it is deployed in adverse acoustic settings. The variations in signal-to-noise ratios, background noises and acoustic interferences make existing SER systems less robust. Here, we focus on a new method that combines adaptive feature learning and noise-resilient techniques using a multiscale-multichannel framework. The objective is to help common SER models generalize better in different acoustic conditions, thus improving the accuracy and robustness of SER in noisy real-life settings. Therefore, these SER models are still not efficient against challenging acoustic conditions and the research area lacks addressing it => thus we utilize state-of-the-art IDCNNs and also explore novel data augmentations to help develop SER systems with great capability for arbitrary acoustical channels that fills the gap of very limited related works till now. © 2024 IEEE.

Powered by - Informatics Publishing Ltd