Publication:
Residual Neural Network precisely quantifies dysarthria severity-level based on short-duration speech segments

dc.contributor.affiliationDA-IICT, Gandhinagar
dc.contributor.authorGupta, Siddhant
dc.contributor.authorPatil, Ankur T
dc.contributor.authorPurohit, Mirali
dc.contributor.authorPatel, Maitreya
dc.contributor.authorGuido, Rodrigo Capobianco
dc.contributor.authorPatil, Hemant
dc.contributor.researcherGupta, Siddhant (201911007)
dc.contributor.researcherPatil, Ankur T (201621008)
dc.contributor.researcherPurohit, Mirali (201811067)
dc.contributor.researcherPatel, Maitreya (201601160)
dc.date.accessioned2025-08-01T13:09:01Z
dc.date.issued01-07-2021
dc.description.abstractRecently, we have witnessed Deep Learning methodologies gaining significant attention for severity-based classification of dysarthric speech. Detecting dysarthria, quantifying its severity, are of paramount importance in various real-life applications, such as the assessment of patients' progression in treatments, which includes an adequate planning of their therapy and the improvement of speech-based interactive systems in order to handle pathologically-affected voices automatically. Notably, current speech-powered tools often deal with short-duration speech segments and, consequently, are less efficient in dealing with impaired speech, even by using Convolutional Neural Networks (CNNs). Thus, detecting dysarthria severity-level based on short speech segments might help in improving the performance and applicability of those systems. To achieve this goal, we propose a novel Residual Network (ResNet)-based technique which receives short-duration speech segments as input. Statistically meaningful objective analysis of our experiments, reported over standard Universal Access corpus, exhibits average values of 21.35% and 22.48% improvement, compared to the baseline CNN, in terms of classification accuracy and F1-score, respectively. For additional comparisons, tests with Gaussian Mixture Models and Light CNNs were also performed. Overall, the values of 98.90% and 98.00% for classification accuracy and F1-score, respectively, were obtained with the proposed ResNet approach, confirming its efficacy and reassuring its practical applicability.
dc.format.extent105-117
dc.identifier.citationSiddhant Gupta, Ankur T. Patil, Mirali Purohit, Maitreya Patel, Patil, Hemant A,, and Rodrigo Capobianco Guido,"Residual Neural Network precisely quantifies dysarthria severity-level based on short-duration speech segments," Neural Networks, Elsevier, vol. 139,pp. 105-117, Jul. 2021. doi:10.1016/j.neunet.2021.02.008.
dc.identifier.doi10.1016/j.neunet.2021.02.008
dc.identifier.issn0893-6080
dc.identifier.scopus2-s2.0-85102061061
dc.identifier.urihttps://ir.daiict.ac.in/handle/dau.ir/1556
dc.identifier.wosWOS:000652750100009
dc.language.isoen
dc.publisherElsevier
dc.relation.ispartofseriesVol. 139; No.
dc.source Neural Networks
dc.source.urihttps://www.sciencedirect.com/science/article/pii/S0893608021000502?via%3Dihub
dc.titleResidual Neural Network precisely quantifies dysarthria severity-level based on short-duration speech segments
dspace.entity.typePublication
relation.isAuthorOfPublicationfdb7041b-280e-498b-b2ee-34f9bc351f4c
relation.isAuthorOfPublication.latestForDiscoveryfdb7041b-280e-498b-b2ee-34f9bc351f4c

Files

Collections