An Application Using Radial Basis Function Classification in Stress Speech Identification

Authors

  • Dhole NP Department of ETE, PRMITR Badnera, Amravati, India
  • Kale SN Department of Applied Electronics, Sant Gadge Baba Amravati University, Amravati, India

DOI:

https://doi.org/10.26438/ijcse/v6i4.448451

Keywords:

RBF, MFCC, Stress Classification, Feature Selection

Abstract

Speech of human beings is the reflection of the state of mind. Proper evaluation of these speech signals into stress types is necessary in order to ensure that the person is in a healthy state of mind. In this work we propose a RBF classifier for speech stress classification algorithm, with sophisticated feature extraction techniques as Mel Frequency Cepstral Coefficients (MFCC). The RBF algorithm assists the system to learn the speech patterns in real time and self-train itself in order to improve the classification accuracy of the overall system. The proposed system is suitable for real time speech and is language and word independent. The human behaviour considers six basic emotions which are happiness, sadness, anger, fear, surprise & disgust. It becomes important to detect emotional state of a person which will be induced by workload, background noise, physical environmental factors (e.g. G-force) & fatigue. Broadly, stress identification becomes a scientific challenge to analyze a human being interaction with environment.

References

Schuller, Bjorn, et al., “Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first Challenge”, Speech Communication 53.9, pp. 1062-1087, 2011.

Anagnostopoulos, Christos-Nikolaos, Theodoros Iliou, and Ioannis Giannoukos, “Features and Classifiers for Emotion Recognition from Speech: A survey from 2000 to 2011”, Artificial Intelligence Review 43.2, pp.155-177, 2015.

Dipti D. Joshi, M. B. Zalte, “Speech Emotion Recognition: A Review”, Journal of Electronics and Communication Engineering (IOSR-JECE) 4.4, pp.34-37, 2013.

Ververidis, Dimitrios, and Constantine Kotropoulos, “Emotional Speech Recognition: Resources, Features, and Methods”, Speech Communication 48.9, pp.1162-1181, 2006.

El Ayadi, Moataz, Mohamed S. Kamel, and Fakhri Karray, “Survey on Speech Emotion Recognition”, Features, classification schemes, and databases, Pattern Recognition 44.3 pp. 572-587,2011.

Scherer, Klaus R., “Vocal Communication of Emotion: A review of research paradigms”, Speech communication 40.1, pp.227-256, 2003.

Vogt, Thurid, Elisabeth Andre, and Johannes Wagner, “Automatic recognition of emotions from speech: a review of the literature and recommendations for practical realization, Affect and emotion in human-computer interaction”, Springer Berlin Heidelberg, pp. 75-91, 2008.

Burkhardt, Felix, et al., “A Database of German Emotional Speech”, INTER-SPEECH, Lisbon, Portugal, vol. 5, pp.1-4, 2005.

Kwon, Oh-Wook, et al, “Emotion Recognition by Speech Signals, INTER-SPEECH, pp.1-4, 2003.

Campbell, N. “Recording and Storing of Speech Data”. In: Proceedings LREC, pp. 12-25, 2002.

Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M., Schroder, M. Feeltrace, “An Instrument for Recording Perceived Emotion in Real Time”, In: Proceedings of the ISCA Workshop on Speech and Emotion, pp.19-24, 2000.

Devillers, L., Cowie, R., Martin, J.-C., Douglas-Cowie, E., Abrilian, S., McRorie, M.: “Real life emotions in French and English TV video clips: an integrated annotation protocol combining continuous and discrete approaches”, 5th International Conference on Language Resources and Evaluation LREC, Genoa, Italy.2006.

Douglas-Cowie, E., Campbell, N., Cowie, R.P. “Emotional speech: Towards a new generation of databases”, Speech Communication 40(1–2), pp.33-60, 2003.

Douglas-Cowie, E., et al.: “The description of naturally occurring emotional speech”. In: Proceedings of 15th International Congress of Phonetic Sciences, Barcelona, 2003.

http://audacity.sourceforge.net/download.

Broomhead, D. S. Lowe, David “Radial basis functions, multi-variable functional interpolation and adaptive networks”, Technical Report- RSRE 4148, 1988.

Broomhead, D. S. Lowe, David "Multivariable functional interpolation and adaptive networks", Complex Systems 2, pp. 321–355, 1988.

Schwenker, Friedhelm; Kestler, Hans A.; Palm, Gunther, "Three learning phases for radial-basis-function networks", Neural Networks 14, pp. 439–458, 2001.

Park, J., I. W. Sandberg, "Universal Approximation Using Radial-Basis-Function Networks", Neural Computation. 3 (2): pp. 246–257, Retrieved 26 March 2013.

BOGERT, B. P.; HEALY, M. J. R.; TURKEY, J. W.: “The Quefrency Alanysis of Time Series for Echoes: Cepstrum, Pseudo Autocovariance, Cross-Cepstrum and Saphe Cracking”, Proceedings of the Symposium on Time Series Analysis, (M. Rosenblatt, Ed) Chapter 15, New York: Wiley, pp.209-243, 1963.

Downloads

Published

2025-11-12
CITATION
DOI: 10.26438/ijcse/v6i4.448451
Published: 2025-11-12

How to Cite

[1]
N. Dhole and S. Kale, “An Application Using Radial Basis Function Classification in Stress Speech Identification”, Int. J. Comp. Sci. Eng., vol. 6, no. 4, pp. 448–451, Nov. 2025.

Issue

Section

Research Article