Documentos de conferencia
URI permanente para esta colección
Examinar
Examinando Documentos de conferencia por Autor "Bastidas, M."
Mostrando 1 - 2 de 2
Resultados por página
Opciones de ordenación
Ítem Double Fourier analysis for Emotion Identification in Voiced Speech(IOP PUBLISHING LTD, 2016-01-01) Sierra-Sosa, D.; Bastidas, M.; Ortiz, P.D.; Quintero, O.L.; Sierra-Sosa, D.; Bastidas, M.; Ortiz, P.D.; Quintero, O.L.; Universidad EAFIT. Departamento de Ciencias; Modelado MatemáticoWe propose a novel analysis alternative, based on two Fourier Transforms for emotion recognition from speech. Fourier analysis allows for display and synthesizes different signals, in terms of power spectral density distributions. A spectrogram of the voice signal is obtained performing a short time Fourier Transform with Gaussian windows, this spectrogram portraits frequency related features, such as vocal tract resonances and quasi-periodic excitations during voiced sounds. Emotions induce such characteristics in speech, which become apparent in spectrogram time-frequency distributions. Later, the signal time-frequency representation from spectrogram is considered an image, and processed through a 2-dimensional Fourier Transform in order to perform the spatial Fourier analysis from it. Finally features related with emotions in voiced speech are extracted and presented.Ítem Multiresolution analysis (discrete wavelet transform) through Daubechies family for emotion recognition in speech.(IOP PUBLISHING LTD, 2016-01-01) Campo, D.; Quintero, O.L.; Bastidas, M.; Campo, D.; Quintero, O.L.; Bastidas, M.; Universidad EAFIT. Departamento de Ciencias; Modelado MatemáticoWe propose a study of the mathematical properties of voice as an audio signal. This work includes signals in which the channel conditions are not ideal for emotion recognition. Multiresolution analysis- discrete wavelet transform - was performed through the use of Daubechies Wavelet Family (Db1-Haar, Db6, Db8, Db10) allowing the decomposition of the initial audio signal into sets of coefficients on which a set of features was extracted and analyzed statistically in order to differentiate emotional states. ANNs proved to be a system that allows an appropriate classification of such states. This study shows that the extracted features using wavelet decomposition are enough to analyze and extract emotional content in audio signals presenting a high accuracy rate in classification of emotional states without the need to use other kinds of classical frequency-time features. Accordingly, this paper seeks to characterize mathematically the six basic emotions in humans: boredom, disgust, happiness, anxiety, anger and sadness, also included the neutrality, for a total of seven states to identify.