The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
In this paper we shall present recent results of two applications for monitoring using acoustical signal classification. The first case study is the problem of context awareness based on acoustic analysis for a service robot. Then we discussed the acoustic classification for wildlife intruder detection. Previous results are briefly recalled and new experimental results are also provided.
In this paper, two models, the I-vector and the Gaussian Mixture Model-Universal Background Model (GMM-UBM), are compared for the speaker identification task. Four feature combinations of I-vectors with seven fusion techniques are considered: maximum, mean, weighted sum, cumulative, interleaving and concatenated for both two and four features. In addition, an Extreme Learning Machine (ELM) is exploited...
The vulnerability of automatic speaker verification (ASV) systems against spoofing attacks is an important security concern about the reliability of ASV technology. Recently, various countermeasures have been developed for spoofing detection. In this paper, we propose to use features derived from linear prediction (LP) residual signal for spoofing detection using simple Gaussian mixture model (GMM)...
Physiological and behavioural human characteristics are exploited in biometrics and performance metrics are used to measure some characteristic of an individual. The measure might lead to a one-to-one match, which is called authentication or one-from-N, and a match represents identification. In this paper, we exploit a speech biometric I-vector with low and fixed dimension of 100 to identify speakers...
Speech uttered by the human beings contains the information about speakers, languages and contents. Language of uttered speech can easily be identified by extracting the language specific information from it. Identification of language of speech is known as Language Identification (LID). Identification of language from speech is helpful in its translation, speech recognition and speech activated automatic...
Voice-based biometric systems are highly prone to spoofing attacks. Recently, various countermeasures have been developed for detecting different kinds of attacks such as replay, speech synthesis (SS) and voice conversion (VC). Most of the existing studies are conducted with a specific training set defined by the evaluation protocol. However, for realistic scenarios, selecting appropriate training...
The performance of a speaker verification system is severely degraded by spoofing attacks generated from artificial speech synthesizers. Recently, several approaches have been proposed for classifying natural and synthetic speech (spoof detection) which can be used in conjunction with a speaker verification system. In this paper, we attempt to develop a joint modelling approach which can detect the...
Emotions exhibited by a speaker can be detected by analyzing his/her speech, facial expressions and gestures or by combining these properties. This paper concentrates on determining the emotional state from speech signals. Various acoustic features such as energy, zero crossing rate(ZCR), fundamental frequency, Mel Frequency Cepstral Coefficients (MFCCs), etc are extracted for short term, overlapping...
Speech is natural vocalized and primary means of communication. Speech is easy, hand-free, fast and do not require any technical knowledge. Communicating with computer using speech is simple and comfortable way for human being. Speech recognition system made this possible. The acoustic and language model for this system are available but mostly in English language. In India there are so many peoples...
Identification of musical instruments from the acoustic signal using speech signal processing methods is a challenging problem. Further, whether this identification can be carried out by a single musical note, like humans are able to do, is an interesting research issue that has several potential applications in the music industry. Attempts have been made earlier using the spectral and temporal features...
This paper presents a text-dependent speaker verification using Mel-Frequency Cepstral Coefficients (MFCC) and Support Vector Machine (SVM). Mel-Frequency Cepstral Coefficients technique has been used to extract the characteristic from the recorded voice spoken by the user and SVM is used to classify the all models of the speakers and impostors. A Malay spoken digit database is utilized for the training...
Phonetic Engine (PE) is a system that is used to determine the sequence of phones in a spoken utterance. In order to transcribe the speech database, International Phonetic Alphabet (IPA) is used. This work focuses on developing multilingual PE for four Indian languages namely, Bengali, Hindi, Urdu and Telugu. The number of languages can be increased to any number. For developing the PE, read speech...
This work projects the importance of phonetic match between train and test session for a text-independent framework under limited test data condition. The robustness of text-independent speaker verification (SV) tends to fall down with the reduction of the amount of speech involved. From a deployable application oriented system point of view, the amount of speech involved, is expected to be less to...
Emotion recognition plays a significant role in affective computing and adds value to machine intelligence. While the emotional state of a person can be manifested in different ways such as facial expressions, gestures, movements and postures, recognition of emotion from speech has gathered much interest over others. However, after years of research, recognizing the emotional state of individuals...
The current work presents a multilingual speech-to-text conversion system. Conversion is based on information in speech signal. Speech is the natural and most important form of communication for human being. Speech-To-Text (STT) system takes a human speech utterance as an input and requires a string of words as output. The objective of this system is to extract, characterize and recognize the information...
The paper presents speaker verification results for six basic emotional states. The database of emotional speech (six acted states: anger, sadness, happiness, fear, disgust, surprise) plus the neutral state were examined with a typical speaker verification system based on MFCC features and GMM classifiers. The obtained results were confronted with the subjective and objective emotion recognition scores...
Acoustic vocalizations are common in marine mammals which can be used for classification purposes. Pinnipeds are a group of carnivore mammals composed by seals, sea lions, and walruses. But although, there is a great interest in research literature about acoustic monitoring of marine mammals, the identification of pinnipeds trough experts systems has been poorly studied. This paper brings a novel...
In this paper, we present our work on speech-smile/shaking vowels classification. An efficient classification system would be a first step towards the estimation (from speech signals only) of amusement levels beyond smile, as indeed shaking vowels represent a transition from smile to laughter superimposed to speech. A database containing examples of both classes has been collected from acted and spontaneous...
This paper develops an algorithm “Discrete Wavelet Transform with Adaptive Filter” (DWTAF) to transform Neutral speech into emotional speech like Angry, Happy or Sad and this is compared with two other emotion transformation algorithms. The other two algorithms are “Speech Transformation using Statistical Parameters and Pitch Contours” (STSPPC) and “Speech Transformation using Mel Frequency Cepstral...
In this paper, classifying and indexing hierarchical video genres using Support Vector Machines (SVMs) are based on only audio features. In fact, segmentation parameters are extracted at block levels, which have a major benefit by capturing local temporal information. The main contribution of our study is to present a powerful combination between the two employed audio descriptors; Mel Frequency Cepstral...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.