Speech

On-Line Speaker Indexing

Unsupervised speaker indexing sequentially detects points where a speaker identity changes in a multi-speaker audio stream, and categorizes each speaker segment, without any prior knowledge about the speakers.

[Read More]

Implicit Pronunciation Modeling for Speech Recognition Using Syllable-Centric Models

Speech recognition is an essential component of any human computer interaction (HCI) scheme, which aspires to be natural.



[Read More]

Hierarchical Speech Recognition

Speech recognition is an essential component of any Human Computer Interaction (HCI) scheme, which aspires to be natural. Thus, high accuracy speech recognition is of critical importance in making natural man-machine interfaces.

[Read More]

Expressive Speech Synthesis and Modeling

Emotions (anger, happiness, sadness, etc.) are inseparable components of the natural human speech. Because of that, the level of human speech can only be achieved with the ability to synthesize emotions.


[Read More]

ChIMP

Spoken language adds naturalness and efficiency to human-machine interactions with both children and adults.


[Read More]

Compression for Speech Recognition and Music Classification

One of the goals of this project is to develop methods for compressing speech signals for a distributed speech recognition task.

[Read More]

Automatic Recognition of Emotions from the Acoustic Speech Signal

This research aims at investigating several feature sets such as acoustic, lexical, and discourse features, and classification algorithms for classifying spoken utterances based on the emotional state of the speaker.

[Read More]