Skip to content

A machine learning project that uses deep learning to recognize emotions in Urdu speech. Implements CNN, LSTM, and hybrid models to classify emotional states from audio signals with up to 93.75% accuracy. Utilizes advanced audio feature extraction and preprocessing techniques.

License

Notifications You must be signed in to change notification settings

ranauzairahmed/UrduSpeechEmotions

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

Emotion Recognition from Urdu Speech Audio

Overview

This project focuses on developing an emotion recognition system for Urdu speech. The process begins with organizing the audio dataset and extracting essential features such as MFCCs, Chroma, and Zero Crossing Rate. Data augmentation techniques, including time stretching and pitch shifting, are applied to enhance the dataset. Machine learning models, including CNN, LSTM, and a hybrid CNN-LSTM, are trained and evaluated on the extracted features. The evaluation metrics include accuracy, confusion matrices, and classification reports to analyze model performance.

Dataset

URDU-Dataset: https://github.com/siddiquelatif/URDU-Dataset

Features

Audio emotion classification

Multiple neural network models: CNN, LSTM, CNN-LSTM

93.75% peak accuracy

Technologies:

Python | TensorFlow | Librosa | scikit-learn | Pandas | NumPy | Matplotlib | Seaborn

About

A machine learning project that uses deep learning to recognize emotions in Urdu speech. Implements CNN, LSTM, and hybrid models to classify emotional states from audio signals with up to 93.75% accuracy. Utilizes advanced audio feature extraction and preprocessing techniques.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published