A method for recognising a speaker based on speech features is known as speaker recognition. Speaker recognition technology is frequently employed in a variety of fields. Most speaker identification algorithms have been developed on regular, clean recordings, but their effectiveness degrades when hearing speech with emotions. This study offers an emotional speech signals system developed using standard machine learning algorithms on an emotional speech database obtained from the University Audio visual Database of Affective Speech and Song using temporal, frequency, and spectral properties Five models (Logistic Regression, Support Vector Machine, Variational Forest, XGBoost, and k-Nearest Neighbor) and three deep learning models (Logistic Regression, Logistic Regression, Random Forest, XGBoost, and k-Nearest Neighbor) were trained and compared in terms of performance (Long Short-Term Memory network, Multilayer Perceptron, and Convolutional Neural Network).Deep neural networks outperformed state-of-the-art models