Faheem Ullah, Dr. Muhammad Irfan Khattak, Muhammad Israr, Khushal Khan, Naveed Ur Rehman, Muhammad Zia
Speaker Recognition is the work out obligation of authorizing a user’s demanded identity using physiognomies removed from their voices. This skill is one of the supreme valuable and standard biometric recognition practices in the world chiefly linked to zones in which security is a foremost concern. It can be used for confirmation, investigation, surveillance, reconnaissance, authentication, forensic speaker recognition and a numeral of associated accomplishments. Speaker recognition can be categorized into identification and verification. Speaker identification is the technique of influencing which registered speaker delivers an assumed utterance. Speaker verification, in contrast, is the technique of accepting or discarding the identity claim of a speaker. The progression of Speaker recognition involves of two segments i.e. feature extraction and feature classifying. Feature extraction is the method in which we extract a minor expanse of data from the voice signal that can be used in future to indicate each speaker. Feature classifying is the procedure of familiarize the system with features. Our proposed work consists of feature extraction from the voices of speakers through Mel frequency Cepstral Coefficients (MFCC) and classifying them by Cartesian Genetic Programing (CGP) to get an efficient logic gates circuit and by Cartesian Genetic Programing Evolved Artificial Neural Network (CGP-E-ANN) to develop an efficient and novel system for speaker verification.
Faheem Ullah Dr. Muhammad Irfan Khattak Muhammad Israr Khushal Khan Naveed Ur Rehman Muhammad Zia “Speaker Verification System Based On Cartesian Genetic Programming (CGP) and Cartesia International Journal of Engineering Works Vol. 9 Issue 09 PP. 166-172 September 2022. https://doi.org/10.34259/ijew.22.909166172.
[1] M. Israr, A. Jan, F. Ullah, and F. Hayat, “Development of a Novel System for Speaker Verification.”
[2] G. Di?ken, Z. Tüfekçi, L. Saribulut, and U. Çevik, “A Review on Feature Extraction for Speaker Recognition under Degraded Conditions,” IETE Tech. Rev. (Institution Electron. Telecommun. Eng. India), vol. 34, no. 3, pp. 321–332, 2017.
[3] D. Reynolds, “Automatic speaker recognition: Current approaches and future trends,” Speak. Verif. From Res. to Real., 2001.
[4] L. Feng, “Speaker Recognition,” 2004.
[5] Y. Zheng, “Text-Independent Speaker Recognition System,” 2012.
[6] R. G. Ayestarán, “Text-Independent Speaker Identification Supervisor,” 2008.
[7] A. Nagraniy, J. S. Chungy, and A. Zisserman, “VoxCeleb: A large-scale speaker identification dataset,” Proc. Annu. Conf. Int. Speech Commun. Assoc. INTERSPEECH, vol. 2017-Augus, pp. 2616–2620, 2017.
[8] S. S. Tirumala, S. R. Shahamiri, A. S. Garhwal, and R. Wang, “Speaker identification features extraction methods: A systematic review,” Expert Syst. Appl., vol. 90, pp. 250–271, 2017.
[9] Z. Wu et al., “ASVspoof: The automatic speaker verification spoofing and countermeasures challenge,” IEEE J. Sel. Top. Signal Process., vol. 11, no. 4, pp. 588–604, 2017.
[10] W. Xiong, L. Wu, F. Alleva, J. Droppo, X. Huang, and A. Stolcke, “The Microsoft 2017 Conversational Speech Recognition System,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. - Proc., vol. 2018-April, pp. 5934–5938, 2018.
[11] S. Diarization, “Speaker Diarization and Identification,” 2012.
[12] R. S. Agrawal and U. N. Agrawal, “A review on emotion recognition using hybrid classifier,” Spec. issue Natl. Conf. Recent Adv. Technol. Manag. Integr. growth 2013 (RATMIG 2013), no. Icicct, 2013.
[13] M. Swain, A. Routray, and P. Kabisatpathy, “Databases, features and classifiers for speech emotion recognition: a review,” Int. J. Speech Technol., vol. 21, no. 1, pp. 93–120, 2018.
[14] A. Churi, A. Bhat, R. Mohite, and P. P. Churi, “E-zip: An electronic lock for secured system,” 2016 IEEE Int. Conf. Adv. Electron. Commun. Comput. Technol. ICAECCT 2016, vol. 2, no. 3, pp. 45–49, 2017.
[15] L. Wan, Q. Wang, A. Papir, and I. L. Moreno, “Generalized end-to-end loss for speaker verification,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. - Proc., vol. 2018-April, pp. 4879–4883, 2018.
[16] A. J. Turner and J. F. Miller, “Recurrent Cartesian Genetic Programming of Artificial Neural Networks,” Genet. Program. Evolvable Mach., vol. 18, no. 2, pp. 185–212, 2017.
[17] M. M. Khan, G. M. Khan, and J. F. Miller, “Evolution of neural networks using Cartesian Genetic Programming,” 2010 IEEE World Congr. Comput. Intell. WCCI 2010 - 2010 IEEE Congr. Evol. Comput. CEC 2010, pp. 1–8, 2010.
[18] A. M. Ahmad and G. M. Khan, “Bio-signal processing using cartesian genetic programming evolved artificial neural networks (CGPANN),” Proc. - 10th Int. Conf. Front. Inf. Technol. FIT 2012, pp. 261–268, 2012.
[19] R. Mukherjee, “Speaker Recognition Using Shifted MFCC,” no. January, 2012.
[20] J. Cao, T. Zhao, J. Wang, R. Wang, and Y. Chen, “Excavation equipment classification based on improved MFCC features and ELM,” Neurocomputing, vol. 261, pp. 231–241, 2017.
[21] J. C. Wang, C. Y. Wang, Y. H. Chin, Y. T. Liu, E. T. Chen, and P. C. Chang, “Spectral-temporal receptive fields and MFCC balanced feature extraction for robust speaker recognition,” Multimed. Tools Appl., vol. 76, no. 3, pp. 4055–4068, 2017.
[22] V. Tiwari, “MFCC and its applications in speaker recognition,” Int. J. Emerg. Technol., vol. 1, no. 1, pp. 19–22, 2010.
[23] I. Nunes and D. Hernane, “Arti fi cial Neural Networks.”
[24] ??? and ????, “Cartesian Genetic Programming ???????????????????,” ?????????, vol. 5, no. 3, pp. 45–52, 2014