ResearchSpace

A cross-platform interface for automatic speaker identification and verification

Show simple item record

dc.contributor.author Mokgonyane, TB
dc.contributor.author Sefara, Tshephisho J
dc.contributor.author Manamela, MJ
dc.contributor.author Modipa, TI
dc.date.accessioned 2021-10-07T07:08:31Z
dc.date.available 2021-10-07T07:08:31Z
dc.date.issued 2021-08
dc.identifier.citation Mokgonyane, T., Sefara, T.J., Manamela, M. & Modipa, T. 2021. A cross-platform interface for automatic speaker identification and verification. http://hdl.handle.net/10204/12123 . en_ZA
dc.identifier.isbn 978-1-7281-8592-7
dc.identifier.isbn 978-1-7281-8591-0
dc.identifier.isbn 978-1-7281-8593-4
dc.identifier.uri DOI: 10.1109/icABCD51485.2021.9519322
dc.identifier.uri http://hdl.handle.net/10204/12123
dc.description.abstract The task of automatically identifying and/or verifying the identity of a speaker from a recording of a speech sample, known as automatic speaker recognition, has been studied for many years and automatic speaker recognition technologies have improved recently and becoming inexpensive and reliable methods for identifying and verifying people. Although automatic speaker recognition research has now spanned over 50 years, there is not adequate research done with regards to low-resourced South African indigenous languages. In this paper, a multi-layer perceptron (MLP) classifier model is trained and deployed on a graphical user interface for real time identification and verification of Sepedi native speakers. Sepedi is a low-resourced language spoken by the majority of residents in the Limpopo province of South Africa. The data used to train the speaker recognition system is obtained from the NCHLT (National Centre for Human Language Technology) project. A total of 34 short-term acoustic features of speech are extracted with the use of py Audio Analysis library and Sklearn is used to train the MLP classifier model which performs well with an accuracy of 95%. The GUI is developed with QT Creator and PyQT4 and it has obtained a true acceptance rate (TAR) of 66.67% and a true rejection rate of (TRR) 13.33%. en_US
dc.format Fulltext en_US
dc.language.iso en en_US
dc.relation.uri https://ieeexplore.ieee.org/document/9519322 en_US
dc.source 2021 International Conference on Artificial Intelligence, Big Data, Computing and Data Communication Systems (icABCD), Durban, South Africa, 5-6 August 2021 en_US
dc.subject Automatic speaker recognition en_US
dc.subject Text-dependent en_US
dc.subject Text-independent en_US
dc.subject Graphical user interface en_US
dc.subject Multi-layer perceptron en_US
dc.title A cross-platform interface for automatic speaker identification and verification en_US
dc.type Conference Presentation en_US
dc.description.pages 6 en_US
dc.description.note Copyright: IEEE 2021. The attached pdf contains the accepted version of the published item. en_US
dc.description.cluster Next Generation Enterprises & Institutions en_US
dc.description.impactarea Data Science en_US
dc.identifier.apacitation Mokgonyane, T., Sefara, T. J., Manamela, M., & Modipa, T. (2021). A cross-platform interface for automatic speaker identification and verification. http://hdl.handle.net/10204/12123 en_ZA
dc.identifier.chicagocitation Mokgonyane, TB, Tshephisho J Sefara, MJ Manamela, and TI Modipa. "A cross-platform interface for automatic speaker identification and verification." <i>2021 International Conference on Artificial Intelligence, Big Data, Computing and Data Communication Systems (icABCD), Durban, South Africa, 5-6 August 2021</i> (2021): http://hdl.handle.net/10204/12123 en_ZA
dc.identifier.vancouvercitation Mokgonyane T, Sefara TJ, Manamela M, Modipa T, A cross-platform interface for automatic speaker identification and verification; 2021. http://hdl.handle.net/10204/12123 . en_ZA
dc.identifier.ris TY - Conference Presentation AU - Mokgonyane, TB AU - Sefara, Tshephisho J AU - Manamela, MJ AU - Modipa, TI AB - The task of automatically identifying and/or verifying the identity of a speaker from a recording of a speech sample, known as automatic speaker recognition, has been studied for many years and automatic speaker recognition technologies have improved recently and becoming inexpensive and reliable methods for identifying and verifying people. Although automatic speaker recognition research has now spanned over 50 years, there is not adequate research done with regards to low-resourced South African indigenous languages. In this paper, a multi-layer perceptron (MLP) classifier model is trained and deployed on a graphical user interface for real time identification and verification of Sepedi native speakers. Sepedi is a low-resourced language spoken by the majority of residents in the Limpopo province of South Africa. The data used to train the speaker recognition system is obtained from the NCHLT (National Centre for Human Language Technology) project. A total of 34 short-term acoustic features of speech are extracted with the use of py Audio Analysis library and Sklearn is used to train the MLP classifier model which performs well with an accuracy of 95%. The GUI is developed with QT Creator and PyQT4 and it has obtained a true acceptance rate (TAR) of 66.67% and a true rejection rate of (TRR) 13.33%. DA - 2021-08 DB - ResearchSpace DP - CSIR J1 - 2021 International Conference on Artificial Intelligence, Big Data, Computing and Data Communication Systems (icABCD), Durban, South Africa, 5-6 August 2021 KW - Automatic speaker recognition KW - Text-dependent KW - Text-independent KW - Graphical user interface KW - Multi-layer perceptron LK - https://researchspace.csir.co.za PY - 2021 SM - 978-1-7281-8592-7 SM - 978-1-7281-8591-0 SM - 978-1-7281-8593-4 T1 - A cross-platform interface for automatic speaker identification and verification TI - A cross-platform interface for automatic speaker identification and verification UR - http://hdl.handle.net/10204/12123 ER - en_ZA
dc.identifier.worklist 24962 en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record