Parameterizing Human Speech Generation

Show simple item record

dc.contributor.author Perepichka, Nazariy
dc.date.accessioned 2020-01-29T11:28:20Z
dc.date.available 2020-01-29T11:28:20Z
dc.date.issued 2020
dc.identifier.citation Perepichka, Nazariy. Parameterizing Human Speech Generation : Master Thesis : manuscript rights / Nazariy Perepichka ; Supervisor Diego Saez-Trumper ; Ukrainian Catholic University, Department of Computer Sciences. – Lviv : [s.n.], 2020. – 35 p. : ill. uk
dc.identifier.uri http://er.ucu.edu.ua/handle/1/1908
dc.language.iso en uk
dc.subject Text-to-speech uk
dc.subject Speech synthesis uk
dc.subject Voice parameterizing en
dc.subject Emotion recognition
dc.subject Sentiment classification
dc.title Parameterizing Human Speech Generation uk
dc.type Preprint uk
dc.status Публікується вперше uk
dc.description.abstracten In modern days synthesis of human images and videos is arguably one of the most popular topics in the Data Science community. The synthesis of human speech is less trendy but deeply bonded to the mentioned topic. Since the publication of WaveNet paper by Google researchers in 2016, the state-of-the-art approach transferred from parametric and concatenative systems to deep learning models. Most of the work on the area focuses on improving the intelligibility and naturalness of the speech. However, almost every significant study also mentions ways to generate speech with the voices of different speakers. Usually, such an enhancement requires the model’s re-training in case of generating audio with the voice of a speaker that was not present in the training set. Additionally,studies focused on highly modular speech generation are rare. There fore there is a room left for research on ways to add new parameters for other aspects of the speech, like sentiment, prosody, and melody. In this work, we aimed to implement a competitive text-to-speech solution with the ability to specify the speaker without model re-training and explore possibilities for adding emotions to the generated speech. Our approach generates good quality speech with the mean opinion score of 3,78 (out of 5) points and the ability to mimic speaker voice in real-time, which is a significant improvement over the baseline that merely obtains 2,08. On top of that, we researched sentiment representation possibilities. We built an emotion classifier that performs on the level of the current state of the art solutions by giving an accuracy of more than eighty percent. uk


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search


Browse

My Account