Show simple item record

dc.contributor.advisorNguyen, Thi Thanh Sang
dc.contributor.authorNguyen, Phu Vinh
dc.date.accessioned2024-03-15T03:20:30Z
dc.date.available2024-03-15T03:20:30Z
dc.date.issued2021
dc.identifier.urihttp://keep.hcmiu.edu.vn:8080/handle/123456789/4568
dc.description.abstractFor speech analysis, the process involves learning and recognizing the various features of an audio clip that can be used to analyze a language. This procedure works by converting audio files to spectrograms and then applying a neural network to learn and recognize the various features of the audio. The main objective of this thesis is to analyze the languages out of the various speakers that were recorded in the Mozilla’s Common Voice1 and VIVOS Corpus2 dataset. The recordings were analyzed by recording 10 seconds of each utterance. The datasets are then split into training and test sets. The results of these tests reveal an overall accuracy of 99%.en_US
dc.language.isoenen_US
dc.subjectDeep learningen_US
dc.titleDeep Learning For Speech Analysis And Evaluationen_US
dc.typeThesisen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record