Detection of question segment in an audio Arabic lecture

dc.contributor.authorKhan, Omair
dc.date2007
dc.date.accessioned2022-05-18T09:15:40Z
dc.date.available2022-05-18T09:15:40Z
dc.degree.departmentCollege of Computer Science and Engineering
dc.degree.grantorKing Fahad for Petrolem University
dc.description.abstractProsody has been widely used in many speech-related applications including speaker and word recognition, emotion and accent identification, topic and sentence segmentation, and text-to-speech applications. An important application we investigated, is that of identifying question sentences in Arabic monologue lectures. Languages other than Arabic have received a lot of attention in this regard. We approached this problem by first segmenting the sentences from the continuous speech using intensity and duration features. Prosodic features are then extracted from each sentence. These features are used as input to decision trees to classify each sentence into either Question or Non-Question sentence. Our results suggest that questions are redundantly marked in natural Arabic speech and automatically extracted prosodic features can make significant contribution in question identification. We classified Questions with an accuracy of 77.43%. Feature specific analysis further reveals that energy and fundamental frequency (F0) features are mainly responsible for discriminating between question and non-question sentences. We found that Bayes Network performed better than SVM, MLP and Decision Trees on our dataset. Removal of correlated features through Correlations Based Feature Selection (CFS) produced more efficient and accurate results than the complete feature set.
dc.identifier.other6082
dc.identifier.urihttps://drepo.sdl.edu.sa/handle/20.500.14154/3202
dc.language.isoen
dc.publisherSaudi Digital Library
dc.thesis.levelMaster
dc.thesis.sourceKing Fahad for Petrolem University
dc.titleDetection of question segment in an audio Arabic lecture
dc.typeThesis

Files

Copyright owned by the Saudi Digital Library (SDL) © 2025