Unsupervised Semantic Change Detection in Arabic

dc.contributor.advisorDubossarsky, Haim
dc.contributor.authorSindi, Kenan
dc.date.accessioned2024-02-01T09:03:46Z
dc.date.available2024-02-01T09:03:46Z
dc.date.issued2023-10-23
dc.description.abstractThis study employs pretrained BERT models— AraBERT, CAMeLBERT (CA), and CAMeLBERT (MSA)—to investigate semantic change in Arabic across distinct time periods. Analyzing word embeddings and cosine distance scores reveals variations in capturing semantic shifts. The research highlights the significance of training data quality and diversity, while acknowledging limitations in data scope. The project's outcome—a list of most stable and changed words—contributes to Arabic NLP by shedding light on semantic change detection, suggesting potential model selection strategies and areas for future exploration.
dc.format.extent15
dc.identifier.urihttps://hdl.handle.net/20.500.14154/71355
dc.language.isoen
dc.publisherQueen Mary University of London
dc.subjectNatural Language Processing
dc.subjectArabic NLP
dc.subjectLangauge Models
dc.subjectBERT
dc.subjectData Science
dc.subjectSemantic Change
dc.subjectUnsupervised
dc.titleUnsupervised Semantic Change Detection in Arabic
dc.typeThesis
sdl.degree.departmentElectronic Engineering and Computer Science
sdl.degree.disciplineData Science and Artificial Intelligence
sdl.degree.grantorQueen Mary University of London
sdl.degree.nameMaster of Science

Files

Copyright owned by the Saudi Digital Library (SDL) © 2024