Saudi Cultural Missions Theses & Dissertations

Permanent URI for this communityhttps://drepo.sdl.edu.sa/handle/20.500.14154/10

Browse

Search Results

Now showing 1 - 10 of 43
  • ItemRestricted
    Deep Learning Approaches for Multivariate Time Series: Advances in Feature Selection, Classification, and Forecasting
    (New Mexico State University, 2024) Alshammari, Khaznah Raghyan; Tran, Son; Hamdi, Shah Muhammad
    In this work, we present the latest developments and advancements in the machine learning-based prediction and feature selection of multivariate time series (MVTS) data. MVTS data, which involves multiple interrelated time series, presents significant challenges due to its high dimensionality, complex temporal dependencies, and inter-variable relationships. These challenges are critical in domains such as space weather prediction, environmental monitoring, healthcare, sensor networks, and finance. Our research addresses these challenges by developing and implementing advanced machine-learning algorithms specifically designed for MVTS data. We introduce innovative methodologies that focus on three key areas: feature selection, classification, and forecasting. Our contributions include the development of deep learning models, such as Long Short-Term Memory (LSTM) networks and Transformer-based architectures, which are optimized to capture and model complex temporal and inter-parameter dependencies in MVTS data. Additionally, we propose a novel feature selection framework that gradually identifies the most relevant variables, enhancing model interpretability and predictive accuracy. Through extensive experimentation and validation, we demonstrate the superior performance of our approaches compared to existing methods. The results highlight the practical applicability of our solutions, providing valuable tools and insights for researchers and practitioners working with high-dimensional time series data. This work advances the state of the art in MVTS analysis, offering robust methodologies that address both theoretical and practical challenges in this field.
    14 0
  • ItemRestricted
    TOWARDS ROBUST AND ACCURATE TEXT-TO-CODE GENERATION
    (University of Central Florida, 2024) almohaimeed, saleh; Wang, Liqiang
    Databases play a vital role in today’s digital landscape, enabling effective data storage, manage- ment, and retrieval for businesses and other organizations. However, interacting with databases often requires knowledge of query (e.g., SQL) and analysis, which can be a barrier for many users. In natural language processing, the text-to-code task, which converts natural language text into query and analysis code, bridges this gap by allowing users to access and manipulate data using everyday language. This dissertation investigates different challenges in text-to-code (including text-to-SQL as a subtask), with a focus on four primary contributions to the field. As a solution to the lack of statistical analysis in current text-to-code tasks, we introduce SIGMA, a text-to- Code dataset with statistical analysis, featuring 6000 questions with Python code labels. Baseline models show promising results, indicating that our new task can support both statistical analysis and SQL queries simultaneously. Second, we present Ar-Spider, the first Arabic cross-domain text-to-SQL dataset that addresses multilingual limitations. We have conducted experiments with LGESQL and S2SQL models, enhanced by our Context Similarity Relationship (CSR) approach, which demonstrates competitive performance, reducing the performance gap between the Arabic and English text-to-SQL datasets. Third, we address context-dependent text-to-SQL task, often overlooked by current models. The SParC dataset was explored by utilizing different question rep- resentations and in-context learning prompt engineering techniques. Then, we propose GAT-SQL, an advanced prompt engineering approach that improves both zero-shot and in-context learning experiments. GAT-SQL sets new benchmarks in both SParC and CoSQL datasets. Finally, we introduce Ar-SParC, a context-dependent Arabic text-to-SQL dataset that enables users to interact with the model through a series of interrelated questions. In total, 40 experiments were conducted to investigate this dataset using various prompt engineering techniques, and a novel technique called GAT Corrector was developed, which significantly improved the performance of all base- line models.
    11 0
  • ItemRestricted
    ADAPTIVE INTRUSION DETECTION SYSTEM FOR THE INTERNET OF MEDICAL THINGS (IOMT): ENHANCING SECURITY THROUGH IMPROVED MUTUAL INFORMATION FEATURE SELECTION AND META-LEARNING
    (Towson University, 2024-12) Alalhareth, Mousa; Hong, Sungchul
    The Internet of Medical Things (IoMT) has revolutionized healthcare by enabling continuous patient monitoring and diagnostics but also introduces significant cybersecurity risks. IoMT devices are vulnerable to cyber-attacks that threaten patient data and safety. To address these challenges, Intrusion Detection Systems (IDS) using machine learning algorithms have been introduced. However, the high data dimensionality in IoMT environments often leads to overfitting and reduced detection accuracy. This dissertation presents several methodologies to enhance IDS performance in IoMT. First, the Logistic Redundancy Coefficient Gradual Upweighting Mutual Information Feature Selection (LRGU-MIFS) method is introduced to balance the trade-off between relevance and redundancy, while improving redundancy estimation in cases of data sparsity. This method achieves 95% accuracy, surpassing the 92% reported in related studies. Second, a fuzzy-based self-tuning Long Short-Term Memory (LSTM) IDS model is proposed, which dynamically adjusts training epochs and uses early stopping to prevent overfitting and underfitting. This model achieves 97% accuracy, a 10% false positive rate, and a 94% detection rate, outperforming prior models that reported 95% accuracy, a 12% false positive rate, and a 93% detection rate. Finally, a performance-driven meta-learning technique for ensemble learning is introduced. This technique dynamically adjusts classifier voting weights based on factors such as accuracy, loss, and prediction confidence levels. As a result, this method achieves 98% accuracy, a 97% detection rate, and a 99% F1 score, while reducing the false positive rate to 10%, surpassing previous results of 97% accuracy, a 93% detection rate, a 97% F1 score, and an 11% false positive rate. These contributions significantly enhance IDS effectiveness in IoMT, providing stronger protection for sensitive medical data and improving the security and reliability of healthcare networks.
    8 0
  • ItemRestricted
    Automatic Detection and Verification System for Arabic Rumor News on Twitter
    (University of Technology Sydney, 2026-04) Karali, Sami; Chin-Teng, Lin
    Language models have been extensively studied and applied in various fields in recent years. However, the majority of the language use models are designed for and perform significantly better in English compared to other languages, such as Arabic. The differences between English and Arabic in terms of grammar, writing, and word-forming structures pose significant challenges in applying English-based language models to Arabic content. Therefore, there is a critical need to develop and refine models and methodologies that can effectively process Arabic content. This research aims to address the gaps in Arabic language models by developing innovative machine learning (ML) and natural language processing (NLP) methodologies. We apply the developed model to Arabic rumor detection on Twitter to test its effectiveness. To achieve this, the research is divided into three fundamental phases: 1) Efficiently collecting and pre-processing a comprehensive dataset of Arabic news tweets; 2) The refinement of ML models through an enhanced Convolutional Neural Network (ECNN) equipped with N-gram feature maps for accurate rumor identification; 3) The augmentation of decision-making precision in rumor verification via sophisticated ensemble learning techniques. In the first phase, the research meticulously develops a methodology for the collection and pre-processing of Arabic news tweets, aiming to establish a dataset optimized for rumor detection analysis. Leveraging a blend of automated and manual processes, the research navigates the intricacies of the Arabic language, enhancing the dataset’s quality for ML applications. This foundational phase ensures removing irrelevant data and normalizing text, setting a precedent for accuracy in subsequent detection tasks. The second phase is to develop an Enhanced Convolutional Neural Network (ECNN) model, which incorporates N-gram feature maps for a deeper linguistic analysis of tweets. This innovative ECNN model, designed specifically for the Arabic language, marks a significant departure from traditional rumor detection models by harnessing the power of spatial feature extraction alongside the contextual insights provided by N-gram analysis. Empirical results underscore the ECNN model’s superior performance, demonstrating a marked improvement in detecting and classifying rumors with heightened accuracy and efficiency. The culmination of the study explores the efficacy of ensemble learning methods in enhancing the robustness and accuracy of rumor detection systems. By synergizing the ECNN model with Long Short-Term Memory (LSTM), Bidirectional LSTM (BiLSTM), and Gated Recurrent Unit (GRU) networks within a stacked ensemble framework, the research pioneers a composite approach that significantly outstrips the capabilities of singular models. This innovation results in a state-of-the-art system for rumor verification that outperforms accuracy in identifying rumors, as demonstrated by empirical testing and analysis. This research contributes to bridging the gap between English-centric language models and Arabic language processing, demonstrating the importance of tailored approaches for different languages in the field of ML and NLP. These contributions signify a monumental step forward in the field of Arabic NLP and ML and offer practical solutions for the real-world challenge of rumor proliferation on social media platforms, ultimately fostering a more reliable digital environment for Arabic-speaking communities.
    17 0
  • ItemRestricted
    A Peer-to-Peer Federated Learning Framework for Intrusion Detection in Autonomous Vehicles
    (Lancaster University, 2024-09) Alotaibi, Bassam; Bradbury, Matthew
    As autonomous vehicles (AVs) increasingly rely on interconnected systems for enhanced functionality, they also face heightened cyberattack vulnerability. This study introduces a decentralized peer-to-peer federated learning framework to improve intrusion detection in AV environments while preserving data privacy. A novel soft-reordering one-dimensional Convolutional Neural Network (SR-1CNN) is proposed as the detection engine, capable of identifying known and unknown threats with high accuracy. The framework allows vehicles to communicate directly in a mesh topology, sharing model parameters asynchronously, thus eliminating dependency on centralized servers and mitigating single points of failure. The SR-1CNN model was tested on two datasets: NSL-KDD and Car Hacking, under both independent and non-independent data distribution scenarios. The results demonstrate the model’s robustness, achieving detection accuracies of 94.39% on the NSL-KDD dataset and 99.97% on the Car Hacking dataset in independent settings while maintaining strong performance in non-independent configurations. These findings underline the framework’s potential to enhance cybersecurity in AV networks by addressing data heterogeneity and preserving user privacy. This research contributes to the field of AV security by offering a scalable, privacy-conscious intrusion detection solution. Future work will focus on optimizing the SR-1CNN architecture, exploring vertical federated learning approaches, and validating the framework in real-world autonomous vehicle environments to ensure its practical applicability and scalability.
    23 0
  • ItemRestricted
    Enhancing a Hyper-parameter Tuning of Convolutional Neural Network Model for Brain Tumor Classification using Whale Optimization and Grey Wolf Optimizer
    (Newcastle University, 2024) Alkhudair, Haifa; Freitas, Leo
    Brain tumors represent a global health issue, with about 11 new cases per 100,000 people annually. Therefore, it is crucial to develop faster and more accurate diagnostic solutions. This study develops and evaluates a convolutional neural network (CNN) model optimized using the Whale Optimization Algorithm (WOA) and Grey Wolf Optimizer (GWO) for classifying brain tumors. To achieve that, this work involved collecting and preprocessing an MRI brain tumor dataset, followed by building and training CNN models. Hyperparameters were optimized using WOA and GWO, and the performance of these optimized mod- els was compared against a non-optimized CNN. The WOA-optimized CNN outperformed both the non-optimized and GWO-optimized mod- els, achieving an accuracy of 93.4% and demonstrating superior general- ization across different classes. This study underscores the effectiveness of WOA in enhancing CNN models for medical image classification, of- fering promising approaches to enhancing the accuracy and reliability of brain tumor classification
    9 0
  • ItemRestricted
    Generative AI for Mitosis Synthesis in Histopathology Images
    (University of Surrey, 2024-09) Alkhadra, Rahaf; Rai, Taran; Wells, Kevin
    Identifying mitotic figures has been established as an effective method of fighting cancer at its most vulnerable stage. Traditional methods rely on manual, slow, and invasive detection methods obtained from sectioned tissue samples to acquire histopathological images. Currently, Artificial Intelligence (AI) in oncology has produced a paradigm shift in the fight against cancer, also known as computational oncology. This is heavily reliant on the availability of mitotic figure datasets to train models; however, such datasets are limited in size, type, and may infringe on patient privacy. It is hypothesised that the potential of computational oncology can be realised by synthesising realistic and diverse histopathological datasets using Generative Artificial Intelligence (GenAI). This report demonstrates a comparison of Denoising Probabilistic Diffusion Models (DDPM) and StyleGAN3 in generating synthetic histopathology images, with mitotic figures. The MIDOG++ dataset containing human and canine samples with 7 types of tumours was used to train the models. Quality and similarity of generated and real images was evaluated using as Frechet Inception Distance (FID), Mean Square Error (MSE), Structural Similarity Index (SSIM), and Area Under the Curve (AUC) as a part of Receiver Operating Characteristic (ROC) study were incorporated. Our results suggests that the DDPM model is superior in terms of structural accuracy, however, StyleGAN3 capture the colour scheme better.
    19 0
  • ItemRestricted
    Automated Pain Assessment Through Facial Expression Using Deep Learning and Image Processing
    (University of Reading, 2024-09-13) Alsama, Morady; Patino, Luis
    As pain is an unavoidable part of life, this study examines the use of facial expression tech nology in assisting individuals with pain. Accurate pain assessment in health care is essential, especially for non-verbal patients, since conventional methods largely fail because of the in herent subjectivity and self-reporting. Therefore, the present study develops and evaluates an automated pain assessment system through advanced analysis of facial expressions driven by contemporary deep learning techniques. It aims to generate a reliable and unbiased system for detecting and classifying pain intensity. A CNN-based system was developed using base models that apply ResNet-18 and ResNext-50 architectures. A custom-designed final layer was added to optimize classification accuracy, tailored explicitly for pain detection. Comprehensive data preprocessing strategies were used in the model to make it robust; it involved downsam pling and augmentation of the data. It was trained and validated on the UNBC-McMaster Shoulder Pain Expression Archive Database and the Radboud Faces Database, showing an impressive accuracy of over 90% on the training data. However, generalizing the models to unseen validation and test data proved challenging. These findings further articulate the crit ical imperative of enhancing generalisability across diverse patient populations for the system to perform effectively in real-world settings. The results underline the huge potential for deep learning in the automation of pain assessment, while future research remains on better mod eling generalization, promoting integration in clinical settings for a more objective, reliable, and consistent approach to pain management in health care settings.
    14 0
  • ItemRestricted
    Enhancing DDoS attack Detection using Machine Learning and Deep Learning Models
    (University of Warwick, 2023-09-26) AlObaidan, Fatimah; Raza, Hassan
    Technology has become an essential part of our daily lives, indispensable for both individuals and enterprises. It facilitates the exchange of an extensive range of information across different spaces. However, Internet security is a critical challenge in today's digital age with growing dependence on IT services. Thus, various network environments can be vulnerable to attacks, causing resource depletion and hindering support for legitimate users. One of these attacks is the Distributed Denial of Service (DDoS) attack. The nature of this type of attack is such that it impacts the availability of the system. The impact to confidentiality is primary due to threat actors using DDoS as method to create chaos whilst lunching cyber-attacks on other part of infrastructures. Therefore, it is essential that DDoS attacks required sharper focus from a research perspective. The network intrusion detection system (NIDSs) are important tool to detect and monitor the network environment from DDoS attacks. However, NIDS tools suffer from several limitation such as detecting new attack and misclassified attacks. Therefore, Machine Learning (ML) and Deep Learning (DL) models are increasingly being used for automated detection of DDoS attacks. While several related works deployed ML for NIDS, most of these approaches ignore the appropriate pre-processing and overfitting problem during the implementation of ML algorithms. As a result, it can impact the robustness of the anomaly detection system and lead to poor model performance for zero-day attacks. In this research study, the researcher is proposing a new ML and DL approach based on hybrid feature selection and appropriate pre-processing operation to classify the network flow into normal or DDoS attacks. The results of the experiments carried out by researcher suggest the efficiency and the reliability of the proposed lightweight models in achieving high detection rate while minimising the detection time with less number of features. This project complies with following two CyBOK Skills areas: Network Security: The project evaluates the network security and introduces efficient, lightweight models for DDoS attack detection. Security Operations and Incident Management: The project enhances incident management capabilities by crafting ML that monitors network flows within NIDS.
    11 0
  • ItemRestricted
    Assessing artificial intelligence MRI autocontouring in Raystation and the AutoConfidence uncertainty model for brain radiotherapy
    (The University of Leeds, 2024-10) Alzahrani, Nouf; Henry, Ann; Nix, Michael; Murray, Louise; Al-qaisieh, Bashar
    Abstract: Background: In radiotherapy, deep learning autosegmentation (DL-AS) and automation of quality assurance (QA) have the potential to efficiently standardize and enhance the quality of contours. Aim: To assess the performance of DL-AS in delineating organs-at-risk (OARs) in brain RT using the RayStation Treatment Planning System. Secondly, to build and test a novel artificial intelligence QA model called AutoConfidence (ACo). Methods: Retrospective MRI and CT cases were randomly selected for training and testing. DL-AS models were evaluated from geometric and dosimetric perspectives, focusing on the impact of pre-training editing. The ACo model was evaluated using two sources of autosegmentation: internal autosegmentations (IAS) produced from the ACo generator and two external DL-AS with different qualities (high and low quality) produced from RayStation models. Results: The edited DL-AS models generated more segmentations than the unedited models. Editing pituitary, orbits, optic nerves, lenses, and optic chiasm on MRI before training significantly improved at least one geometry metric. MRI-based DL-AS performed worse than CT-based in delineating the lacrimal gland, whereas the CT-based performed worse in delineating the optic chiasm. Except for the right orbit, when delineated using MRI models, the dosimetric statistical analysis revealed no superior model in terms of the dosimetric accuracy between the MR and CT DL-AS models. The number of patients where the clinical significance threshold was exceeded was higher for the optic chiasm D1% than for other OARs, for all models. ACo had excellent performance on both internal and external segmentations across all OARs (except lenses). Mathews Correlation Coefficient was higher on IAS and low-quality external segmentations than high-quality ones. Conclusion: MRI DL-AS in RT may improve consistency, quality, and efficiency but requires careful editing of training contours. ACo was a reliable predictor of uncertainty and errors on DL-AS, demonstrating its potential as an independent, reference-free QA tool.
    6 0

Copyright owned by the Saudi Digital Library (SDL) © 2024