Saudi Cultural Missions Theses & Dissertations
Permanent URI for this communityhttps://drepo.sdl.edu.sa/handle/20.500.14154/10
Browse
17 results
Search Results
Item Restricted Human Activity Monitoring for Telemedicine Using an Intelligent Millimeter-Wave System(University of Dayton, 2025) Alhazmi, Abdullah; Chodavarapu, VamsyThe growing aging population requires innovative solutions in the healthcare industry. Telemedicine is one such innovation that can improve healthcare access and delivery to diverse and aging populations. It uses various sensors to facilitate remote monitoring of physiological measures of people, such as heart rate, oxygen saturation, blood glucose, and blood pressure. Similarly, it is capable of monitoring critical events, such as falls. The key challenges in telemonitoring are ensuring accurate remote monitoring of physical activity or falls by preserving privacy and avoiding excessive reliance on expensive and/or obtrusive devices. Our approach initially addressed the need for secure, portable, and low-cost solutions specifically for fall detection. Our proposed system integrates a low-power millimeter-wave (mmWave) sensor with a NVIDIA Jetson Nano system and uses machine learning to accurately and remotely detect falls. Our initial work focused on processing the mmWave sensor's output by using neural network models, mainly employing Doppler signatures and a Long Short-Term Memory (LSTM) architecture. The proposed system achieved 79% accuracy in detecting three classes of human activities. In addition to reasonable accuracy, the system protected privacy by not recording camera images, ensuring real-time fall detection and Human Activity Recognition (HAR) for both single and multiple individuals at the same time. Building on this foundation, we developed an advanced system to enhance accuracy and robustness in continuous monitoring of human activities. This enhanced system also utilized a mmWave radar sensor (IWR6843ISK-ODS) connected to a NVIDIA Jetson Nano board, and focused on improving the accuracy and robustness of the monitoring process. This integration facilitated effective data processing and inference at the edge, making it suitable for telemedicine systems in both residential and institutional settings. By developing a PointNet neural network for real-time human activity monitoring, we achieved an inference accuracy of 99.5% when recognizing five types of activities: standing, walking, sitting, lying, and falling. Furthermore, the proposed system provided activity data reports, tracking maps, and fall alerts and significantly enhanced telemedicine applcations by enabling more timely and targeted interventions based on objective data. The final proposed system facilitates the ability to detect falls and monitor physical activity at both home and institutional settings, demonstrating the potential of Artificial Intelligence (AI) algorithms and mmWave sensors for HAR. In conclusion, our system enhances therapeutic adherence and optimizes healthcare resources by enabling patients to receive physical therapy services remotely. Furthermore, it could reduce the need for hospital visits and improve in-home nursing care, thus saving time and money and improving patient outcomes.8 0Item Restricted Cloud computing efficiency: optimizing resource utilization, energy consumption, latency, availability, and reliability using intelligent algorithms(The Universit of Western Australia, 2024) Alelyani, Abdullah Hamed A; Datta, Amitava; Ghulam, Mubasher HassanCloud computing offers significant potential for transforming service delivery with a cost-efficient, pay-as-you-go model, which has led to a dramatic increase in demand. The advantages of virtual machine (VM) and container technologies further optimize resource utilization in cloud environments. Containers and VMs improve application reliability by distributing replicated tasks across different physical machines (PMs). However, several persistent issues in cloud computing remain, including energy consumption, resource management, network traffic costs, availability, latency, service level agreement (SLA) violations, and reliability. Addressing these issues is critical for ensuring QoS. This thesis proposes approaches to address these issues and improve cloud performance.13 0Item Restricted Multi-Class Part Parsing based on Deep Learning(Cardiff University, 2024) Alsudays, Njuod; Wu, Jing; Lai, Yu-Kun; Ji, ZeMulti-class part parsing is a dense prediction task that seeks to simultaneously detect multiple objects and the semantic parts within these objects in the scene. This problem is important in providing detailed object understanding but is challenging due to the existence of both class-level and part-level ambiguities. This thesis investigates recent advancements in deep learning to tackle the challenges in multi-class part parsing. First, the AFPSNet network is proposed, which integrates scaled attention and feature fusion to tackle part-level ambiguity and thereby improving parts prediction accuracy. The integration of attention enhances feature representations by focusing on important features, while the feature fusion improves the fusion operation for different scales of features. An object-to-part training strategy is also used to address class-level ambiguity, improving the localisation of parts by exploiting prior knowledge of objects. Building on this foundation, the GRPSNet framework is introduced to further enhance the performance of multi-class part parsing. This framework integrates graph reasoning to capture relationships between parts, thereby improving part segmentation. These captured relationships help to enhance the recognition and localisation of parts. Moreover, the relationships of part boundaries are exploited to further enhance the accuracy of part segmentation. To further refine part segmentation, Multi-Class Boundaries integrated into the AFPSNet network. This integration aims to accurately identify and focus on the spatial boundaries of part classes, thereby enhancing the overall segmentation quality. Experimental results demonstrate the effectiveness of the proposed networks. Various evaluations, including ablation studies and comparisons with existing methods, were conducted on the widely used PASCAL-Part benchmark dataset and the large-scale ADE20K-Part benchmark dataset. These experiments validate the research hypotheses, showing notable improvements in part localisation and segmentation accuracy.24 0Item Restricted Diagnosis of Oral and maxillofacial cysts using artificial intelligence: a literature review(University of Manchester, 2024) Almohawis, Alhaitham; Yong, SinAbstract Oral and maxillofacial cysts are cavities that can pose significant risks if not detected and treated promptly. Many of these cysts are asymptomatic, often going unnoticed until complications arise. The introduction of artificial intelligence (AI) presents a promising opportunity for early detection and management of these cysts. Aim: To explore current studies on the use of artificial intelligence in diagnosing oral and maxillofacial cysts. Objectives: To examine the existing literature in this field, assess the accuracy, effectiveness, and limitations of AI models, and identify challenges in implementing AI in clinical practice. Methods: This literature review followed a systematic approach, identifying 223 studies from PUBMED and SCOPUS databases between 1975 and 2024. After applying inclusion and exclusion criteria, 26 retrospective cohort studies were included in the final analysis. A risk of bias assessment was conducted using the ROBINS I tool. Results: The investigation revealed that AI models consistently demonstrate high accuracy in detecting oral cysts in both radiographs and digital histopathology. The ROBINS I tool indicated a moderate risk of bias in most of the included studies. Notable limitations include limited datasets, variable data quality, and a lack of explainability in AI models results. Conclusion: AI models have shown considerable effectiveness and speed in detecting both simple and complex cysts. However, to fully leverage AI's potential in clinical settings, further rigorous studies are needed to evaluate its risks, benefits, and feasibility, ensuring compliance with governmental health regulations on AI.13 0Item Restricted Exploring the Applications of Artificial Intelligence in Enhancing Pre-Hospital Care: A Scoping Review(Queen’s University, Belfast, 2024) Alfaifi, Yahya; Clarke, SusanArtificial Intelligence (AI) has the potential to significantly improve pre-hospital care, especially in emergency medical services (EMS). However, its current application remains scattered, with varying integration levels across care stages. This scoping review aims to map and assess existing research on AI applications within pre-hospital care without focusing on specific AI technologies, such as machine learning (ML), deep learning (DL), or decision support systems (DSS). The review reflects the current research landscape, capturing how AI is utilised across critical stages such as call-taking, dispatch, and on-scene assessment. Using the framework developed by Arksey and O’Malley (2005), a systematic search was conducted across multiple databases to identify studies relevant to AI in pre-hospital care. The scope was deliberately broad to capture a comprehensive view of the available literature, focusing on identifying areas where further research is needed. The findings indicate that DSS is commonly used to support decision-making in call-taking and dispatch, while more advanced AI applications like ML and DL show potential in predictive analytics and real-time decision-making. However, these technologies are still in their early stages of real-world implementation. This review highlights the gaps in AI research, particularly in the later stages of prehospital care, such as transport and handover. Further exploration is necessary to unlock AI’s full potential in enhancing EMS operations and outcomes.45 0Item Restricted The Role of Artificial Intelligence in Breast Cancer Screening Programmes: A Literature Review and Focus Upon Policy Implications(The University of Edinburgh, 2024) Alrabiah, Alanoud; Hellowell, MarkBackground: Breast cancer (BC) is a leading cause of morbidity and mortality amongst older women, leading to the introduction of screening programmes to support earlier detection and improved survivability. Current screening programmes rely upon the performance of radiologists in terms of accuracy; however, evidence shows that both under and overdiagnosis means screening also results in harms to some women. Artificial intelligence is then a promising technology for improving the accuracy of mammogram screening. Aim: To describe the potential roles of AI in BC screening, and the potential benefits, limitations and risks in these roles. Methods: PubMed, SCOPUS, and CINAHL were searched. Primary research studies published in English and in the last ten years, investigating the accuracy of AI systems for screening BC, were eligible for review. Evidence was appraised using the CASP (2024) checklists and data analysed narratively. Results: 14 studies were found eligible for review, mostly adopting a retrospective study design or laboratory study design. Roles for AI in BC screening include as a standalone system replacing radiologists entirely, as risk stratification systems used before radiologist readings, or as reader aids. While some studies reported AI systems to be superior, others reported accuracy to be inferior to radiologist readings. Differences in results could be due to variations in AI system or radiologist performance. Conclusion: There is insufficient evidence to support the use of AI in BC screening programmes, and more robust, prospective studies comparing readings from clinical practice are urgently required. Policy must also be implemented to regulate the use of AI until there is sufficient evidence to support its use.12 0Item Restricted Sketch compression(University of Surre, 2023-09) Alsadoun, Hadeel Mohammed; song, Yi-Zhe; Ashcroft, AlexanderIn the rapidly evolving field of digital art and animation, traditional sketching techniques often rely on pixel-based methods, leading to less meaningful representations. This dissertation aims to transform this paradigm by rigorously investigating the efficacy of autoencoders for vector sketch compression. We conducted experiments using two distinct neural network architectures: Long Short-Term Memory (LSTM) and Transformer-based autoencoders. The Transformer model, which has significantly impacted the field of sequence-to-sequence tasks, especially in natural language processing, serves as a focal point of our study. Our experiment aims to answer a compelling question: Can these impressive results be replicated in the domain of vector sketch compression? The answer is a resounding yes. The Transformer model not only excelled in reconstructing sketches but also simplified the strokes and enhanced the overall quality of the sketch, achieving an impressive 85.03% classification accuracy. The LSTM model, known for its ability to capture temporal dependencies, served as our baseline, achieving a classification accuracy of 56.139% on a pre-trained classifier. Our findings strongly advocate for the adoption of Transformer-based models in vector sketch compression, offering a more compact and semantically rich representation. The LSTM model’s respectable performance also suggests its potential utility in less complex scenarios. Overall, this study opens new avenues for research in digital art, particularly in optimizing Transformer architectures for sketch compression.12 0Item Restricted MISINFORMATION DETECTION IN THE SOCIAL MEDIA ERA(Howard University, 2024-04-22) Alzahrani, Amani; Rawat, Danda B.As social media becomes the main way of getting information, the spread of misinformation is a serious and widespread problem. Misinformation can take many forms, such as text, video, and audio, and it can travel quickly through different platforms, affecting the quality and trustworthiness of the information that users access around the world. Misinformation can have negative effects on how people think, act, and interact, and it can even endanger social peace. This study aims to tackle the complex problem of misinformation by presenting a comprehensive approach that addresses various forms of deceptive content on social media with a focus on Twitter ( currently X). Twitter stands out as a dynamic and influential microblogging service that enables users to share real-time updates, news, and opinions in concise 280-character messages known as tweets. We introduce a hybrid deep learning model that incorporates Feature-based models at both tweet and user levels, complemented by pre-trained text embedding models such as Global Vectors (GloVe) and Universal Sentence Encoders (USE). Through careful evaluation on a real-world dataset, our approach proves effective in detecting textual misinformation. Recognizing the vital need to verify the reliability of information on social media, we propose a method to assess user credibility. Our solution involves evaluating the credibility of users based on their profiles to enhance the rumors detection model. This study proposes a novel mechanism for assessing a user’s credibility. Additionally, we extended our study capabilities to address the challenges posed by deceptive video content spread on social media using DeepFake technology. As the rapid advancement of deepfake technology threatens the integrity of audio and video content, we present a novel approach combining Optical Flow (OF) algorithms with a Convolutional Neural Network (CNN) to enhance deepfake video detection. This comprehensive strategy addresses the diverse challenges posed by misinformation, credibility assessment, and deepfake detection in the dynamic landscape of social media.37 0Item Restricted Deep Learning-Based Digital Human Modeling And Applications(Saudi Digital Library, 2023-12-14) Ali, Ayman; Wang, PuRecent advancements in the domain of deep learning models have engendered remarkable progress across numerous computer vision tasks. Notably, there has been a burgeoning interest in the field of recovering three-dimensional (3D) human models from monocular images in recent years. This heightened interest can be attributed to the extensive practical applications that necessitate the utilization of 3D human models, including but not limited to gaming, human-computer interaction, virtual systems, and digital twin. The focus of this dissertation is to conceptualize and develop a suite of deep learning-based models with the primary objective of enabling the expeditious and high-fidelity digitalization of human subjects. This endeavor further aims to facilitate a multitude of downstream applications that leverage digital 3D human models. The endeavor to estimate a three-dimensional (3D) human mesh from a monocular image necessitates the application of intricate deep-learning models for enhanced feature extraction, albeit at the expense of heightened computational requirements. As an alternative approach, researchers have explored the utilization of a skeleton-based modality, which represents a lightweight abstraction of human pose, aimed at mitigating the computational intensity. However, this approach entails the omission of significant visual cues, particularly shape information, which cannot be entirely derived from the 3D skeletal representation alone. To harness the advantages of both paradigms, a hybrid methodology that integrates the benefits of 3D human mesh and skeletal information offers a promising avenue. Over the past decade, substantial strides have been made in the estimation of two-dimensional (2D) joint coordinates derived from monocular images. Simultaneously, the application of Convolutional Neural Networks (CNNs) for the extraction of intricate visual features from images has demonstrated its prowess in feature extraction. This progress serves as a compelling impetus for our investigation into a hybrid architectural framework that combines CNNs with a lightweight graph transformer-based approach. This innovative architecture is designed to elevate the 2D joint pose to a comprehensive 3D representation and recover essential visual cues essential for the precise estimation of pose and shape parameters. While SOTA results in 3D Human Pose Estimation (HPE) are important, they do not guarantee the accuracy and plausibility required for biomechanical analysis. Our innovative two-stage deep learning model is designed to efficiently estimate 3D human poses and associated kinematic attributes from monocular videos, with a primary focus on mobile device deployment. The paramount significance of this contribution lies in its ability to provide not only accurate 3D pose estimations but also biomechanically plausible results. This plausibility is essential for achieving accurate biomechanical analyses, thereby advancing various applications, including motion tracking, gesture recognition, and ergonomic assessments. Our work significantly contributes to enhancing our understanding of human movement and its interaction with the environment, ultimately impacting a wide range of biomechanics-related studies and applications. In the realm of human movement analysis, one prominent downstream task is the recognition of human actions based on skeletal data, known as Skeleton-based Human Action Recognition (HAR). This domain has garnered substantial attention within the computer vision community, primarily due to its distinctive attributes, such as computational efficiency, the innate representational power of features, and robustness to variations in illumination. In this context, our research demonstrates that, by representing 3D pose sequences as RGB images, conventional Convolutional Neural Network (CNN) architectures, exemplified by ResNet-50, when complemented by as tute training strategies and diverse augmentation techniques, can attain State-of-the-Art (SOTA) accuracy levels, surpassing the widely adopted Graph Neural Network models. The domain of radar-based sensing, rooted in the transmission and reception of radio waves, offers a non-intrusive and versatile means to monitor human movements, gestures, and vital signs. However, despite its vast potential, the lack of comprehensive radar datasets has hindered the broader implementation of deep learning in radar-based human sensing. In response, the application of synthetic data in deep learning training emerges as a crucial advantage. Synthetic datasets provide an expansive and practically limitless resource, enabling models to adapt and generalize proficiently by exposing them to diverse scenarios, transcending the limitations of real-world data. As part of this research’s trajectory, a novel computational framework known as "virtual radar" is introduced, leveraging 3D pose-driven physics-informed principles. This paradigm allows for the generation of high-fidelity synthetic radar data by merging 3D human models and the principles of Physical Optics (PO) approximation for radar cross-section modeling. The introduction of virtual radar marks a groundbreaking path towards establishing foundational models focused on the nuanced understanding of human behavior through privacy-preserving radar-based methodologies.16 0Item Restricted Exploring AI-Powered Image Generation for Fashion Collaborations: A Deep Learning Approach to Blending Shoe Designs from Multiple Brands(Saudi Digital Library, 2023-09-07) AlJoudi, Sarh; Marchioni, EnricoThis thesis presents an approach for fashion shoe design generation using Generative Adversarial Networks (GANs) with a focus on the WGAN-GP, ProGAN, and StyleGAN architectures. The primary objective is to assess the visual quality, diversity, and adherence to the fashion trends of the generated shoe designs. A set of evaluation metrics, including Fréchet Inception Distance (FID), Earth Mover’s Distance (EMD), Maximum Mean Discrepancy (MMD), and K-Nearest Neighbours (KNN) accuracy, are employed to quantitatively evaluate the performance of each model. The evaluation results demonstrate that ProGAN outperforms both WGAN-GP and StyleGAN in all metrics, achieving the lowest FID, EMD, and MMD scores. ProGAN generates visually appealing and diverse shoe designs, that closely resembling real-world fashion trends. While WGAN-GP also achieves acceptable results, StyleGAN faces challenges with the droplet effect and noisy colours in the generated designs, resulting in lower performance in all evaluation metrics. Nevertheless, StyleGAN’s style-mixing capability showcases its potential for creating novel and creative shoe designs. The qualitative evaluation further confirms the ProGAN as the preferred choice for generating high-quality and fashion-forward shoe designs. Its robustness, progressive growing approach, and architectural stability contribute to its outstanding performance. The discussion also highlights the potential implications of GAN-based fashion image generation in the fashion industry and creative domains, such as trend forecasting and virtual garment display. The findings suggest that ProGAN holds promise as a reliable and creative tool for generating fashion-forward shoe designs and adopting StyleGAN 2 with the Noise-Conditional AdaIN (NCAIN) layer may address the droplet effect problem and enhance visual coherency in the generated designs. Ultimately, the results and discussions pave the way for further advancements and applications of GAN-based fashion design in the fashion industry and creative domains. Keywords: Deep learning, shoe design, GANs, style-mixing.49 0