Search results for: low vision
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1031

Search results for: low vision

851 Autonomous Kuka Youbot Navigation Based on Machine Learning and Path Planning

Authors: Carlos Gordon, Patricio Encalada, Henry Lema, Diego Leon, Dennis Chicaiza

Abstract:

The following work presents a proposal of autonomous navigation of mobile robots implemented in an omnidirectional robot Kuka Youbot. We have been able to perform the integration of robotic operative system (ROS) and machine learning algorithms. ROS mainly provides two distributions; ROS hydro and ROS Kinect. ROS hydro allows managing the nodes of odometry, kinematics, and path planning with statistical and probabilistic, global and local algorithms based on Adaptive Monte Carlo Localization (AMCL) and Dijkstra. Meanwhile, ROS Kinect is responsible for the detection block of dynamic objects which can be in the points of the planned trajectory obstructing the path of Kuka Youbot. The detection is managed by artificial vision module under a trained neural network based on the single shot multibox detector system (SSD), where the main dynamic objects for detection are human beings and domestic animals among other objects. When the objects are detected, the system modifies the trajectory or wait for the decision of the dynamic obstacle. Finally, the obstacles are skipped from the planned trajectory, and the Kuka Youbot can reach its goal thanks to the machine learning algorithms.

Keywords: autonomous navigation, machine learning, path planning, robotic operative system, open source computer vision library

Procedia PDF Downloads 143
850 Automatic Detection of Sugarcane Diseases: A Computer Vision-Based Approach

Authors: Himanshu Sharma, Karthik Kumar, Harish Kumar

Abstract:

The major problem in crop cultivation is the occurrence of multiple crop diseases. During the growth stage, timely identification of crop diseases is paramount to ensure the high yield of crops, lower production costs, and minimize pesticide usage. In most cases, crop diseases produce observable characteristics and symptoms. The Surveyors usually diagnose crop diseases when they walk through the fields. However, surveyor inspections tend to be biased and error-prone due to the nature of the monotonous task and the subjectivity of individuals. In addition, visual inspection of each leaf or plant is costly, time-consuming, and labour-intensive. Furthermore, the plant pathologists and experts who can often identify the disease within the plant according to their symptoms in early stages are not readily available in remote regions. Therefore, this study specifically addressed early detection of leaf scald, red rot, and eyespot types of diseases within sugarcane plants. The study proposes a computer vision-based approach using a convolutional neural network (CNN) for automatic identification of crop diseases. To facilitate this, firstly, images of sugarcane diseases were taken from google without modifying the scene, background, or controlling the illumination to build the training dataset. Then, the testing dataset was developed based on the real-time collected images from the sugarcane field from India. Then, the image dataset is pre-processed for feature extraction and selection. Finally, the CNN-based Visual Geometry Group (VGG) model was deployed on the training and testing dataset to classify the images into diseased and healthy sugarcane plants and measure the model's performance using various parameters, i.e., accuracy, sensitivity, specificity, and F1-score. The promising result of the proposed model lays the groundwork for the automatic early detection of sugarcane disease. The proposed research directly sustains an increase in crop yield.

Keywords: automatic classification, computer vision, convolutional neural network, image processing, sugarcane disease, visual geometry group

Procedia PDF Downloads 84
849 Integrated Risk Management as a Framework for Organisational Success

Authors: Olakunle Felix Adekunle

Abstract:

Risk management is recognised as an essential tool to tackle the inevitable uncertainty associated with business and projects at all levels. But it frequently fails to meet expectations, with projects continuing to run late, over budget or under performing, and business is not gaining the expected benefits. The evident disconnect which often occurs between strategic vision and tactical project delivery typically arises from poorly defined project objectives and inadequate attention to the proactive management of risks that could affect those objectives. One of the main failings in the traditional approach to risk management arises from a narrow focus on the downside, restricted to the technical or operational field, addressing tactical threats to processes, performance or people. This shortcoming can be overcome by widening the scope of risk management to encompass both strategic risks and upside opportunities, creating an integrated approach which can bridge the gap between strategy and tactics. Integrated risk management addresses risk across a variety of levels in the organisation, including strategy and tactics, and covering both opportunity and threat. Effective implementation of integrated risk management can produce a number of benefits to the organisation which are not available from the typical limited-scope risk process. This paper explores how to expand risk management to deliver strategic advantage while retaining its use as a tactical tool.

Keywords: risk management, success, organization, strategy, project, tactis, vision

Procedia PDF Downloads 357
848 Design and Development of Multi-Functional Intelligent Robot Arm Gripper

Authors: W. T. Asheber, L. Chyi-Yeu

Abstract:

An intelligent robot arm is expected to recognize the desired object, grasp it with appropriate force without dropping or damaging it, and also manipulate and deliver the object to the desired destination safely. This paper presents an intelligent multi-finger robot arm gripper design along with vision, proximity, and tactile sensor for efficient grasping and manipulation tasks. The generic design of the gripper makes it convenient for improved parts manipulation, multi-tasking and ease for components assembly. The proposed design emulates the human’s hand fingers structure using linkages and direct drive through power screw like transmission. The actuation and transmission mechanism is designed in such a way that it has non-back-drivable capability, which makes the fingers hold their position when even unpowered. The structural elements are optimized for a finest performance in motion and force transmissivity of the gripper fingers. The actuation mechanisms is designed specially to drive each finger and also rotate two of the fingers about the palm to form appropriate configuration to grasp various size and shape objects. The gripper has an automatic tool set fixture incorporated into its palm, which will reduce time wastage and do assembling in one go. It is equipped with camera-in-hand integrated into its palm; subsequently an image based visual-servoing control scheme is employed.

Keywords: gripper, intelligent gripper, transmissivity, vision sensor

Procedia PDF Downloads 332
847 Automatic Identification and Monitoring of Wildlife via Computer Vision and IoT

Authors: Bilal Arshad, Johan Barthelemy, Elliott Pilton, Pascal Perez

Abstract:

Getting reliable, informative, and up-to-date information about the location, mobility, and behavioural patterns of animals will enhance our ability to research and preserve biodiversity. The fusion of infra-red sensors and camera traps offers an inexpensive way to collect wildlife data in the form of images. However, extracting useful data from these images, such as the identification and counting of animals remains a manual, time-consuming, and costly process. In this paper, we demonstrate that such information can be automatically retrieved by using state-of-the-art deep learning methods. Another major challenge that ecologists are facing is the recounting of one single animal multiple times due to that animal reappearing in other images taken by the same or other camera traps. Nonetheless, such information can be extremely useful for tracking wildlife and understanding its behaviour. To tackle the multiple count problem, we have designed a meshed network of camera traps, so they can share the captured images along with timestamps, cumulative counts, and dimensions of the animal. The proposed method takes leverage of edge computing to support real-time tracking and monitoring of wildlife. This method has been validated in the field and can be easily extended to other applications focusing on wildlife monitoring and management, where the traditional way of monitoring is expensive and time-consuming.

Keywords: computer vision, ecology, internet of things, invasive species management, wildlife management

Procedia PDF Downloads 106
846 Deep Learning Application for Object Image Recognition and Robot Automatic Grasping

Authors: Shiuh-Jer Huang, Chen-Zon Yan, C. K. Huang, Chun-Chien Ting

Abstract:

Since the vision system application in industrial environment for autonomous purposes is required intensely, the image recognition technique becomes an important research topic. Here, deep learning algorithm is employed in image system to recognize the industrial object and integrate with a 7A6 Series Manipulator for object automatic gripping task. PC and Graphic Processing Unit (GPU) are chosen to construct the 3D Vision Recognition System. Depth Camera (Intel RealSense SR300) is employed to extract the image for object recognition and coordinate derivation. The YOLOv2 scheme is adopted in Convolution neural network (CNN) structure for object classification and center point prediction. Additionally, image processing strategy is used to find the object contour for calculating the object orientation angle. Then, the specified object location and orientation information are sent to robotic controller. Finally, a six-axis manipulator can grasp the specific object in a random environment based on the user command and the extracted image information. The experimental results show that YOLOv2 has been successfully employed to detect the object location and category with confidence near 0.9 and 3D position error less than 0.4 mm. It is useful for future intelligent robotic application in industrial 4.0 environment.

Keywords: deep learning, image processing, convolution neural network, YOLOv2, 7A6 series manipulator

Procedia PDF Downloads 210
845 Floodnet: Classification for Post Flood Scene with a High-Resolution Aerial Imaginary Dataset

Authors: Molakala Mourya Vardhan Reddy, Kandimala Revanth, Koduru Sumanth, Beena B. M.

Abstract:

Emergency response and recovery operations are severely hampered by natural catastrophes, especially floods. Understanding post-flood scenarios is essential to disaster management because it facilitates quick evaluation and decision-making. To this end, we introduce FloodNet, a brand-new high-resolution aerial picture collection created especially for comprehending post-flood scenes. A varied collection of excellent aerial photos taken during and after flood occurrences make up FloodNet, which offers comprehensive representations of flooded landscapes, damaged infrastructure, and changed topographies. The dataset provides a thorough resource for training and assessing computer vision models designed to handle the complexity of post-flood scenarios, including a variety of environmental conditions and geographic regions. Pixel-level semantic segmentation masks are used to label the pictures in FloodNet, allowing for a more detailed examination of flood-related characteristics, including debris, water bodies, and damaged structures. Furthermore, temporal and positional metadata improve the dataset's usefulness for longitudinal research and spatiotemporal analysis. For activities like flood extent mapping, damage assessment, and infrastructure recovery projection, we provide baseline standards and evaluation metrics to promote research and development in the field of post-flood scene comprehension. By integrating FloodNet into machine learning pipelines, it will be easier to create reliable algorithms that will help politicians, urban planners, and first responders make choices both before and after floods. The goal of the FloodNet dataset is to support advances in computer vision, remote sensing, and disaster response technologies by providing a useful resource for researchers. FloodNet helps to create creative solutions for boosting communities' resilience in the face of natural catastrophes by tackling the particular problems presented by post-flood situations.

Keywords: image classification, segmentation, computer vision, nature disaster, unmanned arial vehicle(UAV), machine learning.

Procedia PDF Downloads 27
844 Analyzing the Causes of Amblyopia among Patients in Tertiary Care Center: Retrospective Study in King Faisal Specialist Hospital and Research Center

Authors: Hebah M. Musalem, Jeylan El-Mansoury, Lin M. Tuleimat, Selwa Alhazza, Abdul-Aziz A. Al Zoba

Abstract:

Background: Amblyopia is a condition that affects the visual system triggering a decrease in visual acuity without a known underlying pathology. It is due to abnormal vision development in childhood or infancy. Most importantly, vision loss is preventable or reversible with the right kind of intervention in most of the cases. Strabismus, sensory defects, and anisometropia are all well-known causes of amblyopia. However, ocular misalignment in Strabismus is considered the most common form of amblyopia worldwide. The risk of developing amblyopia increases in premature children, developmentally delayed or children who had brain lesions affecting the visual pathway. The prevalence of amblyopia varies between 2 to 5 % in the world according to the literature. Objective: To determine the different causes of Amblyopia in pediatric patients seen in ophthalmology clinic of a tertiary care center, i.e. King Faisal Specialist Hospital and Research Center (KFSH&RC). Methods: This is a hospital based, random retrospective, based on reviewing patient’s files in the Ophthalmology Department of KFSH&RC in Riyadh city, Kingdom of Saudi Arabia. Inclusion criteria: amblyopic pediatric patients who attended the clinic from 2015 to 2016, who are between 6 months and 18 years old. Exclusion Criteria: patients above 18 years of age and any patient who is uncooperative to obtain an accurate vision or a proper refraction. Detailed ocular and medical history are recorded. The examination protocol includes a full ocular exam, full cycloplegic refraction, visual acuity measurement, ocular motility and strabismus evaluation. All data were organized in tables and graphs and analyzed by statistician. Results: Our preliminary results will be discussed on spot by our corresponding author. Conclusions: We focused on this study on utilizing various examination techniques which enhanced our results and highlighted a distinguished correlation between amblyopia and its’ causes. This paper recommendation emphasizes on critical testing protocols to be followed among amblyopic patient, especially in tertiary care centers.

Keywords: amblyopia, amblyopia causes, amblyopia diagnostic criterion, amblyopia prevalence, Saudi Arabia

Procedia PDF Downloads 126
843 Machine Learning and Deep Learning Approach for People Recognition and Tracking in Crowd for Safety Monitoring

Authors: A. Degale Desta, Cheng Jian

Abstract:

Deep learning application in computer vision is rapidly advancing, giving it the ability to monitor the public and quickly identify potentially anomalous behaviour from crowd scenes. Therefore, the purpose of the current work is to improve the performance of safety of people in crowd events from panic behaviour through introducing the innovative idea of Aggregation of Ensembles (AOE), which makes use of the pre-trained ConvNets and a pool of classifiers to find anomalies in video data with packed scenes. According to the theory of algorithms that applied K-means, KNN, CNN, SVD, and Faster-CNN, YOLOv5 architectures learn different levels of semantic representation from crowd videos; the proposed approach leverages an ensemble of various fine-tuned convolutional neural networks (CNN), allowing for the extraction of enriched feature sets. In addition to the above algorithms, a long short-term memory neural network to forecast future feature values and a handmade feature that takes into consideration the peculiarities of the crowd to understand human behavior. On well-known datasets of panic situations, experiments are run to assess the effectiveness and precision of the suggested method. Results reveal that, compared to state-of-the-art methodologies, the system produces better and more promising results in terms of accuracy and processing speed.

Keywords: action recognition, computer vision, crowd detecting and tracking, deep learning

Procedia PDF Downloads 120
842 Refined Edge Detection Network

Authors: Omar Elharrouss, Youssef Hmamouche, Assia Kamal Idrissi, Btissam El Khamlichi, Amal El Fallah-Seghrouchni

Abstract:

Edge detection is represented as one of the most challenging tasks in computer vision, due to the complexity of detecting the edges or boundaries in real-world images that contains objects of different types and scales like trees, building as well as various backgrounds. Edge detection is represented also as a key task for many computer vision applications. Using a set of backbones as well as attention modules, deep-learning-based methods improved the detection of edges compared with the traditional methods like Sobel and Canny. However, images of complex scenes still represent a challenge for these methods. Also, the detected edges using the existing approaches suffer from non-refined results while the image output contains many erroneous edges. To overcome this, n this paper, by using the mechanism of residual learning, a refined edge detection network is proposed (RED-Net). By maintaining the high resolution of edges during the training process, and conserving the resolution of the edge image during the network stage, we make the pooling outputs at each stage connected with the output of the previous layer. Also, after each layer, we use an affined batch normalization layer as an erosion operation for the homogeneous region in the image. The proposed methods are evaluated using the most challenging datasets including BSDS500, NYUD, and Multicue. The obtained results outperform the designed edge detection networks in terms of performance metrics and quality of output images.

Keywords: edge detection, convolutional neural networks, deep learning, scale-representation, backbone

Procedia PDF Downloads 65
841 Understanding the Impact of Spatial Light Distribution on Object Identification in Low Vision: A Pilot Psychophysical Study

Authors: Alexandre Faure, Yoko Mizokami, éRic Dinet

Abstract:

These recent years, the potential of light in assisting visually impaired people in their indoor mobility has been demonstrated by different studies. Implementing smart lighting systems for selective visual enhancement, especially designed for low-vision people, is an approach that breaks with the existing visual aids. The appearance of the surface of an object is significantly influenced by the lighting conditions and the constituent materials of the objects. Appearance of objects may appear to be different from expectation. Therefore, lighting conditions lead to an important part of accurate material recognition. The main objective of this work was to investigate the effect of the spatial distribution of light on object identification in the context of low vision. The purpose was to determine whether and what specific lighting approaches should be preferred for visually impaired people. A psychophysical experiment was designed to study the ability of individuals to identify the smallest cube of a pair under different lighting diffusion conditions. Participants were divided into two distinct groups: a reference group of observers with normal or corrected-to-normal visual acuity and a test group, in which observers were required to wear visual impairment simulation glasses. All participants were presented with pairs of cubes in a "miniature room" and were instructed to estimate the relative size of the two cubes. The miniature room replicates real-life settings, adorned with decorations and separated from external light sources by black curtains. The correlated color temperature was set to 6000 K, and the horizontal illuminance at the object level at approximately 240 lux. The objects presented for comparison consisted of 11 white cubes and 11 black cubes of different sizes manufactured with a 3D printer. Participants were seated 60 cm away from the objects. Two different levels of light diffuseness were implemented. After receiving instructions, participants were asked to judge whether the two presented cubes were the same size or if one was smaller. They provided one of five possible answers: "Left one is smaller," "Left one is smaller but unsure," "Same size," "Right one is smaller," or "Right one is smaller but unsure.". The method of constant stimuli was used, presenting stimulus pairs in a random order to prevent learning and expectation biases. Each pair consisted of a comparison stimulus and a reference cube. A psychometric function was constructed to link stimulus value with the frequency of correct detection, aiming to determine the 50% correct detection threshold. Collected data were analyzed through graphs illustrating participants' responses to stimuli, with accuracy increasing as the size difference between cubes grew. Statistical analyses, including 2-way ANOVA tests, showed that light diffuseness had no significant impact on the difference threshold, whereas object color had a significant influence in low vision scenarios. The first results and trends derived from this pilot experiment clearly and strongly suggest that future investigations could explore extreme diffusion conditions to comprehensively assess the impact of diffusion on object identification. For example, the first findings related to light diffuseness may be attributed to the range of manipulation, emphasizing the need to explore how other lighting-related factors interact with diffuseness.

Keywords: Lighting, Low Vision, Visual Aid, Object Identification, Psychophysical Experiment

Procedia PDF Downloads 33
840 Machine Learning Strategies for Data Extraction from Unstructured Documents in Financial Services

Authors: Delphine Vendryes, Dushyanth Sekhar, Baojia Tong, Matthew Theisen, Chester Curme

Abstract:

Much of the data that inform the decisions of governments, corporations and individuals are harvested from unstructured documents. Data extraction is defined here as a process that turns non-machine-readable information into a machine-readable format that can be stored, for instance, in a database. In financial services, introducing more automation in data extraction pipelines is a major challenge. Information sought by financial data consumers is often buried within vast bodies of unstructured documents, which have historically required thorough manual extraction. Automated solutions provide faster access to non-machine-readable datasets, in a context where untimely information quickly becomes irrelevant. Data quality standards cannot be compromised, so automation requires high data integrity. This multifaceted task is broken down into smaller steps: ingestion, table parsing (detection and structure recognition), text analysis (entity detection and disambiguation), schema-based record extraction, user feedback incorporation. Selected intermediary steps are phrased as machine learning problems. Solutions leveraging cutting-edge approaches from the fields of computer vision (e.g. table detection) and natural language processing (e.g. entity detection and disambiguation) are proposed.

Keywords: computer vision, entity recognition, finance, information retrieval, machine learning, natural language processing

Procedia PDF Downloads 85
839 Innovative Technology to Sustain Food Security in Qatar

Authors: Sana Abusin

Abstract:

Food security in Qatar is a research priority of Qatar University (2021-2025) and all national strategies, including the Qatar National Vision 2030 and food security strategy (2018-2023). Achieving food security requires three actions: 1) transforming surplus food to those who are insecure; 2) reducing food loss and waste by recycling food into valuable resources such as compost (“green fertilizer”) that can be used in growing food; and, finally, 3) establishing strong enforcement agencies to protect consumers from outdated food and promote healthy food. Currently, these objectives are approached separately and not in a sustainable fashion. Food security in Qatar is a research priority of Qatar University (2021-2025) and all national strategies, including the Qatar National Vision 2030 and food security strategy (2018-2023). The study aims to develop an innovative mobile application that supports a sustainable solution to food insecurity and food waste in Qatar. The application will provide a common solution for many different users. For producers, it will facilitate easy disposal of excess food. For charities, it will notify them about surplus food ready for redistribution. The application will also benefit the second layer of end-users in the form of food recycling companies, who will receive information about available food waste that is unable to be consumed. We will use self-exoplanetary diagrams and digital pictures to show all the steps to the final stage. The aim is to motivate the young generation toward innovation and creation, and to encourage public-private collaboration in this sector.

Keywords: food security, innovative technology, sustainability, food waste, Qatar

Procedia PDF Downloads 82
838 Contemporary Vision of Islamic Motifs in Decorating Products

Authors: Shuruq Ghazi Nahhas

Abstract:

Islamic art is a decorative art that depends on repeating motifs in various shapes to cover different surfaces. Each motif has its own characteristics and style that may reflect different Islamic periods, such as Umayyad, Abbasid, Fatimid, Seljuk, Nasrid, Ottoman, and Safavid. These periods were the most powerful periods which played an important role in developing the Islamic motifs. Most of these motifs of the Islamic heritage were not used in new applications. This research focused on reviving the vegetal Islamic motifs found on Islamic heritage and redesign them in a new format to decorate various products, including scarfs, cushions, coasters, wallpaper, wall art, and boxes. The scarf is chosen as one element of these decorative products because it is used as accessories to add aesthetic value to fashion. A descriptive-analytical method is used for this research. The process started with extracting and analyzing the original motifs. Then, creating the new motifs by simplifying, deleting, or adding elements based on the original structure. Then, creating repeated patterns and applying them to decorative products. The findings of this research indicated: repeating patterns based on different structures creates unlimited patterns. Also, changing the elements of the motifs of a pattern adds new characteristics to the pattern. Also, creating frames using elements from the repeated motifs adds aesthetic and contemporary value to decorative products. Finally, using various methods of combining colors creates unlimited variations of each pattern. At the end, reviving the Islamic motifs in contemporary vision enriches decorative products with aesthetic, artistic, and historical values of different Islamic periods. This makes the decorative products valuable that adds uniqueness to their surroundings.

Keywords: Islamic motifs, contemporary patterns, scarfs, decorative products

Procedia PDF Downloads 132
837 Robotic Arm-Automated Spray Painting with One-Shot Object Detection and Region-Based Path Optimization

Authors: Iqraq Kamal, Akmal Razif, Sivadas Chandra Sekaran, Ahmad Syazwan Hisaburi

Abstract:

Painting plays a crucial role in the aerospace manufacturing industry, serving both protective and cosmetic purposes for components. However, the traditional manual painting method is time-consuming and labor-intensive, posing challenges for the sector in achieving higher efficiency. Additionally, the current automated robot path planning has been a bottleneck for spray painting processes, as typical manual teaching methods are time-consuming, error-prone, and skill-dependent. Therefore, it is essential to develop automated tool path planning methods to replace manual ones, reducing costs and improving product quality. Focusing on flat panel painting in aerospace manufacturing, this study aims to address issues related to unreliable part identification techniques caused by the high-mixture, low-volume nature of the industry. The proposed solution involves using a spray gun and a UR10 robotic arm with a vision system that utilizes one-shot object detection (OS2D) to identify parts accurately. Additionally, the research optimizes path planning by concentrating on the region of interest—specifically, the identified part, rather than uniformly covering the entire painting tray.

Keywords: aerospace manufacturing, one-shot object detection, automated spray painting, vision-based path optimization, deep learning, automation, robotic arm

Procedia PDF Downloads 36
836 Application of Improved Semantic Communication Technology in Remote Sensing Data Transmission

Authors: Tingwei Shu, Dong Zhou, Chengjun Guo

Abstract:

Semantic communication is an emerging form of communication that realize intelligent communication by extracting semantic information of data at the source and transmitting it, and recovering the data at the receiving end. It can effectively solve the problem of data transmission under the situation of large data volume, low SNR and restricted bandwidth. With the development of Deep Learning, semantic communication further matures and is gradually applied in the fields of the Internet of Things, Uumanned Air Vehicle cluster communication, remote sensing scenarios, etc. We propose an improved semantic communication system for the situation where the data volume is huge and the spectrum resources are limited during the transmission of remote sensing images. At the transmitting, we need to extract the semantic information of remote sensing images, but there are some problems. The traditional semantic communication system based on Convolutional Neural Network cannot take into account the global semantic information and local semantic information of the image, which results in less-than-ideal image recovery at the receiving end. Therefore, we adopt the improved vision-Transformer-based structure as the semantic encoder instead of the mainstream one using CNN to extract the image semantic features. In this paper, we first perform pre-processing operations on remote sensing images to improve the resolution of the images in order to obtain images with more semantic information. We use wavelet transform to decompose the image into high-frequency and low-frequency components, perform bilinear interpolation on the high-frequency components and bicubic interpolation on the low-frequency components, and finally perform wavelet inverse transform to obtain the preprocessed image. We adopt the improved Vision-Transformer structure as the semantic coder to extract and transmit the semantic information of remote sensing images. The Vision-Transformer structure can better train the huge data volume and extract better image semantic features, and adopt the multi-layer self-attention mechanism to better capture the correlation between semantic features and reduce redundant features. Secondly, to improve the coding efficiency, we reduce the quadratic complexity of the self-attentive mechanism itself to linear so as to improve the image data processing speed of the model. We conducted experimental simulations on the RSOD dataset and compared the designed system with a semantic communication system based on CNN and image coding methods such as BGP and JPEG to verify that the method can effectively alleviate the problem of excessive data volume and improve the performance of image data communication.

Keywords: semantic communication, transformer, wavelet transform, data processing

Procedia PDF Downloads 43
835 Enhancing Plant Throughput in Mineral Processing Through Multimodal Artificial Intelligence

Authors: Muhammad Bilal Shaikh

Abstract:

Mineral processing plants play a pivotal role in extracting valuable minerals from raw ores, contributing significantly to various industries. However, the optimization of plant throughput remains a complex challenge, necessitating innovative approaches for increased efficiency and productivity. This research paper investigates the application of Multimodal Artificial Intelligence (MAI) techniques to address this challenge, aiming to improve overall plant throughput in mineral processing operations. The integration of multimodal AI leverages a combination of diverse data sources, including sensor data, images, and textual information, to provide a holistic understanding of the complex processes involved in mineral extraction. The paper explores the synergies between various AI modalities, such as machine learning, computer vision, and natural language processing, to create a comprehensive and adaptive system for optimizing mineral processing plants. The primary focus of the research is on developing advanced predictive models that can accurately forecast various parameters affecting plant throughput. Utilizing historical process data, machine learning algorithms are trained to identify patterns, correlations, and dependencies within the intricate network of mineral processing operations. This enables real-time decision-making and process optimization, ultimately leading to enhanced plant throughput. Incorporating computer vision into the multimodal AI framework allows for the analysis of visual data from sensors and cameras positioned throughout the plant. This visual input aids in monitoring equipment conditions, identifying anomalies, and optimizing the flow of raw materials. The combination of machine learning and computer vision enables the creation of predictive maintenance strategies, reducing downtime and improving the overall reliability of mineral processing plants. Furthermore, the integration of natural language processing facilitates the extraction of valuable insights from unstructured textual data, such as maintenance logs, research papers, and operator reports. By understanding and analyzing this textual information, the multimodal AI system can identify trends, potential bottlenecks, and areas for improvement in plant operations. This comprehensive approach enables a more nuanced understanding of the factors influencing throughput and allows for targeted interventions. The research also explores the challenges associated with implementing multimodal AI in mineral processing plants, including data integration, model interpretability, and scalability. Addressing these challenges is crucial for the successful deployment of AI solutions in real-world industrial settings. To validate the effectiveness of the proposed multimodal AI framework, the research conducts case studies in collaboration with mineral processing plants. The results demonstrate tangible improvements in plant throughput, efficiency, and cost-effectiveness. The paper concludes with insights into the broader implications of implementing multimodal AI in mineral processing and its potential to revolutionize the industry by providing a robust, adaptive, and data-driven approach to optimizing plant operations. In summary, this research contributes to the evolving field of mineral processing by showcasing the transformative potential of multimodal artificial intelligence in enhancing plant throughput. The proposed framework offers a holistic solution that integrates machine learning, computer vision, and natural language processing to address the intricacies of mineral extraction processes, paving the way for a more efficient and sustainable future in the mineral processing industry.

Keywords: multimodal AI, computer vision, NLP, mineral processing, mining

Procedia PDF Downloads 31
834 Higher Education for Knowledge and Technology Transfer in Egypt

Authors: M. A. Zaki Ewiss, S. Afifi

Abstract:

Nahda University (NUB) believes that internationalisation of higher educational is able to provide global society with an education that meets current needs and that can respond efficiently to contemporary demands and challenges, which are characterized by globalisation, interdependence, and multiculturalism. In this paper, we will discuss the the challenges of the Egyptian Higher Education system and the future vision to improve this system> In this report, the following issues will be considered: Increasing knowledge on the development of specialized programs of study at the university. Developing international cooperation programs, which focus on the development of the students and staff skills, and providing academic culture and learning opportunities. Increasing the opportunities for student mobility, and research projects for faculty members. Increased opportunities for staff, faculty and students to continue to learn foreign universities, and to benefit from scholarships in various disciplines. Taking the advantage of the educational experience and modern teaching methods; Providing the opportunities to study abroad without increasing the period of time required for graduation, and through greater integration in the curricula and programs; More cultural interaction through student exchanges.Improving and providing job opportunities for graduates through participation in the global labor market. This document sets out NUB strategy to move towards that vision. We are confident that greater explicit differentiation, greater freedom and greater collaboration are the keys to delivering the further improvement in quality we shall need to retain and strengthen our position as one of the world’s leading higher education systems.

Keywords: technology transfer higher education, knowledge transfer, internationalisation, mobility

Procedia PDF Downloads 408
833 Real-Time Generative Architecture for Mesh and Texture

Authors: Xi Liu, Fan Yuan

Abstract:

In the evolving landscape of physics-based machine learning (PBML), particularly within fluid dynamics and its applications in electromechanical engineering, robot vision, and robot learning, achieving precision and alignment with researchers' specific needs presents a formidable challenge. In response, this work proposes a methodology that integrates neural transformation with a modified smoothed particle hydrodynamics model for generating transformed 3D fluid simulations. This approach is useful for nanoscale science, where the unique and complex behaviors of viscoelastic medium demand accurate neurally-transformed simulations for materials understanding and manipulation. In electromechanical engineering, the method enhances the design and functionality of fluid-operated systems, particularly microfluidic devices, contributing to advancements in nanomaterial design, drug delivery systems, and more. The proposed approach also aligns with the principles of PBML, offering advantages such as multi-fluid stylization and consistent particle attribute transfer. This capability is valuable in various fields where the interaction of multiple fluid components is significant. Moreover, the application of neurally-transformed hydrodynamical models extends to manufacturing processes, such as the production of microelectromechanical systems, enhancing efficiency and cost-effectiveness. The system's ability to perform neural transfer on 3D fluid scenes using a deep learning algorithm alongside physical models further adds a layer of flexibility, allowing researchers to tailor simulations to specific needs across scientific and engineering disciplines.

Keywords: physics-based machine learning, robot vision, robot learning, hydrodynamics

Procedia PDF Downloads 33
832 A Deep Learning Approach to Detect Complete Safety Equipment for Construction Workers Based on YOLOv7

Authors: Shariful Islam, Sharun Akter Khushbu, S. M. Shaqib, Shahriar Sultan Ramit

Abstract:

In the construction sector, ensuring worker safety is of the utmost significance. In this study, a deep learning-based technique is presented for identifying safety gear worn by construction workers, such as helmets, goggles, jackets, gloves, and footwear. The suggested method precisely locates these safety items by using the YOLO v7 (You Only Look Once) object detection algorithm. The dataset utilized in this work consists of labeled images split into training, testing and validation sets. Each image has bounding box labels that indicate where the safety equipment is located within the image. The model is trained to identify and categorize the safety equipment based on the labeled dataset through an iterative training approach. We used custom dataset to train this model. Our trained model performed admirably well, with good precision, recall, and F1-score for safety equipment recognition. Also, the model's evaluation produced encouraging results, with a [email protected] score of 87.7%. The model performs effectively, making it possible to quickly identify safety equipment violations on building sites. A thorough evaluation of the outcomes reveals the model's advantages and points up potential areas for development. By offering an automatic and trustworthy method for safety equipment detection, this research contributes to the fields of computer vision and workplace safety. The proposed deep learning-based approach will increase safety compliance and reduce the risk of accidents in the construction industry.

Keywords: deep learning, safety equipment detection, YOLOv7, computer vision, workplace safety

Procedia PDF Downloads 31
831 Sustainable Agriculture of Tribal Farmers: An Analysis in Koraput and Malkangiri Districts of Odisha, India

Authors: Amrita Mishra, Tushar Kanti Das

Abstract:

Agriculture is the backbone of the economy of Odisha. Sustainability of agriculture holds the key for the development of Odisha. The Sustainable Development Goals are a framework of 17 goals and 169 targets across social, economical and environmental areas of sustainable development. Among all the seventeen goals the second goal is focusing on the promotion of Sustainable Agriculture. In this research our main aim is also to contribute an understanding of effectiveness of sustainable agriculture as a tool for rural development in the selected tribal district (i.e. Koraput and Malkangiri) of Odisha. These two districts are comes under KBK districts of Odisha which are identified as most backward districts of Odisha. The objectives of our study are to investigate the effect of sustainable agriculture on the lives of tribal farmers, to study whether the farmers are empowered by their participation in sustainable agriculture initiatives to move towards their own vision of development and to study the investment and profit ratio in sustainable agriculture. This research will help in filling the major gaps in sociological studies of sustainable agriculture. This information will helpful for farmers, development organisations, donors and policy makers in formulating the development of effective initiatives and policies to support the development of sustainable agriculture. In this study, we have taken 210 respondents and used various statistical techniques like chi-square test, one-way ANOVA and percentage analysis. This research shows that sustainable agriculture is an effective development strategy that benefits the tribal farmers to move towards their own vision of Good Fortune. The poor farmers who struggle to feed their families and maintain viable livelihoods on shrinking land for them sustainable agriculture are really benefited. The farmers are using homemade pesticides, manure and also getting the seeds from different development organisations and Government. So the investment in Sustainable Agriculture is very less. All farmers said their lives are now better than before. The creation of farmers groups for training and marketing for the produces was shown to be very important for empowerment.

Keywords: sustainable, agriculture, tribal farmers, development, empowerment

Procedia PDF Downloads 134
830 The Importance of Visual Communication in Artificial Intelligence

Authors: Manjitsingh Rajput

Abstract:

Visual communication plays an important role in artificial intelligence (AI) because it enables machines to understand and interpret visual information, similar to how humans do. This abstract explores the importance of visual communication in AI and emphasizes the importance of various applications such as computer vision, object emphasis recognition, image classification and autonomous systems. In going deeper, with deep learning techniques and neural networks that modify visual understanding, In addition to AI programming, the abstract discusses challenges facing visual interfaces for AI, such as data scarcity, domain optimization, and interpretability. Visual communication and other approaches, such as natural language processing and speech recognition, have also been explored. Overall, this abstract highlights the critical role that visual communication plays in advancing AI capabilities and enabling machines to perceive and understand the world around them. The abstract also explores the integration of visual communication with other modalities like natural language processing and speech recognition, emphasizing the critical role of visual communication in AI capabilities. This methodology explores the importance of visual communication in AI development and implementation, highlighting its potential to enhance the effectiveness and accessibility of AI systems. It provides a comprehensive approach to integrating visual elements into AI systems, making them more user-friendly and efficient. In conclusion, Visual communication is crucial in AI systems for object recognition, facial analysis, and augmented reality, but challenges like data quality, interpretability, and ethics must be addressed. Visual communication enhances user experience, decision-making, accessibility, and collaboration. Developers can integrate visual elements for efficient and accessible AI systems.

Keywords: visual communication AI, computer vision, visual aid in communication, essence of visual communication.

Procedia PDF Downloads 47
829 Impact of Adolescent Smoking on the Behaviour, Academic and Health Aspects in Qatar

Authors: Abdelsalam Gomaa, Mahjabeen Ramzan, Tooba Ali Akbar, Huma Nadeem

Abstract:

The use of tobacco and the health risks linked to it are well known in this day and age due to the presence of easily available information through the internet. The media is a powerful platform that is used by many anti-smoking awareness campaigns to reach their target audience; yet, it has been found that adolescents are taking up smoking every passing day. Half of this smoking population of youngsters resides in Asia alone, which includes Qatar, the focus country of this study. As smoking happens to be one of the largest avoidable causes of serious diseases like cancers and heart problems, children are taking up smoking at an alarming rate everywhere including Qatar. Importance of the health of the citizens of Qatar is one of the pillars of the Qatar vision 2030, which is to ensure a healthy population, both physically and mentally. Since the youth makes up a significant percentage of the population and in order to achieve the health objectives of the Qatar vision 2030, it is essential to ensure the health and well-being of this part of the population of the country as they are the future of Qatar. Children, especially boys who tend to be more aggressive by nature, are highly likely to develop behavioral and health issues due to smoking at an early age. Research conducted around the world has also emphasized on this association between the smokers developing a bad behaviour as well as poor social communication skills. However, due to lack of research into this association, very little is known about the extent to which smoking impacts the children’s academics, health and behaviour. Moreover, a study of this nature has not yet been conducted in Qatar previously as most of the studies focus on adult smokers and ways to minimize the number of smoking habits in universities and workplaces. This study solely focuses on identifying a relationship between smoking and its impacts on the adolescents by conducting a research on different schools across Qatar.

Keywords: adolescents, modelling techniques, Qatar, smoking

Procedia PDF Downloads 208
828 Vehicle Speed Estimation Using Image Processing

Authors: Prodipta Bhowmik, Poulami Saha, Preety Mehra, Yogesh Soni, Triloki Nath Jha

Abstract:

In India, the smart city concept is growing day by day. So, for smart city development, a better traffic management and monitoring system is a very important requirement. Nowadays, road accidents increase due to more vehicles on the road. Reckless driving is mainly responsible for a huge number of accidents. So, an efficient traffic management system is required for all kinds of roads to control the traffic speed. The speed limit varies from road to road basis. Previously, there was a radar system but due to high cost and less precision, the radar system is unable to become favorable in a traffic management system. Traffic management system faces different types of problems every day and it has become a researchable topic on how to solve this problem. This paper proposed a computer vision and machine learning-based automated system for multiple vehicle detection, tracking, and speed estimation of vehicles using image processing. Detection of vehicles and estimating their speed from a real-time video is tough work to do. The objective of this paper is to detect vehicles and estimate their speed as accurately as possible. So for this, a real-time video is first captured, then the frames are extracted from that video, then from that frames, the vehicles are detected, and thereafter, the tracking of vehicles starts, and finally, the speed of the moving vehicles is estimated. The goal of this method is to develop a cost-friendly system that can able to detect multiple types of vehicles at the same time.

Keywords: OpenCV, Haar Cascade classifier, DLIB, YOLOV3, centroid tracker, vehicle detection, vehicle tracking, vehicle speed estimation, computer vision

Procedia PDF Downloads 46
827 Vision Zero for the Caribbean Using the Systemic Approach for Road Safety: A Case Study Analyzing Jamaican Road Crash Data (Ongoing)

Authors: Rachelle McFarlane

Abstract:

The Second Decade of Action Road Safety has begun with increased focus on countries who are disproportionately affected by road fatalities. Researchers highlight the low effectiveness of road safety campaigns in Latin America and the Caribbean (LAC) still reporting approximately 130,000 deaths and six million injuries annually. The regional fatality rate 19.2 per 100,000 with heightened concern for persons 15 to 44 years. In 2021, 483 Jamaicans died in 435 crashes, with 33% of these fatalities occurring during Covid-19 curfew hours. The study objective is to conduct a systemic safety review of Jamaican road crashes and provide a framework for its use in complementing traditional methods. The methodology involves the use of the FHWA Systemic Safety Project Selection Tool for analysis. This tool reviews systemwide data in order to identify risk factors across the network associated with severe and fatal crashes, rather that only hotspots. A total of 10,379 crashes with 745 fatalities and serious injuries were reviewed. Of the focus crash types listed, 50% of ‘Pedestrian Accidents’ resulted in fatalities and serious injuries, followed by 32% ‘Bicycle’, 24% ‘Single’ and 12% of ‘Head-on’. This study seeks to understand the associated risk factors with these priority crash types across the network and recommend cost-effective countermeasures across common sites. As we press towards Vision Zero, the inclusion of the systemic safety review method, complementing traditional methods, may create a wider impact in reducing road fatalities and serious injury by targeting issues across network with similarities; focus crash types and contributing factors.

Keywords: systemic safety review, risk factors, road crashes, crash types

Procedia PDF Downloads 59
826 Facilitating Waste Management to Achieve Sustainable Residential Built Environments

Authors: Ingy Ibrahim El-Darwish, Neveen Youssef Azmy

Abstract:

The endowment of a healthy environment can be implemented by endorsing sustainable fundamentals. Design of sustainable buildings through recycling of waste, can reduce health problems, provide good environments and contribute to the aesthetically pleasing entourage. Such environments can help in providing energy-saving alternatives to consolidate the principles of sustainability. The poor community awareness and the absence of laws and legislation in Egypt for waste management specifically in residential areas have led to an inability to provide an integrated system for waste management in urban and rural areas. Many problems and environmental challenges face the Egyptian urban environments. From these problems, is the lack of a cohesive vision for waste collection and recycling for energy-saving. The second problem is the lack public awareness of the short term and long term vision of waste management. Bad practices have adversely affected the efficiency of environmental management systems due to lack of urban legislations that codify collection and recycling of residential communities in Egyptian urban environments. Hence, this research tries to address residents on waste management matters to facilitate legislative process on waste collection and classification within residential units and outside them in a preparation phase for recycling in the Egyptian urban environments. In order to achieve this goal, one of the Egyptian communities has been addressed, analyzed and studied. Waste collection, classification, separation and access to recycling places in the urban city are proposed in preparation for a legislation ruling and regulating the process. Hence, sustainable principles are to be achieved.

Keywords: recycling, residential buildings, sustainability, waste

Procedia PDF Downloads 284
825 Role of Ologen in Previously Failed Trabeculectomy in Advanced Glaucoma

Authors: Reetika Sharma, Lalit Tejwani, Himanshu Shekhar, Arun Singhvi

Abstract:

Purpose: Advanced Glaucoma with Failed trab is not an uncommon sight in glaucoma clinic, and such cases usually tend to present with high intraocular pressure (IOP) and advanced cupping, or even glaucomatous atrophy stage. Re-surgery is needed for such cases, and wound modulation poses a major challenge in these cases. We share our experience in this case series with the use of Ologen (collagen matrix implant) along with MMC 0.04% used in surgery. The purpose of the study was to evaluate the efficacy and outcome of collagen matrix implant in re-trabeculectomy in advanced glaucoma cases. Methodology: Eleven eyes of 11 patients (one eye of one patient) underwent re-trabeculectomy surgery with MMC and Ologen. Ologen implant was used in sub scleral and subconjunctival space, as a spacer and wound modulator. In five cases, triple modulation with implant soaked in anti-VEGF was used. Results: All patients had cupping more than 0.9, and one case was GOA. All cases were on maximal medication at presentation and majority were on systemic anti-glaucoma therapy also. Post-surgery, follow-up ranged from 13 – 34 months, and all cases had a follow longer than the gap between previous surgery (which was failed) and re-trab. One case needed AC reformation and one needling was done. Phaco was done at same sitting in four cases. All cases had their IOP lowered post surgery, and vision was maintained in all, however one case was considered as failed re-surgery case. Topical medication was needed in seven cases post-surgery also. Conclusion: Ologen as adjuvant should be considered in all re-trab cases and all high risk and advanced cases, and triple modulation can be next step in these cases. Aggressive IOP control and non- reluctance to continue topical medications post second surgery should be considered in such cases, to give them best possible vision.

Keywords: failed trabeculectomy, ologen, trabeculectomy, advanced glaucoma

Procedia PDF Downloads 290
824 A Review of Deep Learning Methods in Computer-Aided Detection and Diagnosis Systems based on Whole Mammogram and Ultrasound Scan Classification

Authors: Ian Omung'a

Abstract:

Breast cancer remains to be one of the deadliest cancers for women worldwide, with the risk of developing tumors being as high as 50 percent in Sub-Saharan African countries like Kenya. With as many as 42 percent of these cases set to be diagnosed late when cancer has metastasized and or the prognosis has become terminal, Full Field Digital [FFD] Mammography remains an effective screening technique that leads to early detection where in most cases, successful interventions can be made to control or eliminate the tumors altogether. FFD Mammograms have been proven to multiply more effective when used together with Computer-Aided Detection and Diagnosis [CADe] systems, relying on algorithmic implementations of Deep Learning techniques in Computer Vision to carry out deep pattern recognition that is comparable to the level of a human radiologist and decipher whether specific areas of interest in the mammogram scan image portray abnormalities if any and whether these abnormalities are indicative of a benign or malignant tumor. Within this paper, we review emergent Deep Learning techniques that will prove relevant to the development of State-of-The-Art FFD Mammogram CADe systems. These techniques will span self-supervised learning for context-encoded occlusion, self-supervised learning for pre-processing and labeling automation, as well as the creation of a standardized large-scale mammography dataset as a benchmark for CADe systems' evaluation. Finally, comparisons are drawn between existing practices that pre-date these techniques and how the development of CADe systems that incorporate them will be different.

Keywords: breast cancer diagnosis, computer aided detection and diagnosis, deep learning, whole mammogram classfication, ultrasound classification, computer vision

Procedia PDF Downloads 64
823 Late Presentation of Pseudophakic Macula Edema from Oral Kinase Inhibitors: A Case and Literature Review

Authors: Christolyn Raj, Lewis Levitz

Abstract:

Introduction: Two cases of late presentation ( > five years ) of bilateral pseudophakic macula edema related to oral tyrosine kinase inhibitors are described. These cases are the first of their type in the published literature. A review of ocular inflammatory complications of tyrosine kinase inhibitors in the current literature is explored. Case Presentations(s): Case 1 is an 83-year-old female who has been stable on Ibrutinib (Imbruvica ®) for chronic lymphocytic leukemia (CLL). She presented with bilateral blurred vision from severe cystoid macula edema seven years following routine cataract surgery. She was treated with intravitreal steroids with complete resolution without relapse. Case 2 is a 76-year-old female who was on therapy for polycythemia vera with Ruxolitinib (Jakafi®). She presented with bilateral blurred vision from mild cystoid macula edema six years following routine cataract surgery. She responded well to topical steroids without relapse. In both cases, oral tyrosine kinase inhibitor agents were presumed to be the underlying cause and were ceased. Over the last five years, there have been increasing reports in the literature of the inflammatory effects of tyrosine kinase inhibitors on the retina, uvea and optic nerve. Conclusion: Late presentation of pseudophakic macula edema following routine cataract surgery is rare. Such presentations should prompt investigation of the chronic use of systemic medications, especially oral kinase inhibitors. Patients who must remain on these agents require ongoing ophthalmologic assessment in view of their long-term inflammatory side effects.

Keywords: macula edema, oral kinase inhibitors, retinal toxicity, pseudo-phakia

Procedia PDF Downloads 58
822 Event Data Representation Based on Time Stamp for Pedestrian Detection

Authors: Yuta Nakano, Kozo Kajiwara, Atsushi Hori, Takeshi Fujita

Abstract:

In association with the wave of electric vehicles (EV), low energy consumption systems have become more and more important. One of the key technologies to realize low energy consumption is a dynamic vision sensor (DVS), or we can call it an event sensor, neuromorphic vision sensor and so on. This sensor has several features, such as high temporal resolution, which can achieve 1 Mframe/s, and a high dynamic range (120 DB). However, the point that can contribute to low energy consumption the most is its sparsity; to be more specific, this sensor only captures the pixels that have intensity change. In other words, there is no signal in the area that does not have any intensity change. That is to say, this sensor is more energy efficient than conventional sensors such as RGB cameras because we can remove redundant data. On the other side of the advantages, it is difficult to handle the data because the data format is completely different from RGB image; for example, acquired signals are asynchronous and sparse, and each signal is composed of x-y coordinate, polarity (two values: +1 or -1) and time stamp, it does not include intensity such as RGB values. Therefore, as we cannot use existing algorithms straightforwardly, we have to design a new processing algorithm to cope with DVS data. In order to solve difficulties caused by data format differences, most of the prior arts make a frame data and feed it to deep learning such as Convolutional Neural Networks (CNN) for object detection and recognition purposes. However, even though we can feed the data, it is still difficult to achieve good performance due to a lack of intensity information. Although polarity is often used as intensity instead of RGB pixel value, it is apparent that polarity information is not rich enough. Considering this context, we proposed to use the timestamp information as a data representation that is fed to deep learning. Concretely, at first, we also make frame data divided by a certain time period, then give intensity value in response to the timestamp in each frame; for example, a high value is given on a recent signal. We expected that this data representation could capture the features, especially of moving objects, because timestamp represents the movement direction and speed. By using this proposal method, we made our own dataset by DVS fixed on a parked car to develop an application for a surveillance system that can detect persons around the car. We think DVS is one of the ideal sensors for surveillance purposes because this sensor can run for a long time with low energy consumption in a NOT dynamic situation. For comparison purposes, we reproduced state of the art method as a benchmark, which makes frames the same as us and feeds polarity information to CNN. Then, we measured the object detection performances of the benchmark and ours on the same dataset. As a result, our method achieved a maximum of 7 points greater than the benchmark in the F1 score.

Keywords: event camera, dynamic vision sensor, deep learning, data representation, object recognition, low energy consumption

Procedia PDF Downloads 60