Search results for: 3D computer vision
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3047

Search results for: 3D computer vision

2927 A Deep Learning Approach to Detect Complete Safety Equipment for Construction Workers Based on YOLOv7

Authors: Shariful Islam, Sharun Akter Khushbu, S. M. Shaqib, Shahriar Sultan Ramit

Abstract:

In the construction sector, ensuring worker safety is of the utmost significance. In this study, a deep learning-based technique is presented for identifying safety gear worn by construction workers, such as helmets, goggles, jackets, gloves, and footwear. The suggested method precisely locates these safety items by using the YOLO v7 (You Only Look Once) object detection algorithm. The dataset utilized in this work consists of labeled images split into training, testing and validation sets. Each image has bounding box labels that indicate where the safety equipment is located within the image. The model is trained to identify and categorize the safety equipment based on the labeled dataset through an iterative training approach. We used custom dataset to train this model. Our trained model performed admirably well, with good precision, recall, and F1-score for safety equipment recognition. Also, the model's evaluation produced encouraging results, with a [email protected] score of 87.7%. The model performs effectively, making it possible to quickly identify safety equipment violations on building sites. A thorough evaluation of the outcomes reveals the model's advantages and points up potential areas for development. By offering an automatic and trustworthy method for safety equipment detection, this research contributes to the fields of computer vision and workplace safety. The proposed deep learning-based approach will increase safety compliance and reduce the risk of accidents in the construction industry.

Keywords: deep learning, safety equipment detection, YOLOv7, computer vision, workplace safety

Procedia PDF Downloads 32
2926 The Importance of Visual Communication in Artificial Intelligence

Authors: Manjitsingh Rajput

Abstract:

Visual communication plays an important role in artificial intelligence (AI) because it enables machines to understand and interpret visual information, similar to how humans do. This abstract explores the importance of visual communication in AI and emphasizes the importance of various applications such as computer vision, object emphasis recognition, image classification and autonomous systems. In going deeper, with deep learning techniques and neural networks that modify visual understanding, In addition to AI programming, the abstract discusses challenges facing visual interfaces for AI, such as data scarcity, domain optimization, and interpretability. Visual communication and other approaches, such as natural language processing and speech recognition, have also been explored. Overall, this abstract highlights the critical role that visual communication plays in advancing AI capabilities and enabling machines to perceive and understand the world around them. The abstract also explores the integration of visual communication with other modalities like natural language processing and speech recognition, emphasizing the critical role of visual communication in AI capabilities. This methodology explores the importance of visual communication in AI development and implementation, highlighting its potential to enhance the effectiveness and accessibility of AI systems. It provides a comprehensive approach to integrating visual elements into AI systems, making them more user-friendly and efficient. In conclusion, Visual communication is crucial in AI systems for object recognition, facial analysis, and augmented reality, but challenges like data quality, interpretability, and ethics must be addressed. Visual communication enhances user experience, decision-making, accessibility, and collaboration. Developers can integrate visual elements for efficient and accessible AI systems.

Keywords: visual communication AI, computer vision, visual aid in communication, essence of visual communication.

Procedia PDF Downloads 47
2925 Vehicle Speed Estimation Using Image Processing

Authors: Prodipta Bhowmik, Poulami Saha, Preety Mehra, Yogesh Soni, Triloki Nath Jha

Abstract:

In India, the smart city concept is growing day by day. So, for smart city development, a better traffic management and monitoring system is a very important requirement. Nowadays, road accidents increase due to more vehicles on the road. Reckless driving is mainly responsible for a huge number of accidents. So, an efficient traffic management system is required for all kinds of roads to control the traffic speed. The speed limit varies from road to road basis. Previously, there was a radar system but due to high cost and less precision, the radar system is unable to become favorable in a traffic management system. Traffic management system faces different types of problems every day and it has become a researchable topic on how to solve this problem. This paper proposed a computer vision and machine learning-based automated system for multiple vehicle detection, tracking, and speed estimation of vehicles using image processing. Detection of vehicles and estimating their speed from a real-time video is tough work to do. The objective of this paper is to detect vehicles and estimate their speed as accurately as possible. So for this, a real-time video is first captured, then the frames are extracted from that video, then from that frames, the vehicles are detected, and thereafter, the tracking of vehicles starts, and finally, the speed of the moving vehicles is estimated. The goal of this method is to develop a cost-friendly system that can able to detect multiple types of vehicles at the same time.

Keywords: OpenCV, Haar Cascade classifier, DLIB, YOLOV3, centroid tracker, vehicle detection, vehicle tracking, vehicle speed estimation, computer vision

Procedia PDF Downloads 46
2924 Traumatic Chiasmal Syndrome Following Traumatic Brain Injury

Authors: Jiping Cai, Ningzhi Wangyang, Jun Shao

Abstract:

Traumatic brain injury (TBI) is one of the major causes of morbidity and mortality that leads to structural and functional damage in several parts of the brain, such as cranial nerves, optic nerve tract or other circuitry involved in vision and occipital lobe, depending on its location and severity. As a result, the function associated with vision processing and perception are significantly affected and cause blurred vision, double vision, decreased peripheral vision and blindness. Here two cases complaining of monocular vision loss (actually temporal hemianopia) due to traumatic chiasmal syndrome after frontal head injury were reported, and were compared the findings with individual case reports published in the literature. Reported cases of traumatic chiasmal syndrome appear to share some common features, such as injury to the frontal bone and fracture of the anterior skull base. The degree of bitemporal hemianopia and visual loss acuity have a variable presentation and was not necessarily related to the severity of the craniocerebral trauma. Chiasmal injury may occur even in the absence bony chip impingement. Isolated bitemporal hemianopia is rare and clinical improvement usually may not occur. Mechanisms of damage to the optic chiasm after trauma include direct tearing, contusion haemorrhage and contusion necrosis, and secondary mechanisms such as cell death, inflammation, edema, neurogenesis impairment and axonal damage associated with TBI. Beside visual field test, MRI evaluation of optic pathways seems to the strong objective evidence to demonstrate the impairment of the integrity of visual systems following TBI. Therefore, traumatic chiasmal syndrome should be considered as a differential diagnosis by both neurosurgeons and ophthalmologists in patients presenting with visual impairment, especially bitemporal hemianopia after head injury causing frontal and anterior skull base fracture.

Keywords: bitemporal hemianopia, brain injury, optic chiasma, traumatic chiasmal syndrome.

Procedia PDF Downloads 41
2923 Laser Corneoplastique™: A Refractive Surgery for Corneal Scars

Authors: Arun C. Gulani, Aaishwariya A. Gulani, Amanda Southall

Abstract:

Background: Laser Corneoplastique™ as a least interventional, visually promising technique for patients with vision disability from corneal scars of varied causes has been retrospectively reviewed and proves to cause a paradigm shift in mindset and approach towards corneal scars as a Refractive surgery aiming for emmetropic, unaided vision of 20;/20 in most cases. Three decades of work on this technique has been compiled in this 15-year study. Subject and Methods: The objective of this study was to determine the success of Laser Corneoplastique™ surgery as a treatment of corneal scar cases. A survey of corneal scar cases caused by various medical histories that had undergone Laser Corneoplastique™ surgery over the past twenty years by a single surgeon Arun C. Gulani, M.D. were retrospectively reviewed. The details of each of the cases were retrieved from their medical records and analyzed. Each patient had been examined thoroughly at their preoperative appointments for stability of refraction and vision, depth of scar, pachymetry, topography, pattern of the scar and uncorrected and best corrected vision potential, which were all taken into account in the patients' treatment plans. Results: 64 eyes of 53 patients were investigated for scar etiology, keratometry, visual acuity, and complications. There were 25 different etiologies seen, with the most common being a Herpetic scar. The average visual acuity post-op was, on average, 20/23.55 (±7.05). Laser parameters used were depth and pulses. Overall, the mean Laser ablation depth was 30.67 (±19.05), ranging from 2 to 73 µm. Number of Laser pulses averaged 191.85 (±112.02). Conclusion: Refractive Laser Corneoplastique™ surgery, when practiced as an art, can address all levels of ametropia while reversing complex corneas and scars from refractive surgery complications back to 20/20 vision.

Keywords: corneal scar, refractive surgery, corneal transplant, laser corneoplastique

Procedia PDF Downloads 143
2922 Usability Evaluation of a Mobile Application to Enhance the Use of Smartphone, by Visually Impaired Users in Indonesia

Authors: Johanna Renny Octavia, Kamila Okta Saarah

Abstract:

Smartphone nowadays is widely used by many people all over the world. However, people with vision impairment may experience difficulties that interfere with the proper usage of the smartphone. In Indonesia, the population of visually impaired is about 13 million people (estimated 285 million people worldwide). There are a number of mobile applications developed to enhance the use of smartphone by visually impaired. This paper discusses the usability evaluation of a mobile application, namely Ray Vision, designed to help visually impaired in using smartphone. A series of usability testing with a number of Indonesian visually impaired revealed 28 usability problems in the mobile application that led to 14 design recommendations. The redesigned application was then re-evaluated through another usability testing series. The results showed that all five usability criteria assessed were increased (usefulness by 13%, effectiveness by 27%, efficiency by 27%, satisfaction by 23%, and learnability by 12%). The System Usability Score (SUS) was also increased by 14.92%.

Keywords: mobile application, smartphone, usability evaluation, vision impaired

Procedia PDF Downloads 279
2921 Static and Dynamic Hand Gesture Recognition Using Convolutional Neural Network Models

Authors: Keyi Wang

Abstract:

Similar to the touchscreen, hand gesture based human-computer interaction (HCI) is a technology that could allow people to perform a variety of tasks faster and more conveniently. This paper proposes a training method of an image-based hand gesture image and video clip recognition system using a CNN (Convolutional Neural Network) with a dataset. A dataset containing 6 hand gesture images is used to train a 2D CNN model. ~98% accuracy is achieved. Furthermore, a 3D CNN model is trained on a dataset containing 4 hand gesture video clips resulting in ~83% accuracy. It is demonstrated that a Cozmo robot loaded with pre-trained models is able to recognize static and dynamic hand gestures.

Keywords: deep learning, hand gesture recognition, computer vision, image processing

Procedia PDF Downloads 104
2920 Generative Adversarial Network for Bidirectional Mappings between Retinal Fundus Images and Vessel Segmented Images

Authors: Haoqi Gao, Koichi Ogawara

Abstract:

Retinal vascular segmentation of color fundus is the basis of ophthalmic computer-aided diagnosis and large-scale disease screening systems. Early screening of fundus diseases has great value for clinical medical diagnosis. The traditional methods depend on the experience of the doctor, which is time-consuming, labor-intensive, and inefficient. Furthermore, medical images are scarce and fraught with legal concerns regarding patient privacy. In this paper, we propose a new Generative Adversarial Network based on CycleGAN for retinal fundus images. This method can generate not only synthetic fundus images but also generate corresponding segmentation masks, which has certain application value and challenge in computer vision and computer graphics. In the results, we evaluate our proposed method from both quantitative and qualitative. For generated segmented images, our method achieves dice coefficient of 0.81 and PR of 0.89 on DRIVE dataset. For generated synthetic fundus images, we use ”Toy Experiment” to verify the state-of-the-art performance of our method.

Keywords: retinal vascular segmentations, generative ad-versarial network, cyclegan, fundus images

Procedia PDF Downloads 103
2919 Critical Vision Innovation and Creativity in the Architecture and Urbanism of the Land in Islam between Traditionalism and Positivism

Authors: Wafeek Mohamed Ibrahim Mohamed

Abstract:

In the era of globalization and openness informational. Anyone who thinks about innovation in the earth population in Islam in our contemporary reality, he will find that it is not destined to its civilized extension to last. The purpose of the research is a trial to reach a realistic vision for creative, innovative and intellectual thought for the earth population in Islam as an instrument to Confrontation and observe the changes that have affected in the architecture of the land during different eras. Through knowing the controls of the ruling legitimacy(that served as definitions and laws which formulate its features) and using customs, traditions, and conventions as a telescope for the earth population in Islam, It explained the impact of them on features of creative formation for the architecture of the land in our contemporary reality. The study shows a modern vision to identify innovation in the earth population in Islam. As well as reformulating its mental image and monitoring its changes in Islamic heritage cities. This will be done through a two main branches: firstly, set forth a theory represented in studying creative concepts which formulate the population of the earth in Islam. Such as initiative and responsibility for reviving the dead land, the lane [alley] as formation unit and social solidarity,… Etc.. The second branch is preparing a practical, critical vision for innovative conceptual thought for the architecture of the land of Islam, through studying the development of a traditional Islamic city., The conceptual thought of making the birth festival ["Al-Refaee"] and its emulation for governing roles in the traditional city building. The research concludes The necessity of forming the suggested a creative vision for identifying how to re-form the conceptual for our contemporary population of the earth. It poses an important question which is how to return to creativity in the architecture of the land of Islam in our built environments.

Keywords: innovation and creation, architecture, the land in Islam, criticism of design

Procedia PDF Downloads 425
2918 Analysing Social Media Coverage of Political Speeches in Relation to Discourse and Context

Authors: Yaser Mohammed Altameemi

Abstract:

This research looks at the representation of the social media for the Saudi Government decrees regarding the developmental projects of the Saudi 2030 vision. The paper analyses a television interview with the Crown Prince Mohammed Bin Salman who talks about the progress of the Saudi vision of 2030, and how the government had acted as response to the COVID-19 pandemic. The interview was on 28/4/2021. The paper analyses the tweets on Twitter that cover the interview for the purpose of investigating the development of concepts and meanings regarding the Saudi peoples’ orientations towards the Saudi projects. The data include all related tweets from the day of the interview and the following seven days after the interview. The finding of the collocation analysis suggests that nationalism notion is explicitly expressed by users in Twitter. The main finding of this paper suggests the importance of further analyses for the concordance lines. However, the collocation network suggests that there is a clear highlight for nationalism.

Keywords: social media, twitter, political interview, prince Mohammed Bin Salman, Saudi vision 2030

Procedia PDF Downloads 154
2917 Machine Vision System for Measuring the Quality of Bulk Sun-dried Organic Raisins

Authors: Navab Karimi, Tohid Alizadeh

Abstract:

An intelligent vision-based system was designed to measure the quality and purity of raisins. A machine vision setup was utilized to capture the images of bulk raisins in ranges of 5-50% mixed pure-impure berries. The textural features of bulk raisins were extracted using Grey-level Histograms, Co-occurrence Matrix, and Local Binary Pattern (a total of 108 features). Genetic Algorithm and neural network regression were used for selecting and ranking the best features (21 features). As a result, the GLCM features set was found to have the highest accuracy (92.4%) among the other sets. Followingly, multiple feature combinations of the previous stage were fed into the second regression (linear regression) to increase accuracy, wherein a combination of 16 features was found to be the optimum. Finally, a Support Vector Machine (SVM) classifier was used to differentiate the mixtures, producing the best efficiency and accuracy of 96.2% and 97.35%, respectively.

Keywords: sun-dried organic raisin, genetic algorithm, feature extraction, ann regression, linear regression, support vector machine, south azerbaijan.

Procedia PDF Downloads 39
2916 Using Artificial Vision Techniques for Dust Detection on Photovoltaic Panels

Authors: Gustavo Funes, Eduardo Peters, Jose Delpiano

Abstract:

It is widely known that photovoltaic technology has been massively distributed over the last decade despite its low-efficiency ratio. Dust deposition reduces this efficiency even more, lowering the energy production and module lifespan. In this work, we developed an artificial vision algorithm based on CIELAB color space to identify dust over panels in an autonomous way. We performed several experiments photographing three different types of panels, 30W, 340W and 410W. Those panels were soiled artificially with uniform and non-uniform distributed dust. The algorithm proposed uses statistical tools to provide a simulation with a 100% soiled panel and then performs a comparison to get the percentage of dirt in the experimental data set. The simulation uses a seed that is obtained by taking a dust sample from the maximum amount of dust from the dataset. The final result is the dirt percentage and the possible distribution of dust over the panel. Dust deposition is a key factor for plant owners to determine cleaning cycles or identify nonuniform depositions that could lead to module failure and hot spots.

Keywords: dust detection, photovoltaic, artificial vision, soiling

Procedia PDF Downloads 18
2915 Design of Speed Bump Recognition System Integrated with Adjustable Shock Absorber Control

Authors: Ming-Yen Chang, Sheng-Hung Ke

Abstract:

This research focuses on the development of a speed bump identification system for real-time control of adjustable shock absorbers in vehicular suspension systems. The study initially involved the collection of images of various speed bumps, and rubber speed bump profiles found on roadways. These images were utilized for training and recognition purposes through the deep learning object detection algorithm YOLOv5. Subsequently, the trained speed bump identification program was integrated with an in-vehicle camera system for live image capture during driving. These images were instantly transmitted to a computer for processing. Using the principles of monocular vision ranging, the distance between the vehicle and an approaching speed bump was determined. The appropriate control distance was established through both practical vehicle measurements and theoretical calculations. Collaboratively, with the electronically adjustable shock absorbers equipped in the vehicle, a shock absorber control system was devised to dynamically adapt the damping force just prior to encountering a speed bump. This system effectively mitigates passenger discomfort and enhances ride quality.

Keywords: adjustable shock absorbers, image recognition, monocular vision ranging, ride

Procedia PDF Downloads 34
2914 A Comparative Study of Social Entrepreneurship Centers in Universities of the World

Authors: Farnoosh Alami, Nazgol Azimi

Abstract:

Universities have recently paid much attention to the subject of social entrepreneurship. As a result, many of the highly ranked universities have established centers in this regard. The present research aims to investigate vision and mission of social entrepreneurship centers of the best universities ranked under 50 by Shanghai List 2013. It tries to find the common goals and features of their mission, vision, and activities which lead to their present success. This investigation is based on the web content of the first top 10 universities; among which six had social entrepreneurship centers. This is a qualitative research, and the findings are based on content analysis of documents. The findings confirm that education, research, talent development, innovative solutions, and supporting social innovation, are shared in the vision of these centers. In regard to their missions, social participation, networking, and leader education are the most shared features. Their common activities are focused on five categories of education, research, support, promotion, and networking.

Keywords: comparative study, qualitative research, social entrepreneurship centers, universities in the world

Procedia PDF Downloads 263
2913 Development of a Social Assistive Robot for Elderly Care

Authors: Edwin Foo, Woei Wen, Lui, Meijun Zhao, Shigeru Kuchii, Chin Sai Wong, Chung Sern Goh, Yi Hao He

Abstract:

This presentation presents an elderly care and assistive social robot development work. We named this robot JOS and he is restricted to table top operation. JOS is designed to have a maximum volume of 3600 cm3 with its base restricted to 250 mm and his mission is to provide companion, assist and help the elderly. In order for JOS to accomplish his mission, he will be equipped with perception, reaction and cognition capability. His appearance will be not human like but more towards cute and approachable type. JOS will also be designed to be neutral gender. However, the robot will still have eyes, eyelid and a mouth. For his eyes and eyelids, they will be built entirely with Robotis Dynamixel AX18 motor. To realize this complex task, JOS will be also be equipped with micro-phone array, vision camera and Intel i5 NUC computer and a powered by a 12 V lithium battery that will be self-charging. His face is constructed using 1 motor each for the eyelid, 2 motors for the eyeballs, 3 motors for the neck mechanism and 1 motor for the lips movement. The vision senor will be house on JOS forehead and the microphone array will be somewhere below the mouth. For the vision system, Omron latest OKAO vision sensor is used. It is a compact and versatile sensor that is only 60mm by 40mm in size and operates with only 5V supply. In addition, OKAO vision sensor is capable of identifying the user and recognizing the expression of the user. With these functions, JOS is able to track and identify the user. If he cannot recognize the user, JOS will ask the user if he would want him to remember the user. If yes, JOS will store the user information together with the capture face image into a database. This will allow JOS to recognize the user the next time the user is with JOS. In addition, JOS is also able to interpret the mood of the user through the facial expression of the user. This will allow the robot to understand the user mood and behavior and react according. Machine learning will be later incorporated to learn the behavior of the user so as to understand the mood of the user and requirement better. For the speech system, Microsoft speech and grammar engine is used for the speech recognition. In order to use the speech engine, we need to build up a speech grammar database that captures the commonly used words by the elderly. This database is built from research journals and literature on elderly speech and also interviewing elderly what do they want to robot to assist them with. Using the result from the interview and research from journal, we are able to derive a set of common words the elderly frequently used to request for the help. It is from this set that we build up our grammar database. In situation where there is more than one person near JOS, he is able to identify the person who is talking to him through an in-house developed microphone array structure. In order to make the robot more interacting, we have also included the capability for the robot to express his emotion to the user through the facial expressions by changing the position and movement of the eyelids and mouth. All robot emotions will be in response to the user mood and request. Lastly, we are expecting to complete this phase of project and test it with elderly and also delirium patient by Feb 2015.

Keywords: social robot, vision, elderly care, machine learning

Procedia PDF Downloads 411
2912 Use of Computer and Machine Learning in Facial Recognition

Authors: Neha Singh, Ananya Arora

Abstract:

Facial expression measurement plays a crucial role in the identification of emotion. Facial expression plays a key role in psychophysiology, neural bases, and emotional disorder, to name a few. The Facial Action Coding System (FACS) has proven to be the most efficient and widely used of the various systems used to describe facial expressions. Coders can manually code facial expressions with FACS and, by viewing video-recorded facial behaviour at a specified frame rate and slow motion, can decompose into action units (AUs). Action units are the most minor visually discriminable facial movements. FACS explicitly differentiates between facial actions and inferences about what the actions mean. Action units are the fundamental unit of FACS methodology. It is regarded as the standard measure for facial behaviour and finds its application in various fields of study beyond emotion science. These include facial neuromuscular disorders, neuroscience, computer vision, computer graphics and animation, and face encoding for digital processing. This paper discusses the conceptual basis for FACS, a numerical listing of discrete facial movements identified by the system, the system's psychometric evaluation, and the software's recommended training requirements.

Keywords: facial action, action units, coding, machine learning

Procedia PDF Downloads 73
2911 Improving Lane Detection for Autonomous Vehicles Using Deep Transfer Learning

Authors: Richard O’Riordan, Saritha Unnikrishnan

Abstract:

Autonomous Vehicles (AVs) are incorporating an increasing number of ADAS features, including automated lane-keeping systems. In recent years, many research papers into lane detection algorithms have been published, varying from computer vision techniques to deep learning methods. The transition from lower levels of autonomy defined in the SAE framework and the progression to higher autonomy levels requires increasingly complex models and algorithms that must be highly reliable in their operation and functionality capacities. Furthermore, these algorithms have no room for error when operating at high levels of autonomy. Although the current research details existing computer vision and deep learning algorithms and their methodologies and individual results, the research also details challenges faced by the algorithms and the resources needed to operate, along with shortcomings experienced during their detection of lanes in certain weather and lighting conditions. This paper will explore these shortcomings and attempt to implement a lane detection algorithm that could be used to achieve improvements in AV lane detection systems. This paper uses a pre-trained LaneNet model to detect lane or non-lane pixels using binary segmentation as the base detection method using an existing dataset BDD100k followed by a custom dataset generated locally. The selected roads will be modern well-laid roads with up-to-date infrastructure and lane markings, while the second road network will be an older road with infrastructure and lane markings reflecting the road network's age. The performance of the proposed method will be evaluated on the custom dataset to compare its performance to the BDD100k dataset. In summary, this paper will use Transfer Learning to provide a fast and robust lane detection algorithm that can handle various road conditions and provide accurate lane detection.

Keywords: ADAS, autonomous vehicles, deep learning, LaneNet, lane detection

Procedia PDF Downloads 54
2910 Mathematics Vision of the Companies' Growth with Educational Technologies

Authors: Valencia P. L. Rodrigo, Morita A. Adelina, Vargas V. Martin

Abstract:

This proposal consists of an analysis of macro concepts involved within an organization growth using educational technologies, which will relate each concept, in a mathematical way with a vision of harmonic work. Working collaboratively, competitively and cooperatively so that this growth is harmonious and homogenous, coining a new term, Harmonic Work. The Harmonic Work ensures that the organization grows in all business directions, allowing managers to project a much more accurate growth, making clear the contribution of each department, resulting in an algorithm that analyzes each of the variables both endogenous and exogenous, establishing different performance indicators in its process of growth.

Keywords: business projection, collaboration, competitiveness, educational technology, harmonious growth

Procedia PDF Downloads 289
2909 Autonomous Kuka Youbot Navigation Based on Machine Learning and Path Planning

Authors: Carlos Gordon, Patricio Encalada, Henry Lema, Diego Leon, Dennis Chicaiza

Abstract:

The following work presents a proposal of autonomous navigation of mobile robots implemented in an omnidirectional robot Kuka Youbot. We have been able to perform the integration of robotic operative system (ROS) and machine learning algorithms. ROS mainly provides two distributions; ROS hydro and ROS Kinect. ROS hydro allows managing the nodes of odometry, kinematics, and path planning with statistical and probabilistic, global and local algorithms based on Adaptive Monte Carlo Localization (AMCL) and Dijkstra. Meanwhile, ROS Kinect is responsible for the detection block of dynamic objects which can be in the points of the planned trajectory obstructing the path of Kuka Youbot. The detection is managed by artificial vision module under a trained neural network based on the single shot multibox detector system (SSD), where the main dynamic objects for detection are human beings and domestic animals among other objects. When the objects are detected, the system modifies the trajectory or wait for the decision of the dynamic obstacle. Finally, the obstacles are skipped from the planned trajectory, and the Kuka Youbot can reach its goal thanks to the machine learning algorithms.

Keywords: autonomous navigation, machine learning, path planning, robotic operative system, open source computer vision library

Procedia PDF Downloads 143
2908 Image Based Landing Solutions for Large Passenger Aircraft

Authors: Thierry Sammour Sawaya, Heikki Deschacht

Abstract:

In commercial aircraft operations, almost half of the accidents happen during approach or landing phases. Automatic guidance and automatic landings have proven to bring significant safety value added for this challenging landing phase. This is why Airbus and ScioTeq have decided to work together to explore the capability of image-based landing solutions as additional landing aids to further expand the possibility to perform automatic approach and landing to runways where the current guiding systems are either not fitted or not optimum. Current systems for automated landing often depend on radio signals provided by airport ground infrastructure on the airport or satellite coverage. In addition, these radio signals may not always be available with the integrity and performance required for safe automatic landing. Being independent from these radio signals would widen the operations possibilities and increase the number of automated landings. Airbus and ScioTeq are joining their expertise in the field of Computer Vision in the European Program called Clean Sky 2 Large Passenger Aircraft, in which they are leading the IMBALS (IMage BAsed Landing Solutions) project. The ultimate goal of this project is to demonstrate, develop, validate and verify a certifiable automatic landing system guiding an airplane during the approach and landing phases based on an onboard camera system capturing images, enabling automatic landing independent from radio signals and without precision instrument for landing. In the frame of this project, ScioTeq is responsible for the development of the Image Processing Platform (IPP), while Airbus is responsible for defining the functional and system requirements as well as the testing and integration of the developed equipment in a Large Passenger Aircraft representative environment. The aim of this paper will be to describe the system as well as the associated methods and tools developed for validation and verification.

Keywords: aircraft landing system, aircraft safety, autoland, avionic system, computer vision, image processing

Procedia PDF Downloads 60
2907 Seashore Debris Detection System Using Deep Learning and Histogram of Gradients-Extractor Based Instance Segmentation Model

Authors: Anshika Kankane, Dongshik Kang

Abstract:

Marine debris has a significant influence on coastal environments, damaging biodiversity, and causing loss and damage to marine and ocean sector. A functional cost-effective and automatic approach has been used to look up at this problem. Computer vision combined with a deep learning-based model is being proposed to identify and categorize marine debris of seven kinds on different beach locations of Japan. This research compares state-of-the-art deep learning models with a suggested model architecture that is utilized as a feature extractor for debris categorization. The model is being proposed to detect seven categories of litter using a manually constructed debris dataset, with the help of Mask R-CNN for instance segmentation and a shape matching network called HOGShape, which can then be cleaned on time by clean-up organizations using warning notifications of the system. The manually constructed dataset for this system is created by annotating the images taken by fixed KaKaXi camera using CVAT annotation tool with seven kinds of category labels. A pre-trained HOG feature extractor on LIBSVM is being used along with multiple templates matching on HOG maps of images and HOG maps of templates to improve the predicted masked images obtained via Mask R-CNN training. This system intends to timely alert the cleanup organizations with the warning notifications using live recorded beach debris data. The suggested network results in the improvement of misclassified debris masks of debris objects with different illuminations, shapes, viewpoints and litter with occlusions which have vague visibility.

Keywords: computer vision, debris, deep learning, fixed live camera images, histogram of gradients feature extractor, instance segmentation, manually annotated dataset, multiple template matching

Procedia PDF Downloads 63
2906 The Effect of Postural Sway and Technical Parameters of 8 Weeks Technical Training Performed with Restrict of Visual Input on the 10-12 Ages Soccer Players

Authors: Nurtekin Erkmen, Turgut Kaplan, Halil Taskin, Ahmet Sanioglu, Gokhan Ipekoglu

Abstract:

The aim of this study was to determine the effects of an 8 week soccerspecific technical training with limited vision perception on postural control and technical parameters in 10-12 aged soccer players. Subjects in this study were 24 male young soccer players (age: 11.00 ± 0.56 years, height: 150.5 ± 4.23 cm, body weight: 41.49 ± 7.56 kg). Subjects were randomly divided as two groups: Training and control. Balance performance was measured by Biodex Balance System (BBS). Short pass, speed dribbling, 20 m speed with ball, ball control, juggling tests were used to measure soccer players’ technical performances with a ball. Subjects performed soccer training 3 times per week for 8 weeks. In each session, training group with limited vision perception and control group with normal vision perception committed soccer-specific technical drills for 20 min. Data analyzed with t-test for independent samples and Mann-Whitney U between groups and paired t-test and Wilcoxon test between pre-posttests. No significant difference was found balance scores and with eyes open and eyes closed and LOS test between training and control groups after training (p>0.05). After eight week of training there are no significant difference in balance score with eyes open for both training and control groups (p>0.05). Balance scores decreased in training and control groups after the training (p<0.05). The completion time of LOS test shortened in both training and control groups after training (p<0.05). The training developed speed dribbling performance of training group (p<0.05). On the other hand, soccer players’ performance in training and control groups increased in 20 m speed with a ball after eight week training (p<0.05). In conclusion; the results of this study indicate that soccer-specific training with limited vision perception may not improves balance performance in 10-12 aged soccer players, but it develops speed dribbling performance.

Keywords: Young soccer players, vision perception, postural control, technical

Procedia PDF Downloads 444
2905 An Effective Change in the Strategic Structure of Quality Management Systems: The Organization’s Needs Management

Authors: Joel Carlos Vieira Reinhardt, Mariana de Freitas Dewes, Odair Lelis Gonçalez

Abstract:

This paper proposes a method to implement a strategic framework for the quality management system that considers the analysis of prospective scenarios in the determination of policy, mission, vision, objectives, processes, monitoring, and goals. Semantic categorization of qualitative testimonial research on employee perception shows it was possible to implement an effective change in the organizations at the Department of Aerospace Science and Technology through the focus on the organization's needs management, producing a rupture with the historical managerial practice.

Keywords: management of company needs, mission, prospective scenarios, quality management, quality policy, vision

Procedia PDF Downloads 72
2904 Development a Battery of Measurements to Assess Giftedness Initiatives in Light of the Objectives of Saudi Arabia's Future Vision of Gifted Education

Authors: Saeed M. Al Qahtani, Alaa Eldin A. Ayoub

Abstract:

The study aimed to develop a battery of measures to assessment gifted initiatives in Saudi Arabia. The battery consisted of 17 measures developed in light of Saudi Arabia's future vision objectives for gifted education. A battery was applied to 193 gifted students who benefit from gifted initiatives and programs, 42 teachers of gifted as well as, 40 experts of gifted. Samples were taken from three main regions: Riyadh, Sharqia, Gharbia in Saudi Arabia. The results indicated that battery measures have a reliability and stability index ranging from 0.6 to 0.87. Besides that, results showed that the educational environment lacks many basic components such as facilities, laboratories, and activities that may stimulate creativity and innovation. Furthermore, results showed that there is a weakness in private sector involvement in the construction of educational buildings, special centers for gifted people and the provision of certain facilities that support talented programs. The recommendations of the study indicate the need for the private sector participation in the provision of services and projects for the care of gifted students in Saudi Arabia.

Keywords: battery of measures, gifted care initiatives, Saudi future vision, gifted student

Procedia PDF Downloads 140
2903 Computer Fraud from the Perspective of Iran's Law and International Documents

Authors: Babak Pourghahramani

Abstract:

One of the modern crimes against property and ownership in the cyber-space is the computer fraud. Despite being modern, the aforementioned crime has its roots in the principles of religious jurisprudence. In some cases, this crime is compatible with the traditional regulations and that is when the computer is considered as a crime commitment device and also some computer frauds that take place in the context of electronic exchanges are considered as crime based on the E-commerce Law (approved in 2003) but the aforementioned regulations are flawed and until recent years there was no comprehensive law in this regard; yet after some years the Computer Crime Act was approved in 2009/26/5 and partly solved the problem of legal vacuum. The present study intends to investigate the computer fraud according to Iran's Computer Crime Act and by taking into consideration the international documents.

Keywords: fraud, cyber fraud, computer fraud, classic fraud, computer crime

Procedia PDF Downloads 301
2902 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments

Authors: Rahul Paul, Peter Mctaggart, Luke Skinner

Abstract:

Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.

Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry

Procedia PDF Downloads 51
2901 Simplified Mobile AR Platform Design for Augmented Tourism

Authors: Eric Hawkinson, Edgaras Artemciukas

Abstract:

This study outlines iterations of designing mobile augmented reality (MAR) applications for tourism specific contexts. Using a design based research model, several cycles of development to implementation were analyzed and refined upon with the goal of building a MAR platform that would facilitate the creation of augmented tours and environments by non-technical users. The project took on several stages, and through the process, a simple framework was begun to be established that can inform the design and use of MAR applications for tourism contexts. As a result of these iterations of development, a platform was developed that can allow novice computer users to create augmented tourism environments. This system was able to connect existing tools in widespread use such as Google Forms and connect them to computer vision algorithms needed for more advanced augmented tourism environments. The study concludes with a discussion of this MAR platform and reveals design elements that have implications for tourism contexts. The study also points to future case uses and design approaches for augmented tourism.

Keywords: augmented tourism, augmented reality, user experience, mobile design, e-tourism

Procedia PDF Downloads 175
2900 The Yield of Neuroimaging in Patients Presenting to the Emergency Department with Isolated Neuro-Ophthalmological Conditions

Authors: Dalia El Hadi, Alaa Bou Ghannam, Hala Mostafa, Hana Mansour, Ibrahim Hashim, Soubhi Tahhan, Tharwat El Zahran

Abstract:

Introduction: Neuro-ophthalmological emergencies require prompt assessment and management to avoid vision or life-threatening sequelae. Some would require neuroimaging. Most commonly used are the CT and MRI of the Brain. They can be over-used when not indicated. Their yield remains dependent on multiple factors relating to the clinical scenario. Methods: A retrospective cross-sectional study was conducted by reviewing the electronic medical records of patients presenting to the Emergency Department (ED) with isolated neuro-ophthalmologic complaints. For each patient, data were collected on the clinical presentation, whether neuroimaging was performed (and which type), and the result of neuroimaging. Analysis of the performed neuroimaging was made, and its yield was determined. Results: A total of 211 patients were reviewed. The complaints or symptoms at presentation were: blurry vision, change in the visual field, transient vision loss, floaters, double vision, eye pain, eyelid droop, headache, dizziness and others such as nausea or vomiting. In the ED, a total of 126 neuroimaging procedures were performed. Ninety-four imagings (74.6%) were normal, while 32 (25.4%) had relevant abnormal findings. Only 2 symptoms were significant for abnormal imaging: blurry vision (p-value= 0.038) and visual field change (p-value= 0.014). While 4 physical exam findings had significant abnormal imaging: visual field defect (p-value= 0.016), abnormal pupil reactivity (p-value= 0.028), afferent pupillary defect (p-value= 0.018), and abnormal optic disc exam (p-value= 0.009). Conclusion: Risk indicators for abnormal neuroimaging in the setting of neuro-ophthalmological emergencies are blurred vision or changes in the visual field on history taking. While visual field irregularities, abnormal pupil reactivity with or without afferent pupillary defect, or abnormal optic discs, are risk factors related to physical testing. These findings, when present, should sway the ED physician towards neuroimaging but still individualizing each case is of utmost importance to prevent time-consuming, resource-draining, and sometimes unnecessary workup. In the end, it suggests a well-structured patient-centered algorithm to be followed by ED physicians.

Keywords: emergency department, neuro-ophthalmology, neuroimaging, risk indicators

Procedia PDF Downloads 146
2899 Public-Private Partnership in Tourism Development: Kuwait Experience within 2035 Vision

Authors: Obaid Alotaibi

Abstract:

Tourism and recreation have become one of the important and influential sectors in most of the modern economies. This sector has been accepted as one of the alternative sources of national income, employment, and foreign exchange. Kuwait has many potentialities in tourism and recreation, and exploitation of this leads to more diversification of the economy besides augmenting its contribution to the GDP. It is an import-oriented economy; it requires hard currencies (foreign exchange) to meet the import costs as well as to maintain stability in the international market. To compensate for the revenue fall stemmed from fluctuations in oil prices -where the agriculture, fisheries, and industrial sectors are too immune and inelastic- the only alternative solution is the regeneration of the tourism and recreation to surface. This study envisages the characteristics of tourism and recreation, the economic and social importance for the society, the physical and human endowments, as well as the tourist pattern and plans for promoting and sustaining tourism in the country. The study summarizes many recommendations, including the necessity of establishing authority or a council for tourism, linking the planning of tourism development with the comprehensive planning for economic and social development in Kuwait in the shadow of 2035 vision, and to encourage the investors to develop new tourist and recreation projects.

Keywords: Kuwait, public-private, partnership, tourism, 2035 vision

Procedia PDF Downloads 88
2898 Anajaa-Visual Substitution System: A Navigation Assistive Device for the Visually Impaired

Authors: Juan Pablo Botero Torres, Alba Avila, Luis Felipe Giraldo

Abstract:

Independent navigation and mobility through unknown spaces pose a challenge for the autonomy of visually impaired people (VIP), who have relied on the use of traditional assistive tools like the white cane and trained dogs. However, emerging visually assistive technologies (VAT) have proposed several human-machine interfaces (HMIs) that could improve VIP’s ability for self-guidance. Hereby, we introduce the design and implementation of a visually assistive device, Anajaa – Visual Substitution System (AVSS). This system integrates ultrasonic sensors with custom electronics, and computer vision models (convolutional neural networks), in order to achieve a robust system that acquires information of the surrounding space and transmits it to the user in an intuitive and efficient manner. AVSS consists of two modules: the sensing and the actuation module, which are fitted to a chest mount and belt that communicate via Bluetooth. The sensing module was designed for the acquisition and processing of proximity signals provided by an array of ultrasonic sensors. The distribution of these within the chest mount allows an accurate representation of the surrounding space, discretized in three different levels of proximity, ranging from 0 to 6 meters. Additionally, this module is fitted with an RGB-D camera used to detect potentially threatening obstacles, like staircases, using a convolutional neural network specifically trained for this purpose. Posteriorly, the depth data is used to estimate the distance between the stairs and the user. The information gathered from this module is then sent to the actuation module that creates an HMI, by the means of a 3x2 array of vibration motors that make up the tactile display and allow the system to deliver haptic feedback. The actuation module uses vibrational messages (tactones); changing both in amplitude and frequency to deliver different awareness levels according to the proximity of the obstacle. This enables the system to deliver an intuitive interface. Both modules were tested under lab conditions, and the HMI was additionally tested with a focal group of VIP. The lab testing was conducted in order to establish the processing speed of the computer vision algorithms. This experimentation determined that the model can process 0.59 frames per second (FPS); this is considered as an adequate processing speed taking into account that the walking speed of VIP is 1.439 m/s. In order to test the HMI, we conducted a focal group composed of two females and two males between the ages of 35-65 years. The subject selection was aided by the Colombian Cooperative of Work and Services for the Sightless (COOTRASIN). We analyzed the learning process of the haptic messages throughout five experimentation sessions using two metrics: message discrimination and localization success. These correspond to the ability of the subjects to recognize different tactones and locate them within the tactile display. Both were calculated as the mean across all subjects. Results show that the focal group achieved message discrimination of 70% and a localization success of 80%, demonstrating how the proposed HMI leads to the appropriation and understanding of the feedback messages, enabling the user’s awareness of its surrounding space.

Keywords: computer vision on embedded systems, electronic trave aids, human-machine interface, haptic feedback, visual assistive technologies, vision substitution systems

Procedia PDF Downloads 45