Search results for: enterprise intelligence
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2113

Search results for: enterprise intelligence

853 Geovisualisation for Defense Based on a Deep Learning Monocular Depth Reconstruction Approach

Authors: Daniel R. dos Santos, Mateus S. Maldonado, Estevão J. R. Batista

Abstract:

The military commanders increasingly dependent on spatial awareness, as knowing where enemy are, understanding how war battle scenarios change over time, and visualizing these trends in ways that offer insights for decision-making. Thanks to advancements in geospatial technologies and artificial intelligence algorithms, the commanders are now able to modernize military operations on a universal scale. Thus, geovisualisation has become an essential asset in the defense sector. It has become indispensable for better decisionmaking in dynamic/temporal scenarios, operation planning and management for the war field, situational awareness, effective planning, monitoring, and others. For example, a 3D visualization of war field data contributes to intelligence analysis, evaluation of postmission outcomes, and creation of predictive models to enhance decision-making and strategic planning capabilities. However, old-school visualization methods are slow, expensive, and unscalable. Despite modern technologies in generating 3D point clouds, such as LIDAR and stereo sensors, monocular depth values based on deep learning can offer a faster and more detailed view of the environment, transforming single images into visual information for valuable insights. We propose a dedicated monocular depth reconstruction approach via deep learning techniques for 3D geovisualisation of satellite images. It introduces scalability in terrain reconstruction and data visualization. First, a dataset with more than 7,000 satellite images and associated digital elevation model (DEM) is created. It is based on high resolution optical and radar imageries collected from Planet and Copernicus, on which we fuse highresolution topographic data obtained using technologies such as LiDAR and the associated geographic coordinates. Second, we developed an imagery-DEM fusion strategy that combine feature maps from two encoder-decoder networks. One network is trained with radar and optical bands, while the other is trained with DEM features to compute dense 3D depth. Finally, we constructed a benchmark with sparse depth annotations to facilitate future research. To demonstrate the proposed method's versatility, we evaluated its performance on no annotated satellite images and implemented an enclosed environment useful for Geovisualisation applications. The algorithms were developed in Python 3.0, employing open-source computing libraries, i.e., Open3D, TensorFlow, and Pythorch3D. The proposed method provides fast and accurate decision-making with GIS for localization of troops, position of the enemy, terrain and climate conditions. This analysis enhances situational consciousness, enabling commanders to fine-tune the strategies and distribute the resources proficiently.

Keywords: depth, deep learning, geovisualisation, satellite images

Procedia PDF Downloads 10
852 Modeling Visual Memorability Assessment with Autoencoders Reveals Characteristics of Memorable Images

Authors: Elham Bagheri, Yalda Mohsenzadeh

Abstract:

Image memorability refers to the phenomenon where certain images are more likely to be remembered by humans than others. It is a quantifiable and intrinsic attribute of an image. Understanding how visual perception and memory interact is important in both cognitive science and artificial intelligence. It reveals the complex processes that support human cognition and helps to improve machine learning algorithms by mimicking the brain's efficient data processing and storage mechanisms. To explore the computational underpinnings of image memorability, this study examines the relationship between an image's reconstruction error, distinctiveness in latent space, and its memorability score. A trained autoencoder is used to replicate human-like memorability assessment inspired by the visual memory game employed in memorability estimations. This study leverages a VGG-based autoencoder that is pre-trained on the vast ImageNet dataset, enabling it to recognize patterns and features that are common to a wide and diverse range of images. An empirical analysis is conducted using the MemCat dataset, which includes 10,000 images from five broad categories: animals, sports, food, landscapes, and vehicles, along with their corresponding memorability scores. The memorability score assigned to each image represents the probability of that image being remembered by participants after a single exposure. The autoencoder is finetuned for one epoch with a batch size of one, attempting to create a scenario similar to human memorability experiments where memorability is quantified by the likelihood of an image being remembered after being seen only once. The reconstruction error, which is quantified as the difference between the original and reconstructed images, serves as a measure of how well the autoencoder has learned to represent the data. The reconstruction error of each image, the error reduction, and its distinctiveness in latent space are calculated and correlated with the memorability score. Distinctiveness is measured as the Euclidean distance between each image's latent representation and its nearest neighbor within the autoencoder's latent space. Different structural and perceptual loss functions are considered to quantify the reconstruction error. The results indicate that there is a strong correlation between the reconstruction error and the distinctiveness of images and their memorability scores. This suggests that images with more unique distinct features that challenge the autoencoder's compressive capacities are inherently more memorable. There is also a negative correlation between the reduction in reconstruction error compared to the autoencoder pre-trained on ImageNet, which suggests that highly memorable images are harder to reconstruct, probably due to having features that are more difficult to learn by the autoencoder. These insights suggest a new pathway for evaluating image memorability, which could potentially impact industries reliant on visual content and mark a step forward in merging the fields of artificial intelligence and cognitive science. The current research opens avenues for utilizing neural representations as instruments for understanding and predicting visual memory.

Keywords: autoencoder, computational vision, image memorability, image reconstruction, memory retention, reconstruction error, visual perception

Procedia PDF Downloads 90
851 Nest-Building Using Place Cells for Spatial Navigation in an Artificial Neural Network

Authors: Thomas E. Portegys

Abstract:

An animal behavior problem is presented in the form of a nest-building task that involves two cooperating virtual birds, a male and female. The female builds a nest into which she lays an egg. The male's job is to forage in a forest for food for both himself and the female. In addition, the male must fetch stones from a nearby desert for the female to use as nesting material. The task is completed when the nest is built, and an egg is laid in it. A goal-seeking neural network and a recurrent neural network were trained and tested with little success. The goal-seeking network was then enhanced with “place cells”, allowing the birds to spatially navigate the world, building the nest while keeping themselves fed. Place cells are neurons in the hippocampus that map space.

Keywords: artificial animal intelligence, artificial life, goal-seeking neural network, nest-building, place cells, spatial navigation

Procedia PDF Downloads 59
850 Improving Fingerprinting-Based Localization System Using Generative AI

Authors: Getaneh Berie Tarekegn, Li-Chia Tai

Abstract:

With the rapid advancement of artificial intelligence, low-power built-in sensors on Internet of Things devices, and communication technologies, location-aware services have become increasingly popular and have permeated every aspect of people’s lives. Global navigation satellite systems (GNSSs) are the default method of providing continuous positioning services for ground and aerial vehicles, as well as consumer devices (smartphones, watches, notepads, etc.). However, the environment affects satellite positioning systems, particularly indoors, in dense urban and suburban cities enclosed by skyscrapers, or when deep shadows obscure satellite signals. This is because (1) indoor environments are more complicated due to the presence of many objects surrounding them; (2) reflection within the building is highly dependent on the surrounding environment, including the positions of objects and human activity; and (3) satellite signals cannot be reached in an indoor environment, and GNSS doesn't have enough power to penetrate building walls. GPS is also highly power-hungry, which poses a severe challenge for battery-powered IoT devices. Due to these challenges, IoT applications are limited. Consequently, precise, seamless, and ubiquitous Positioning, Navigation and Timing (PNT) systems are crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarms, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 42
849 Corporate Governance Development in Mongolia: The Role of Professional Accountants

Authors: Ernest Nweke

Abstract:

The work of Professional Accountants and Corporate governance are synonymous and cannot be divorced from each other. Organizations, profit and non-profit alike cannot implement sound corporate practices without inputs from Professional Accountants. In today’s dynamic corporate world, good corporate governance practice is a sine qua non. More so, following the corporate failures of the past decades like Enron and WorldCom, governments around the world, including Mongolia are becoming more proactive in ensuring sound corporate governance mechanisms. In the past fifteen years, the Mongolian government has taken several measures to establish and strengthen internal corporate governance structures in firms. This paper highlights the role of professional accountants and auditors play in ensuring that good corporate governance mechanisms are entrenched in listed companies in Mongolia. Both primary and secondary data are utilized in this research. In collection of primary data, Delphi method was used, securing responses from only knowledgeable senior employees, top managers, and some CEOs. Using this method, a total of 107 top-level company employees and executives randomly selected from 22 companies were surveyed; maximum of 5 and minimum of 4 from each company. These companies cut across several sectors. It was concluded that Professional Accountants play key roles in setting and maintaining firm governance. They do this by ensuring full compliance with all the requirements of good and sound corporate governance, establishing reporting, monitoring and evaluating standards, assisting in the setting up of proper controls, efficient and effective audit systems, sound fraud risk management and putting in place an overall vision for the enterprise. Companies with effective corporate governance mechanisms are usually strong and fraud-resilient. It was also discovered that companies with big 4 audit firms tend to have better governance structures in Mongolia.

Keywords: accountants, corporate disclosure, corporate failure, corporate governance

Procedia PDF Downloads 278
848 A Survey of Response Generation of Dialogue Systems

Authors: Yifan Fan, Xudong Luo, Pingping Lin

Abstract:

An essential task in the field of artificial intelligence is to allow computers to interact with people through natural language. Therefore, researches such as virtual assistants and dialogue systems have received widespread attention from industry and academia. The response generation plays a crucial role in dialogue systems, so to push forward the research on this topic, this paper surveys various methods for response generation. We sort out these methods into three categories. First one includes finite state machine methods, framework methods, and instance methods. The second contains full-text indexing methods, ontology methods, vast knowledge base method, and some other methods. The third covers retrieval methods and generative methods. We also discuss some hybrid methods based knowledge and deep learning. We compare their disadvantages and advantages and point out in which ways these studies can be improved further. Our discussion covers some studies published in leading conferences such as IJCAI and AAAI in recent years.

Keywords: deep learning, generative, knowledge, response generation, retrieval

Procedia PDF Downloads 134
847 Blueprinting of a Normalized Supply Chain Processes: Results in Implementing Normalized Software Systems

Authors: Bassam Istanbouli

Abstract:

With the technology evolving every day and with the increase in global competition, industries are always under the pressure to be the best. They need to provide good quality products at competitive prices, when and how the customer wants them.  In order to achieve this level of service, products and their respective supply chain processes need to be flexible and evolvable; otherwise changes will be extremely expensive, slow and with many combinatorial effects. Those combinatorial effects impact the whole organizational structure, from a management, financial, documentation, logistics and specially the information system Enterprise Requirement Planning (ERP) perspective. By applying the normalized system concept/theory to segments of the supply chain, we believe minimal effects, especially at the time of launching an organization global software project. The purpose of this paper is to point out that if an organization wants to develop a software from scratch or implement an existing ERP software for their business needs and if their business processes are normalized and modular then most probably this will yield to a normalized and modular software system that can be easily modified when the business evolves. Another important goal of this paper is to increase the awareness regarding the design of the business processes in a software implementation project. If the blueprints created are normalized then the software developers and configurators will use those modular blueprints to map them into modular software. This paper only prepares the ground for further studies;  the above concept will be supported by going through the steps of developing, configuring and/or implementing a software system for an organization by using two methods: The Software Development Lifecycle method (SDLC) and the Accelerated SAP implementation method (ASAP). Both methods start with the customer requirements, then blue printing of its business processes and finally mapping those processes into a software system.  Since those requirements and processes are the starting point of the implementation process, then normalizing those processes will end up in a normalizing software.

Keywords: blueprint, ERP, modular, normalized

Procedia PDF Downloads 139
846 Identifying Confirmed Resemblances in Problem-Solving Engineering, Both in the Past and Present

Authors: Colin Schmidt, Adrien Lecossier, Pascal Crubleau, Philippe Blanchard, Simon Richir

Abstract:

Introduction:The widespread availability of artificial intelligence, exemplified by Generative Pre-trained Transformers (GPT) relying on large language models (LLM), has caused a seismic shift in the realm of knowledge. Everyone now has the capacity to swiftly learn how these models can either serve them well or not. Today, conversational AI like ChatGPT is grounded in neural transformer models, a significant advance in natural language processing facilitated by the emergence of renowned LLMs constructed using neural transformer architecture. Inventiveness of an LLM : OpenAI's GPT-3 stands as a premier LLM, capable of handling a broad spectrum of natural language processing tasks without requiring fine-tuning, reliably producing text that reads as if authored by humans. However, even with an understanding of how LLMs respond to questions asked, there may be lurking behind OpenAI’s seemingly endless responses an inventive model yet to be uncovered. There may be some unforeseen reasoning emerging from the interconnection of neural networks here. Just as a Soviet researcher in the 1940s questioned the existence of Common factors in inventions, enabling an Under standing of how and according to what principles humans create them, it is equally legitimate today to explore whether solutions provided by LLMs to complex problems also share common denominators. Theory of Inventive Problem Solving (TRIZ) : We will revisit some fundamentals of TRIZ and how Genrich ALTSHULLER was inspired by the idea that inventions and innovations are essential means to solve societal problems. It's crucial to note that traditional problem-solving methods often fall short in discovering innovative solutions. The design team is frequently hampered by psychological barriers stemming from confinement within a highly specialized knowledge domain that is difficult to question. We presume ChatGPT Utilizes TRIZ 40. Hence, the objective of this research is to decipher the inventive model of LLMs, particularly that of ChatGPT, through a comparative study. This will enhance the efficiency of sustainable innovation processes and shed light on how the construction of a solution to a complex problem was devised. Description of the Experimental Protocol : To confirm or reject our main hypothesis that is to determine whether ChatGPT uses TRIZ, we will follow a stringent protocol that we will detail, drawing on insights from a panel of two TRIZ experts. Conclusion and Future Directions : In this endeavor, we sought to comprehend how an LLM like GPT addresses complex challenges. Our goal was to analyze the inventive model of responses provided by an LLM, specifically ChatGPT, by comparing it to an existing standard model: TRIZ 40. Of course, problem solving is our main focus in our endeavours.

Keywords: artificial intelligence, Triz, ChatGPT, inventiveness, problem-solving

Procedia PDF Downloads 74
845 Deep Neural Network Approach for Navigation of Autonomous Vehicles

Authors: Mayank Raj, V. G. Narendra

Abstract:

Ever since the DARPA challenge on autonomous vehicles in 2005, there has been a lot of buzz about ‘Autonomous Vehicles’ amongst the major tech giants such as Google, Uber, and Tesla. Numerous approaches have been adopted to solve this problem, which can have a long-lasting impact on mankind. In this paper, we have used Deep Learning techniques and TensorFlow framework with the goal of building a neural network model to predict (speed, acceleration, steering angle, and brake) features needed for navigation of autonomous vehicles. The Deep Neural Network has been trained on images and sensor data obtained from the comma.ai dataset. A heatmap was used to check for correlation among the features, and finally, four important features were selected. This was a multivariate regression problem. The final model had five convolutional layers, followed by five dense layers. Finally, the calculated values were tested against the labeled data, where the mean squared error was used as a performance metric.

Keywords: autonomous vehicles, deep learning, computer vision, artificial intelligence

Procedia PDF Downloads 158
844 Utilizing Federated Learning for Accurate Prediction of COVID-19 from CT Scan Images

Authors: Jinil Patel, Sarthak Patel, Sarthak Thakkar, Deepti Saraswat

Abstract:

Recently, the COVID-19 outbreak has spread across the world, leading the World Health Organization to classify it as a global pandemic. To save the patient’s life, the COVID-19 symptoms have to be identified. But using an AI (Artificial Intelligence) model to identify COVID-19 symptoms within the allotted time was challenging. The RT-PCR test was found to be inadequate in determining the COVID status of a patient. To determine if the patient has COVID-19 or not, a Computed Tomography Scan (CT scan) of patient is a better alternative. It will be challenging to compile and store all the data from various hospitals on the server, though. Federated learning, therefore, aids in resolving this problem. Certain deep learning models help to classify Covid-19. This paper will have detailed work of certain deep learning models like VGG19, ResNet50, MobileNEtv2, and Deep Learning Aggregation (DLA) along with maintaining privacy with encryption.

Keywords: federated learning, COVID-19, CT-scan, homomorphic encryption, ResNet50, VGG-19, MobileNetv2, DLA

Procedia PDF Downloads 73
843 Intelligent Swarm-Finding in Formation Control of Multi-Robots to Track a Moving Target

Authors: Anh Duc Dang, Joachim Horn

Abstract:

This paper presents a new approach to control robots, which can quickly find their swarm while tracking a moving target through the obstacles of the environment. In this approach, an artificial potential field is generated between each free-robot and the virtual attractive point of the swarm. This artificial potential field will lead free-robots to their swarm. The swarm-finding of these free-robots dose not influence the general motion of their swarm and nor other robots. When one singular robot approaches the swarm then its swarm-search will finish, and it will further participate with its swarm to reach the position of the target. The connections between member-robots with their neighbours are controlled by the artificial attractive/repulsive force field between them to avoid collisions and keep the constant distances between them in ordered formation. The effectiveness of the proposed approach has been verified in simulations.

Keywords: formation control, potential field method, obstacle avoidance, swarm intelligence, multi-agent systems

Procedia PDF Downloads 440
842 A Hybrid Distributed Algorithm for Multi-Objective Dynamic Flexible Job Shop Scheduling Problem

Authors: Aydin Teymourifar, Gurkan Ozturk

Abstract:

In this paper, a hybrid distributed algorithm has been suggested for multi-objective dynamic flexible job shop scheduling problem. The proposed algorithm is high level, in which several algorithms search the space on different machines simultaneously also it is a hybrid algorithm that takes advantages of the artificial intelligence, evolutionary and optimization methods. Distribution is done at different levels and new approaches are used for design of the algorithm. Apache spark and Hadoop frameworks have been used for the distribution of the algorithm. The Pareto optimality approach is used for solving the multi-objective benchmarks. The suggested algorithm that is able to solve large-size problems in short times has been compared with the successful algorithms of the literature. The results prove high speed and efficiency of the algorithm.

Keywords: distributed algorithms, apache-spark, Hadoop, flexible dynamic job shop scheduling, multi-objective optimization

Procedia PDF Downloads 354
841 Digital Environment as a Factor of the City's Competitiveness in Attracting Tourists: The Case of Yekaterinburg

Authors: Alexander S. Burnasov, Anatoly V. Stepanov, Maria Y. Ilyushkina

Abstract:

In the conditions of transition to the digital economy, the digital environment of the city becomes one of the key factors of its tourism attractiveness. Modern digital environment makes travelling more accessible, improves the quality of travel services and the attractiveness of many tourist destinations. The digitalization of the industry allows to use resources more efficiently, to simplify business processes, to minimize risks, and to improve travel safety. The city promotion as a tourist destination in the foreign market becomes decisive in the digital environment. Information technologies are extremely important for the functioning of not only any tourist enterprise but also the city as a whole. In addition to solving traditional problems, it is also possible to implement some innovations from the tourism industry, such as the availability of city services in international systems of booking tickets and booking rooms in hotels, the possibility of early booking of theater and museum tickets, the possibility of non-cash payment by cards of international payment systems, Internet access in the urban environment for travelers. The availability of the city's digital services makes it possible to reduce ordering costs, contributes to the optimal selection of tourist products that meet the requirements of the tourist, provides increased transparency of transactions. The users can compare prices, features, services, and reviews of the travel service. The ability to share impressions with friends thousands of miles away directly affects the image of the city. It is possible to promote the image of the city in the digital environment not only through world-scale events (such as World Cup 2018, international summits, etc.) but also through the creation and management of services in the digital environment aimed at supporting tourism services, which will help to improve the positioning of the city in the global tourism market.

Keywords: competitiveness, digital environment, travelling, Yekaterinburg

Procedia PDF Downloads 136
840 Critical Review of Web Content Mining Extraction Mechanisms

Authors: Rabia Bashir, Sajjad Akbar

Abstract:

There is an inevitable demand of web mining due to rapid increase of huge information on the Internet, but the striking variety of web structures has made required content retrieval a difficult task. To counter this issue, Web Content Mining (WCM) emerges as a potential candidate which extracts and integrates suitable resources of data to users. In past few years, research has been done on several extraction techniques for WCM i.e. agent-based, template-based, assumption-based, statistic-based, wrapper-based and machine learning. However, it is still unclear that either these approaches are efficiently tackling the significant challenges of WCM or not. To answer this question, this paper identifies these challenges such as language independency, structure flexibility, performance, automation, dynamicity, redundancy handling, intelligence, relevant content retrieval, and privacy. Further, mapping of these challenges is done with existing extraction mechanisms which helps to adopt the most suitable WCM approach, given some conditions and characteristics at hand.

Keywords: content mining challenges, web content mining, web content extraction approaches, web information retrieval

Procedia PDF Downloads 548
839 A Thorough Analysis of the Literature on the Airport Service Quality and Patron Satisfaction

Authors: Mohammed Saad Alanazi

Abstract:

Satisfaction of travelers with services provided in the airports is a sign of competitiveness and the corporate image of the airport. This study conducted a systematic literature review of recent studies published after 2017 regarding the factors that positively influence travelers’ satisfaction and encourage them to report positive reviews online. This study found variations among the studies found. They used several research methodologies, and datasets and focused on different airports, yet, they commonly categorized airport services into seven categories that should receive high intention because their qualities were found increasing review rate and positivity. It was found that studies targeting travelers’ satisfaction and intention of revisiting tended to use primary sources of data (survey); meanwhile, studies concerned positivity and negativity of comments towards airport services often used online reviews provided by travelers.

Keywords: business Intelligence, airport service quality, passenger satisfaction, thorough analysis

Procedia PDF Downloads 80
838 Medical Advances in Diagnosing Neurological and Genetic Disorders

Authors: Simon B. N. Thompson

Abstract:

Retinoblastoma is a rare type of childhood genetic cancer that affects children worldwide. The diagnosis is often missed due to lack of education and difficulty in presentation of the tumor. Frequently, the tumor on the retina is noticed by photography when the red-eye flash, commonly seen in normal eyes, is not produced. Instead, a yellow or white colored patch is seen or the child has a noticeable strabismus. Early detection can be life-saving though often results in removal of the affected eye. Remaining functioning in the healthy eye when the child is young has resulted in super-vision and high or above-average intelligence. Technological advancement of cameras has helped in early detection. Brain imaging has also made possible early detection of neurological diseases and, together with the monitoring of cortisol levels and yawning frequency, promises to be the next new early diagnostic tool for the detection of neurological diseases where cortisol insufficiency is particularly salient, such as multiple sclerosis and Cushing’s disease.

Keywords: cortisol, neurological disease, retinoblastoma, Thompson cortisol hypothesis, yawning

Procedia PDF Downloads 386
837 DURAFILE: A Collaborative Tool for Preserving Digital Media Files

Authors: Santiago Macho, Miquel Montaner, Raivo Ruusalepp, Ferran Candela, Xavier Tarres, Rando Rostok

Abstract:

During our lives, we generate a lot of personal information such as photos, music, text documents and videos that link us with our past. This data that used to be tangible is now digital information stored in our computers, which implies a software dependence to make them accessible in the future. Technology, however, constantly evolves and goes through regular shifts, quickly rendering various file formats obsolete. The need for accessing data in the future affects not only personal users but also organizations. In a digital environment, a reliable preservation plan and the ability to adapt to fast changing technology are essential for maintaining data collections in the long term. We present in this paper the European FP7 project called DURAFILE that provides the technology to preserve media files for personal users and organizations while maintaining their quality.

Keywords: artificial intelligence, digital preservation, social search, digital preservation plans

Procedia PDF Downloads 445
836 Smoker Recognition from Lung X-Ray Images Using Convolutional Neural Network

Authors: Moumita Chanda, Md. Fazlul Karim Patwary

Abstract:

Smoking is one of the most popular recreational drug use behaviors, and it contributes to birth defects, COPD, heart attacks, and erectile dysfunction. To completely eradicate this disease, it is imperative that it be identified and treated. Numerous smoking cessation programs have been created, and they demonstrate how beneficial it may be to help someone stop smoking at the ideal time. A tomography meter is an effective smoking detector. Other wearables, such as RF-based proximity sensors worn on the collar and wrist to detect when the hand is close to the mouth, have been proposed in the past, but they are not impervious to deceptive variables. In this study, we create a machine that can discriminate between smokers and non-smokers in real-time with high sensitivity and specificity by watching and collecting the human lung and analyzing the X-ray data using machine learning. If it has the highest accuracy, this machine could be utilized in a hospital, in the selection of candidates for the army or police, or in university entrance.

Keywords: CNN, smoker detection, non-smoker detection, OpenCV, artificial Intelligence, X-ray Image detection

Procedia PDF Downloads 84
835 The Impact of Artificial Intelligence on the Behavior of Children and Autism

Authors: Sara Fayez Fawzy Mikhael

Abstract:

Inclusive education services for students with Autism remains in its early developmental stages in Thailand. Despite many more children with autism are attending schools since the Thai government introduced the Education Provision for People with Disabilities Act in 2008, the services students with autism and their families receive are generally lacking. This quantitative study used Attitude and Preparedness to Teach Students with Autism Scale (APTSAS) to investigate 110 primary school teachers’ attitude and preparedness to teach students with autism in the general education classroom. Descriptive statistical analysis of the data found that student behavior was the most significant factor in building teachers’ negative attitudes students with autism. The majority of teachers also indicated that their pre-service education did not prepare them to meet the learning needs of children with autism in particular, those who are non-verbal. The study is significant and provides direction for enhancing teacher education for inclusivity in Thailand.

Keywords: attitude, autism, teachers, thailandsports activates, movement skills, motor skills

Procedia PDF Downloads 100
834 Parental Investment in Education: A Pathway for the Children's Access to Quality Education

Authors: Tukur Husaini Nahuche

Abstract:

The parent resources play a vital role in the life of the offspring. It help give children basic necessities of life like food, clothing, and housing. In a like manner financial assets allow parents to move into neighborhood with more affluent school systems, to pay school bills, purchase expensive technologies like personal computer, save money for tutoring books, magazines, journals, Newspapers etc. Making of proper provision in the home environment conducive for learning after school hours and creation of other outdoor activities for them are what necessitate in enhancing and accelerating children’s learning opportunities. Indeed, this paper intends to discuss parental investment in education, parent income resources, parental education, occupation, and income as relatively influencing children’s access to quality education. With the hope that families would provide equal opportunities for children irrespective of their sex, intelligence, subject choice,etc.

Keywords: parental investment, children's access, quality education

Procedia PDF Downloads 551
833 Wireless Sensor Anomaly Detection Using Soft Computing

Authors: Mouhammd Alkasassbeh, Alaa Lasasmeh

Abstract:

We live in an era of rapid development as a result of significant scientific growth. Like other technologies, wireless sensor networks (WSNs) are playing one of the main roles. Based on WSNs, ZigBee adds many features to devices, such as minimum cost and power consumption, and increasing the range and connect ability of sensor nodes. ZigBee technology has come to be used in various fields, including science, engineering, and networks, and even in medicinal aspects of intelligence building. In this work, we generated two main datasets, the first being based on tree topology and the second on star topology. The datasets were evaluated by three machine learning (ML) algorithms: J48, meta.j48 and multilayer perceptron (MLP). Each topology was classified into normal and abnormal (attack) network traffic. The dataset used in our work contained simulated data from network simulation 2 (NS2). In each database, the Bayesian network meta.j48 classifier achieved the highest accuracy level among other classifiers, of 99.7% and 99.2% respectively.

Keywords: IDS, Machine learning, WSN, ZigBee technology

Procedia PDF Downloads 543
832 The Impact of Artificial Intelligence on Autism Attitude and Skills

Authors: Samwail Fahmi Francis Yacoub

Abstract:

Inclusive education services for students with Autism remains in its early developmental stages in Thailand. Despite many more children with autism are attending schools since the Thai government introduced the Education Provision for People with Disabilities Act in 2008, the services students with autism and their families receive are generally lacking. This quantitative study used Attitude and Preparedness to Teach Students with Autism Scale (APTSAS) to investigate 110 primary school teachers’ attitude and preparedness to teach students with autism in the general education classroom. Descriptive statistical analysis of the data found that student behavior was the most significant factor in building teachers’ negative attitudes students with autism. The majority of teachers also indicated that their pre-service education did not prepare them to meet the learning needs of children with autism in particular, those who are non-verbal. The study is significant and provides direction for enhancing teacher education for inclusivity in Thailand.

Keywords: attitude, autism, teachers, movement skills, motor skills, children, behavior.

Procedia PDF Downloads 52
831 New Approach for Load Modeling

Authors: Slim Chokri

Abstract:

Load forecasting is one of the central functions in power systems operations. Electricity cannot be stored, which means that for electric utility, the estimate of the future demand is necessary in managing the production and purchasing in an economically reasonable way. A majority of the recently reported approaches are based on neural network. The attraction of the methods lies in the assumption that neural networks are able to learn properties of the load. However, the development of the methods is not finished, and the lack of comparative results on different model variations is a problem. This paper presents a new approach in order to predict the Tunisia daily peak load. The proposed method employs a computational intelligence scheme based on the Fuzzy neural network (FNN) and support vector regression (SVR). Experimental results obtained indicate that our proposed FNN-SVR technique gives significantly good prediction accuracy compared to some classical techniques.

Keywords: neural network, load forecasting, fuzzy inference, machine learning, fuzzy modeling and rule extraction, support vector regression

Procedia PDF Downloads 435
830 Smart Airport: Application of Internet of Things for Confronting Airport Challenges

Authors: Ali Safaeianpour, Nima Shamandi

Abstract:

As air traffic expands, many airports have evolved into transit centers for people, information, and commerce, and technology implementation is an absolute part of airport development. Several challenges are in the way of implementing technology in an airport. Airport 4.0 proposes the "Smart Airport" concept, which focuses on using modern technologies such as Big Data, the Internet of Things (IoT), advanced biometric systems, blockchain, and cloud computing to alter and enhance passengers' journeys. Several common IoT concrete topics as partial keys to smart airports are discussed and introduced, ranging from automated check-in systems to exterior tracking processes, with the goal of enlightening more and more insightful ideas and proposals about smart airport solutions. IoT will dramatically alter people's lives by infusing intelligence, boosting the quality of life, and assembling it smarter. This paper reviews the approaches to transforming an airport into a smart airport and describes several enabling components of IoT and challenges that can hinder the implementation of a smart airport's function, which require to be addressed.

Keywords: airport 4.0, digital airport, smart airport, IoT

Procedia PDF Downloads 113
829 (Re)Framing the Muslim Subject: Studying the Artistic Representation of Guantanamo and Abu Ghraib Detainees

Authors: Iqra Raza

Abstract:

This paper attempts to conceptualize the (de)humanization of the Muslim subject in Karen J. Greenberg and Janet Hamlin’s transmedia Sketching Guantanamo through a close study of the aesthetics and semiotics of the text. The Muslim experience, the paper shall argue, is mediated through a (de)humanization confined and incarcerated within the chains of artistic representation. Hamlin’s reliance on the distortions offered by stereotypes is reminiscent of the late Victorian epistemology on criminality, as evidenced most starkly in the sketch of Khalid Sheikh Mohammad. The position of the white artist thus becomes suspect in the enterprise of neo-Victorian ethnography. The visual stories of movement from within Guantanamo become potent; the paper shall argue, especially in juxtaposition with the images of stillness that came out from the detention centers, which portrayed the enactment of violence on individual bodies with a deliberate erasure of faces. So, while art becomes a way for reclaiming subjectivity or humanizing these identifiable bodies, the medium predicates itself on their objectification. The paper shall explore various questions about what it means for the (criminal?) subjects to be rendered into art rather than being photographed. Does art entail a necessary departure from the assumed objectivity of the photographic images? What makes art the preferred medium for (de)humanization of the violated Muslim bodies? What happens when art is produced without a recognition of the ‘precariousness’ of the life being portrayed? Rendering the detainees into art becomes a slippery task complicated by Hamlin’s privileged position outside the glass walls of the court. The paper shall adjourn analysis at the many dichotomies that exist in the text viz. between the White men and the brown, the Muslims and the Christians, Occident and the Orient problematized by Hamlin’s politics, that of a ‘neutral outsider’ which quickly turns on its head and becomes complicity in her deliberate erasure of the violence that shaped and still shapes Guantanamo.

Keywords: Abu Ghraib, Derrida, Guantanamo, graphic journalism, Muslimness, orient, spectrality

Procedia PDF Downloads 154
828 Settlement Prediction for Tehran Subway Line-3 via FLAC3D and ANFIS

Authors: S. A. Naeini, A. Khalili

Abstract:

Nowadays, tunnels with different applications are developed, and most of them are related to subway tunnels. The excavation of shallow tunnels that pass under municipal utilities is very important, and the surface settlement control is an important factor in the design. The study sought to analyze the settlement and also to find an appropriate model in order to predict the behavior of the tunnel in Tehran subway line-3. The displacement in these sections is also determined by using numerical analyses and numerical modeling. In addition, the Adaptive Neuro-Fuzzy Inference System (ANFIS) method is utilized by Hybrid training algorithm. The database pertinent to the optimum network was obtained from 46 subway tunnels in Iran and Turkey which have been constructed by the new Austrian tunneling method (NATM) with similar parameters based on type of their soil. The surface settlement was measured, and the acquired results were compared to the predicted values. The results disclosed that computing intelligence is a good substitute for numerical modeling.

Keywords: settlement, Subway Line, FLAC3D, ANFIS Method

Procedia PDF Downloads 233
827 A Review Paper on Data Security in Precision Agriculture Using Internet of Things

Authors: Tonderai Muchenje, Xolani Mkhwanazi

Abstract:

Precision agriculture uses a number of technologies, devices, protocols, and computing paradigms to optimize agricultural processes. Big data, artificial intelligence, cloud computing, and edge computing are all used to handle the huge amounts of data generated by precision agriculture. However, precision agriculture is still emerging and has a low level of security features. Furthermore, future solutions will demand data availability and accuracy as key points to help farmers, and security is important to build robust and efficient systems. Since precision agriculture comprises a wide variety and quantity of resources, security addresses issues such as compatibility, constrained resources, and massive data. Moreover, conventional protection schemes used in the traditional internet may not be useful for agricultural systems, creating extra demands and opportunities. Therefore, this paper aims at reviewing state of the art of precision agriculture security, particularly in open field agriculture, discussing its architecture, describing security issues, and presenting the major challenges and future directions.

Keywords: precision agriculture, security, IoT, EIDE

Procedia PDF Downloads 90
826 Technology Impact in Learning and Teaching English Language Writing

Authors: Laura Naka

Abstract:

The invention of computer writing programs has changed the way of teaching second language writing. This artificial intelligence engine can provide students with feedback on their essays, on their grammatical and spelling errors, convenient writing and editing tools to facilitate student’s writing process. However, it is not yet proved if this technology is helping students to improve their writing skills. There are several programs that are of great assistance for students concerning their writing skills. New technology provides students with different software programs which enable them to be more creative, to express their opinions and ideas in words, pictures and sounds, but at the end main and most correct feedback should be given by their teachers. No matter how new technology affects in writing skills, always comes from their teachers. This research will try to present some of the advantages and disadvantages that new technology has in writing process for students. The research takes place in the University of Gjakova ‘’Fehmi Agani’’ Faculty of Education-Preschool Program. The research aims to provide random sample response by using questionnaires and observation.

Keywords: English language learning, technology, academic writing, teaching L2.

Procedia PDF Downloads 571
825 The Significance of ‘Practice’ in Art Research: Indian and Western Perspective

Authors: Mukta Avachat-Shirke

Abstract:

The process of manifestation in art has been studied deeply by various Indian and Western philosophers through times. In the art of painting, ‘Practice’ is always considered as techniques or making and ‘Theory’ is related to intelligence or the ‘conceptual.' The question about the significance of ‘Practice’ in artistic research has been a topic of debate. The aim of this qualitative study is to find the relevance of practice and theory while creating artworks. This study analyzes the thoughts and philosophy of Abhinavgupta, Hegel, and Croce to find a new perspective for looking at practice and theory within artistic research. With the method of grounded theory, the study attempts to establish the importance of both in artistic research. It discusses the issues like stages of creating art, role of tacit knowledge and importance of the decision-making the ability of the artist. This comparative analysis of these three philosophers along with the present systems can be used as a point of reference for further developments in the pedagogy of art research and artists, to understand the psychology and to follow the process of creativity effectively.

Keywords: artistic research, Indian philosophy, practice, Western Philosophy

Procedia PDF Downloads 299
824 Black Swans Public Administration and Informatics

Authors: Anastasis Petrou

Abstract:

Black Swan Theories (BSTs) have existed since the 2nd Century BC. However, problematisation in the interdisciplinary field of Public Administration and Informatics (PA&I) about the impact of Black Swans as rare events in Society is a more recent phenomenon but with a growing, although dispersed, body of research literature. This paper offers a synopsis of core issues and questions raised in PA&I literature about the impacts of rare events in Society, the need for knowledge accumulation and explainability processes about rare events and asks what could help explain the occurrence, severity, heterogeneity, overall impact of Black Swans and the challenges they represent to established scientific methods. The second part of the paper considers how the use of Artificial Intelligence (AI) could assist researchers in better explaining rare events in PA&I. However, the research shows that whilst AI use at the start of knowledge accumulation and explainability processes about rare events is beneficial it is also fraught with challenges discussed herein. The paper concludes with recommendations for future research.

Keywords: black swans, public administration, AI, informatics

Procedia PDF Downloads 15