Search results for: adaptive algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4316

Search results for: adaptive algorithm

986 Study on Errors in Estimating the 3D Gaze Point for Different Pupil Sizes Using Eye Vergences

Authors: M. Pomianek, M. Piszczek, M. Maciejewski

Abstract:

The binocular eye tracking technology is increasingly being used in industry, entertainment and marketing analysis. In the case of virtual reality, eye tracking systems are already the basis for user interaction with the environment. In such systems, the high accuracy of determining the user's eye fixation point is very important due to the specificity of the virtual reality head-mounted display (HMD). Often, however, there are unknown errors occurring in the used eye tracking technology, as well as those resulting from the positioning of the devices in relation to the user's eyes. However, can the virtual environment itself influence estimation errors? The paper presents mathematical analyses and empirical studies of the determination of the fixation point and errors resulting from the change in the size of the pupil in response to the intensity of the displayed scene. The article contains both static laboratory tests as well as on the real user. Based on the research results, optimization solutions were proposed that would reduce the errors of gaze estimation errors. Studies show that errors in estimating the fixation point of vision can be minimized both by improving the pupil positioning algorithm in the video image and by using more precise methods to calibrate the eye tracking system in three-dimensional space.

Keywords: eye tracking, fixation point, pupil size, virtual reality

Procedia PDF Downloads 110
985 A Real Time Ultra-Wideband Location System for Smart Healthcare

Authors: Mingyang Sun, Guozheng Yan, Dasheng Liu, Lei Yang

Abstract:

Driven by the demand of intelligent monitoring in rehabilitation centers or hospitals, a high accuracy real-time location system based on UWB (ultra-wideband) technology was proposed. The system measures precise location of a specific person, traces his movement and visualizes his trajectory on the screen for doctors or administrators. Therefore, doctors could view the position of the patient at any time and find them immediately and exactly when something emergent happens. In our design process, different algorithms were discussed, and their errors were analyzed. In addition, we discussed about a , simple but effective way of correcting the antenna delay error, which turned out to be effective. By choosing the best algorithm and correcting errors with corresponding methods, the system attained a good accuracy. Experiments indicated that the ranging error of the system is lower than 7 cm, the locating error is lower than 20 cm, and the refresh rate exceeds 5 times per second. In future works, by embedding the system in wearable IoT (Internet of Things) devices, it could provide not only physical parameters, but also the activity status of the patient, which would help doctors a lot in performing healthcare.

Keywords: intelligent monitoring, ultra-wideband technology, real-time location, IoT devices, smart healthcare

Procedia PDF Downloads 111
984 Implementing Mindfulness into Wellness Plans: Assisting Individuals with Substance Abuse and Addiction

Authors: Michele M. Mahr

Abstract:

The purpose of this study is to educate, inform, and facilitate scholarly conversation and discussion regarding the implementation of mindfulness techniques when working with individuals with substance use disorder (SUD) or addictive behaviors in mental health. Mindfulness can be recognized as the present moment, non-judgmental awareness, initiated by concentrated attention that is non-reactive and as openheartedly as possible. Individuals with SUD or addiction typically are challenged with triggers, environmental situations, cravings, or social pressures which may deter them from remaining abstinent from their drug of choice or addictive behavior. Also, mindfulness is recognized as one of the cognitive and behavioral treatment approaches and is both a physical and mental practice that encompasses individuals to become aware of internal situations and experiences with undivided attention. That said, mindfulness may be an effective strategy for individuals to employ during these experiences. This study will reveal how mental health practitioners and addiction counselors may find mindfulness to be an essential component of increasing wellness when working with individuals seeking mental health treatment. To this end, mindfulness is simply the ability individuals have to know what is actually happening as it is occurring and what they are experiencing at the moment. In the context of substance abuse and addiction, individuals may employ breathing techniques, meditation, and cognitive restructuring of the mind to become aware of present moment experiences. Furthermore, the notion of mindfulness has been directly connected to the development of neuropathways. The creation of the neural pathways then leads to creating thoughts which leads to developing new coping strategies and adaptive behaviors. Mindfulness strategies can assist individuals in connecting the mind with the body, allowing the individual to remain centered and focused. All of these mentioned above are vital components to recovery during substance abuse and addiction treatment. There are a variety of therapeutic modalities applying the key components of mindfulness, such as Mindfulness-Based Stress Reduction (MBSR) and Mindfulness-Based Cognitive Therapy for depression (MBCT). This study will provide an overview of both MBSR and MBCT in relation to treating individuals with substance abuse and addiction. The author will also provide strategies for readers to employ when working with clients. Lastly, the author will create and foster a safe space for discussion and engaging conversation among participants to ask questions, share perspectives, and be educated on the numerous benefits of mindfulness within wellness.

Keywords: mindfulness, wellness, substance abuse, mental health

Procedia PDF Downloads 46
983 Neighborhood Graph-Optimized Preserving Discriminant Analysis for Image Feature Extraction

Authors: Xiaoheng Tan, Xianfang Li, Tan Guo, Yuchuan Liu, Zhijun Yang, Hongye Li, Kai Fu, Yufang Wu, Heling Gong

Abstract:

The image data collected in reality often have high dimensions, and it contains noise and redundant information. Therefore, it is necessary to extract the compact feature expression of the original perceived image. In this process, effective use of prior knowledge such as data structure distribution and sample label is the key to enhance image feature discrimination and robustness. Based on the above considerations, this paper proposes a local preserving discriminant feature learning model based on graph optimization. The model has the following characteristics: (1) Locality preserving constraint can effectively excavate and preserve the local structural relationship between data. (2) The flexibility of graph learning can be improved by constructing a new local geometric structure graph using label information and the nearest neighbor threshold. (3) The L₂,₁ norm is used to redefine LDA, and the diagonal matrix is introduced as the scale factor of LDA, and the samples are selected, which improves the robustness of feature learning. The validity and robustness of the proposed algorithm are verified by experiments in two public image datasets.

Keywords: feature extraction, graph optimization local preserving projection, linear discriminant analysis, L₂, ₁ norm

Procedia PDF Downloads 127
982 Reliability Analysis of Computer Centre at Yobe State University Using LRU Algorithm

Authors: V. V. Singh, Yusuf Ibrahim Gwanda, Rajesh Prasad

Abstract:

In this paper, we focus on the reliability and performance analysis of Computer Centre (CC) at Yobe State University, Damaturu, Nigeria. The CC consists of three servers: one database mail server, one redundant and one for sharing with the client computers in the CC (called as a local server). Observing the different possibilities of the functioning of the CC, the analysis has been done to evaluate the various popular measures of reliability such as availability, reliability, mean time to failure (MTTF), profit analysis due to the operation of the system. The system can ultimately fail due to the failure of router, redundant server before repairing the mail server and switch failure. The system can also partially fail when a local server fails. The failed devices have restored according to Least Recently Used (LRU) techniques. The system can also fail entirely due to a cooling failure of the server, electricity failure or some natural calamity like earthquake, fire tsunami, etc. All the failure rates are assumed to be constant and follow exponential time distribution, while the repair follows two types of distributions: i.e. general and Gumbel-Hougaard family copula distribution.

Keywords: reliability, availability Gumbel-Hougaard family copula, MTTF, internet data centre

Procedia PDF Downloads 504
981 Tape-Shaped Multiscale Fiducial Marker: A Design Prototype for Indoor Localization

Authors: Marcell Serra de Almeida Martins, Benedito de Souza Ribeiro Neto, Gerson Lima Serejo, Carlos Gustavo Resque Dos Santos

Abstract:

Indoor positioning systems use sensors such as Bluetooth, ZigBee, and Wi-Fi, as well as cameras for image capture, which can be fixed or mobile. These computer vision-based positioning approaches are low-cost to implement, mainly when it uses a mobile camera. The present study aims to create a design of a fiducial marker for a low-cost indoor localization system. The marker is tape-shaped to perform a continuous reading employing two detection algorithms, one for greater distances and another for smaller distances. Therefore, the location service is always operational, even with variations in capture distance. A minimal localization and reading algorithm were implemented for the proposed marker design, aiming to validate it. The accuracy tests consider readings varying the capture distance between [0.5, 10] meters, comparing the proposed marker with others. The tests showed that the proposed marker has a broader capture range than the ArUco and QRCode, maintaining the same size. Therefore, reducing the visual pollution and maximizing the tracking since the ambient can be covered entirely.

Keywords: multiscale recognition, indoor localization, tape-shaped marker, fiducial marker

Procedia PDF Downloads 105
980 On the Construction of Some Optimal Binary Linear Codes

Authors: Skezeer John B. Paz, Ederlina G. Nocon

Abstract:

Finding an optimal binary linear code is a central problem in coding theory. A binary linear code C = [n, k, d] is called optimal if there is no linear code with higher minimum distance d given the length n and the dimension k. There are bounds giving limits for the minimum distance d of a linear code of fixed length n and dimension k. The lower bound which can be taken by construction process tells that there is a known linear code having this minimum distance. The upper bound is given by theoretic results such as Griesmer bound. One way to find an optimal binary linear code is to make the lower bound of d equal to its higher bound. That is, to construct a binary linear code which achieves the highest possible value of its minimum distance d, given n and k. Some optimal binary linear codes were presented by Andries Brouwer in his published table on bounds of the minimum distance d of binary linear codes for 1 ≤ n ≤ 256 and k ≤ n. This was further improved by Markus Grassl by giving a detailed construction process for each code exhibiting the lower bound. In this paper, we construct new optimal binary linear codes by using some construction processes on existing binary linear codes. Particularly, we developed an algorithm applied to the codes already constructed to extend the list of optimal binary linear codes up to 257 ≤ n ≤ 300 for k ≤ 7.

Keywords: bounds of linear codes, Griesmer bound, construction of linear codes, optimal binary linear codes

Procedia PDF Downloads 726
979 Improving Cryptographically Generated Address Algorithm in IPv6 Secure Neighbor Discovery Protocol through Trust Management

Authors: M. Moslehpour, S. Khorsandi

Abstract:

As transition to widespread use of IPv6 addresses has gained momentum, it has been shown to be vulnerable to certain security attacks such as those targeting Neighbor Discovery Protocol (NDP) which provides the address resolution functionality in IPv6. To protect this protocol, Secure Neighbor Discovery (SEND) is introduced. This protocol uses Cryptographically Generated Address (CGA) and asymmetric cryptography as a defense against threats on integrity and identity of NDP. Although SEND protects NDP against attacks, it is computationally intensive due to Hash2 condition in CGA. To improve the CGA computation speed, we parallelized CGA generation process and used the available resources in a trusted network. Furthermore, we focused on the influence of the existence of malicious nodes on the overall load of un-malicious ones in the network. According to the evaluation results, malicious nodes have adverse impacts on the average CGA generation time and on the average number of tries. We utilized a Trust Management that is capable of detecting and isolating the malicious node to remove possible incentives for malicious behavior. We have demonstrated the effectiveness of the Trust Management System in detecting the malicious nodes and hence improving the overall system performance.

Keywords: CGA, ICMPv6, IPv6, malicious node, modifier, NDP, overall load, SEND, trust management

Procedia PDF Downloads 165
978 Manufacturing and Calibration of Material Standards for Optical Microscopy in Industrial Environments

Authors: Alberto Mínguez-Martínez, Jesús De Vicente Y Oliva

Abstract:

It seems that we live in a world in which the trend in industrial environments is the miniaturization of systems and materials and the fabrication of parts at the micro-and nano-scale. The problem arises when manufacturers want to study the quality of their production. This characteristic is becoming crucial due to the evolution of the industry and the development of Industry 4.0. As Industry 4.0 is based on digital models of production and processes, having accurate measurements becomes capital. At this point, the metrology field plays an important role as it is a powerful tool to ensure more stable production to reduce scrap and the cost of non-conformities. The most extended measuring instruments that allow us to carry out accurate measurements at these scales are optical microscopes, whether they are traditional, confocal, focus variation microscopes, profile projectors, or any other similar measurement system. However, the accuracy of measurements is connected to the traceability of them to the SI unit of length (the meter). The fact of providing adequate traceability to 2D and 3D dimensional measurements at micro-and nano-scale in industrial environments is a problem that is being studied, and it does not have a unique answer. In addition, if commercial material standards for micro-and nano-scale are considered, we can find that there are two main problems. On the one hand, those material standards that could be considered complete and very interesting do not give traceability of dimensional measurements and, on the other hand, their calibration is very expensive. This situation implies that these kinds of standards will not succeed in industrial environments and, as a result, they will work in the absence of traceability. To solve this problem in industrial environments, it becomes necessary to have material standards that are easy to use, agile, adaptive to different forms, cheap to manufacture and, of course, traceable to the definition of meter with simple methods. By using these ‘customized standards’, it would be possible to adapt and design measuring procedures for each application and manufacturers will work with some traceability. It is important to note that, despite the fact that this traceability is clearly incomplete, this situation is preferable to working in the absence of it. Recently, it has been demonstrated the versatility and the utility of using laser technology and other AM technologies to manufacture customized material standards. In this paper, the authors propose to manufacture a customized material standard using an ultraviolet laser system and a method to calibrate it. To conclude, the results of the calibration carried out in an accredited dimensional metrology laboratory are presented.

Keywords: industrial environment, material standards, optical measuring instrument, traceability

Procedia PDF Downloads 97
977 Modeling the Acquisition of Expertise in a Sequential Decision-Making Task

Authors: Cristóbal Moënne-Loccoz, Rodrigo C. Vergara, Vladimir López, Domingo Mery, Diego Cosmelli

Abstract:

Our daily interaction with computational interfaces is plagued of situations in which we go from inexperienced users to experts through self-motivated exploration of the same task. In many of these interactions, we must learn to find our way through a sequence of decisions and actions before obtaining the desired result. For instance, when drawing cash from an ATM machine, choices are presented in a step-by-step fashion so that a specific sequence of actions must be performed in order to produce the expected outcome. But, as they become experts in the use of such interfaces, do users adopt specific search and learning strategies? Moreover, if so, can we use this information to follow the process of expertise development and, eventually, predict future actions? This would be a critical step towards building truly adaptive interfaces that can facilitate interaction at different moments of the learning curve. Furthermore, it could provide a window into potential mechanisms underlying decision-making behavior in real world scenarios. Here we tackle this question using a simple game interface that instantiates a 4-level binary decision tree (BDT) sequential decision-making task. Participants have to explore the interface and discover an underlying concept-icon mapping in order to complete the game. We develop a Hidden Markov Model (HMM)-based approach whereby a set of stereotyped, hierarchically related search behaviors act as hidden states. Using this model, we are able to track the decision-making process as participants explore, learn and develop expertise in the use of the interface. Our results show that partitioning the problem space into such stereotyped strategies is sufficient to capture a host of exploratory and learning behaviors. Moreover, using the modular architecture of stereotyped strategies as a Mixture of Experts, we are able to simultaneously ask the experts about the user's most probable future actions. We show that for those participants that learn the task, it becomes possible to predict their next decision, above chance, approximately halfway through the game. Our long-term goal is, on the basis of a better understanding of real-world decision-making processes, to inform the construction of interfaces that can establish dynamic conversations with their users in order to facilitate the development of expertise.

Keywords: behavioral modeling, expertise acquisition, hidden markov models, sequential decision-making

Procedia PDF Downloads 228
976 Steepest Descent Method with New Step Sizes

Authors: Bib Paruhum Silalahi, Djihad Wungguli, Sugi Guritman

Abstract:

Steepest descent method is a simple gradient method for optimization. This method has a slow convergence in heading to the optimal solution, which occurs because of the zigzag form of the steps. Barzilai and Borwein modified this algorithm so that it performs well for problems with large dimensions. Barzilai and Borwein method results have sparked a lot of research on the method of steepest descent, including alternate minimization gradient method and Yuan method. Inspired by previous works, we modified the step size of the steepest descent method. We then compare the modification results against the Barzilai and Borwein method, alternate minimization gradient method and Yuan method for quadratic function cases in terms of the iterations number and the running time. The average results indicate that the steepest descent method with the new step sizes provide good results for small dimensions and able to compete with the results of Barzilai and Borwein method and the alternate minimization gradient method for large dimensions. The new step sizes have faster convergence compared to the other methods, especially for cases with large dimensions.

Keywords: steepest descent, line search, iteration, running time, unconstrained optimization, convergence

Procedia PDF Downloads 521
975 Multiscale Modelling of Textile Reinforced Concrete: A Literature Review

Authors: Anicet Dansou

Abstract:

Textile reinforced concrete (TRC)is increasingly used nowadays in various fields, in particular civil engineering, where it is mainly used for the reinforcement of damaged reinforced concrete structures. TRC is a composite material composed of multi- or uni-axial textile reinforcements coupled with a fine-grained cementitious matrix. The TRC composite is an alternative solution to the traditional Fiber Reinforcement Polymer (FRP) composite. It has good mechanical performance and better temperature stability but also, it makes it possible to meet the criteria of sustainable development better.TRCs are highly anisotropic composite materials with nonlinear hardening behavior; their macroscopic behavior depends on multi-scale mechanisms. The characterization of these materials through numerical simulation has been the subject of many studies. Since TRCs are multiscale material by definition, numerical multi-scale approaches have emerged as one of the most suitable methods for the simulation of TRCs. They aim to incorporate information pertaining to microscale constitute behavior, mesoscale behavior, and macro-scale structure response within a unified model that enables rapid simulation of structures. The computational costs are hence significantly reduced compared to standard simulation at a fine scale. The fine scale information can be implicitly introduced in the macro scale model: approaches of this type are called non-classical. A representative volume element is defined, and the fine scale information are homogenized over it. Analytical and computational homogenization and nested mesh methods belong to these approaches. On the other hand, in classical approaches, the fine scale information are explicitly introduced in the macro scale model. Such approaches pertain to adaptive mesh refinement strategies, sub-modelling, domain decomposition, and multigrid methods This research presents the main principles of numerical multiscale approaches. Advantages and limitations are identified according to several criteria: the assumptions made (fidelity), the number of input parameters required, the calculation costs (efficiency), etc. A bibliographic study of recent results and advances and of the scientific obstacles to be overcome in order to achieve an effective simulation of textile reinforced concrete in civil engineering is presented. A comparative study is further carried out between several methods for the simulation of TRCs used for the structural reinforcement of reinforced concrete structures.

Keywords: composites structures, multiscale methods, numerical modeling, textile reinforced concrete

Procedia PDF Downloads 82
974 Intelligent Rheumatoid Arthritis Identification System Based Image Processing and Neural Classifier

Authors: Abdulkader Helwan

Abstract:

Rheumatoid joint inflammation is characterized as a perpetual incendiary issue which influences the joints by hurting body tissues Therefore, there is an urgent need for an effective intelligent identification system of knee Rheumatoid arthritis especially in its early stages. This paper is to develop a new intelligent system for the identification of Rheumatoid arthritis of the knee utilizing image processing techniques and neural classifier. The system involves two principle stages. The first one is the image processing stage in which the images are processed using some techniques such as RGB to gryascale conversion, rescaling, median filtering, background extracting, images subtracting, segmentation using canny edge detection, and features extraction using pattern averaging. The extracted features are used then as inputs for the neural network which classifies the X-ray knee images as normal or abnormal (arthritic) based on a backpropagation learning algorithm which involves training of the network on 400 X-ray normal and abnormal knee images. The system was tested on 400 x-ray images and the network shows good performance during that phase, resulting in a good identification rate 97%.

Keywords: rheumatoid arthritis, intelligent identification, neural classifier, segmentation, backpropoagation

Procedia PDF Downloads 514
973 Evaluation of Cyclic Thermo-Mechanical Responses of an Industrial Gas Turbine Rotor

Authors: Y. Rae, A. Benaarbia, J. Hughes, Wei Sun

Abstract:

This paper describes an elasto-visco-plastic computational modelling method which can be used to assess the cyclic plasticity responses of high temperature structures operating under thermo-mechanical loadings. The material constitutive equation used is an improved unified multi-axial Chaboche-Lemaitre model, which takes into account non-linear kinematic and isotropic hardening. The computational methodology is a three-dimensional framework following an implicit formulation and based on a radial return mapping algorithm. The associated user material (UMAT) code is developed and calibrated across isothermal hold-time low cycle fatigue tests for a typical turbine rotor steel for use in finite element (FE) implementation. The model is applied to a realistic industrial gas turbine rotor, where the study focuses its attention on the deformation heterogeneities and critical high stress areas within the rotor structure. The potential improvements of such FE visco-plastic approach are discussed. An integrated life assessment procedure based on R5 and visco-plasticity modelling, is also briefly addressed.

Keywords: unified visco-plasticity, thermo-mechanical, turbine rotor, finite element modelling

Procedia PDF Downloads 108
972 Optimal Scheduling of Load and Operational Strategy of a Load Aggregator to Maximize Profit with PEVs

Authors: Md. Shafiullah, Ali T. Al-Awami

Abstract:

This project proposes optimal scheduling of imported power of a load aggregator with the utilization of EVs to maximize its profit. As with the increase of renewable energy resources, electricity price in competitive market becomes more uncertain and, on the other hand, with the penetration of renewable distributed generators in the distribution network the predicted load of a load aggregator also becomes uncertain in real time. Though there is uncertainties in both load and price, the use of EVs storage capacity can make the operation of load aggregator flexible. LA submits its offer to day-ahead market based on predicted loads and optimized use of its EVs to maximize its profit, as well as in real time operation it uses its energy storage capacity in such a way that it can maximize its profit. In this project, load aggregators profit maximization algorithm is formulated and the optimization problem is solved with the help of CVX. As in real time operation the forecasted loads differ from actual load, the mismatches are settled in real time balancing market. Simulation results compare the profit of a load aggregator with a hypothetical group of 1000 EVs and without EVs.

Keywords: CVX, electricity market, load aggregator, load and price uncertainties, profit maximization, real time balancing operation

Procedia PDF Downloads 394
971 Introduction to Various Innovative Techniques Suggested for Seismic Hazard Assessment

Authors: Deepshikha Shukla, C. H. Solanki, Mayank K. Desai

Abstract:

Amongst all the natural hazards, earthquakes have the potential for causing the greatest damages. Since the earthquake forces are random in nature and unpredictable, the quantification of the hazards becomes important in order to assess the hazards. The time and place of a future earthquake are both uncertain. Since earthquakes can neither be prevented nor be predicted, engineers have to design and construct in such a way, that the damage to life and property are minimized. Seismic hazard analysis plays an important role in earthquake design structures by providing a rational value of input parameter. In this paper, both mathematical, as well as computational methods adopted by researchers globally in the past five years, will be discussed. Some mathematical approaches involving the concepts of Poisson’s ratio, Convex Set Theory, Empirical Green’s Function, Bayesian probability estimation applied for seismic hazard and FOSM (first-order second-moment) algorithm methods will be discussed. Computational approaches and numerical model SSIFiBo developed in MATLAB to study dynamic soil-structure interaction problem is discussed in this paper. The GIS-based tool will also be discussed which is predominantly used in the assessment of seismic hazards.

Keywords: computational methods, MATLAB, seismic hazard, seismic measurements

Procedia PDF Downloads 313
970 Quantum Decision Making with Small Sample for Network Monitoring and Control

Authors: Tatsuya Otoshi, Masayuki Murata

Abstract:

With the development and diversification of applications on the Internet, applications that require high responsiveness, such as video streaming, are becoming mainstream. Application responsiveness is not only a matter of communication delay but also a matter of time required to grasp changes in network conditions. The tradeoff between accuracy and measurement time is a challenge in network control. We people make countless decisions all the time, and our decisions seem to resolve tradeoffs between time and accuracy. When making decisions, people are known to make appropriate choices based on relatively small samples. Although there have been various studies on models of human decision-making, a model that integrates various cognitive biases, called ”quantum decision-making,” has recently attracted much attention. However, the modeling of small samples has not been examined much so far. In this paper, we extend the model of quantum decision-making to model decision-making with a small sample. In the proposed model, the state is updated by value-based probability amplitude amplification. By analytically obtaining a lower bound on the number of samples required for decision-making, we show that decision-making with a small number of samples is feasible.

Keywords: quantum decision making, small sample, MPEG-DASH, Grover's algorithm

Procedia PDF Downloads 52
969 The Impact Of Environmental Management System ISO 14001 Adoption on Firm Performance

Authors: Raymond Treacy, Paul Humphreys, Ronan McIvor, Trevor Cadden, Alan McKittrick

Abstract:

This study employed event study methodology to examine the role of institutions, resources and dynamic capabilities in the relationship between the Environmental Management System ISO 14001 adoption and firm performance. Utilising financial data from 140 ISO 14001 certified firms and 320 non-certified firms, the results of the study suggested that the UK and Irish manufacturers were not implementing ISO 14001 solely to gain legitimacy. In contrast, the results demonstrated that firms were fully integrating the ISO 14001 standard within their operations as certified firms were able to improve both financial and operating performance when compared to non-certified firms. However, while there were significant and long lasting improvements for employee productivity, manufacturing cost efficiency, return on assets and sales turnover, the sample firms operating cycle and fixed asset efficiency displayed evidence of diminishing returns in the long-run, underlying the observation that no operating advantage based on incremental improvements can be everlasting. Hence, there is an argument for investing in dynamic capabilities which help renew and refresh the resource base and help the firm adapt to changing environments. Indeed, the results of the regression analysis suggest that dynamic capabilities for innovation acted as a moderator in the relationship between ISO 14001 certification and firm performance. This, in turn, will have a significant and symbiotic influence on sustainability practices within the participating organisations. The study not only provides new and original insights, but demonstrates pragmatically how firms can take advantage of environmental management systems as a moderator to significantly enhance firm performance. However, while it was shown that firm innovation aided both short term and long term ROA performance, adaptive market capabilities only aided firms in the short-term at the marketing strategy deployment stage. Finally, the results have important implications for firms operating in an economic recession as the results suggest that firms should scale back investment in R&D while operating in an economic downturn. Conversely, under normal trading conditions, consistent and long term investments in R&D was found to moderate the relationship between ISO 14001 certification and firm performance. Hence, the results of the study have important implications for academics and management alike.

Keywords: supply chain management, environmental management systems, quality management, sustainability, firm performance

Procedia PDF Downloads 287
968 Parents' Motivating Factors for Their Deaf and Mute Children to Participate in Physical and Recreational Activities

Authors: Ruben L. Tagare, Jr

Abstract:

This study was conducted to determine the parents’ motivating factors for their deaf and mute children to participate in physical and recreational activities. Data were collected from the 17 parents of the deaf and mute children using a specifically designed survey questionnaire as the primary instrument used in the study. Data analysis and interpretation were done with the aid of descriptive statistics, such as frequency, percentage, weighted mean and multiple responses. Most of the respondents were female with a mean average age of 38 years old. The average age of their children was 10 years old. In terms of monthly income, the respondents had an average monthly income of PhP 13,945. Furthermore, most of the respondents lived in the urban area and were all Catholic by faith or religion. As to the factors that parents used to motivate their deaf and mute children to engage in physical and recreational activities, these included the followings: First; to make my child experience and explore more meaningful things through physical and recreational activities; second; to gain other’s respect; third; to build friendship and interact with his peers; fourth; to experience the feeling of belongingness; and fifth: to learn and discover new things. On the other hand, the least chosen factors were: first; to help achieve and maintain a healthy weight; second; to reduce fats and lowering blood pressure; third; to improve balance, coordination and strength; fourth; to improve posture; and fifth; to assist the child in the development of gross motor and fine motor skills. Based on the findings of the study, it is hereby recommended that since the first factor is 'to make my child experience and explore more meaningful things through physical and recreational activities' and the other top factors are more on social aspect, the school should design extra-curricular activities such as theatrical play and other similar activities that the students will find interesting while the parents will be more motivated to engage their children into physical and recreational activities. Also, since the least chosen factors are more on physical aspect, the school should organize or conduct a seminar for the parents to be aware of the benefits of participating in physical and recreational activities for their deaf and mute children. They can also conduct an information campaign to encourage the other parents of deaf and mute children, whom they keep only inside their home to enroll in the school and let their children be exposed to the natural world. Considering that parents are the primary motivators that can best help their children become more interested in physical and recreational activities for their own development, the school should always remain motivated by creating activities for the deaf and mute children with their parents. The study also recommends conducting further study on the level of knowledge/understanding of the parents on the benefits that can be derived from participating in physical and recreational activities.

Keywords: deaf and mute, participation, physical and recreational activities, adaptive PE

Procedia PDF Downloads 159
967 Analyzing Medical Workflows Using Market Basket Analysis

Authors: Mohit Kumar, Mayur Betharia

Abstract:

Healthcare domain, with the emergence of Electronic Medical Record (EMR), collects a lot of data which have been attracting Data Mining expert’s interest. In the past, doctors have relied on their intuition while making critical clinical decisions. This paper presents the means to analyze the Medical workflows to get business insights out of huge dumped medical databases. Market Basket Analysis (MBA) which is a special data mining technique, has been widely used in marketing and e-commerce field to discover the association between products bought together by customers. It helps businesses in increasing their sales by analyzing the purchasing behavior of customers and pitching the right customer with the right product. This paper is an attempt to demonstrate Market Basket Analysis applications in healthcare. In particular, it discusses the Market Basket Analysis Algorithm ‘Apriori’ applications within healthcare in major areas such as analyzing the workflow of diagnostic procedures, Up-selling and Cross-selling of Healthcare Systems, designing healthcare systems more user-friendly. In the paper, we have demonstrated the MBA applications using Angiography Systems, but can be extrapolated to other modalities as well.

Keywords: data mining, market basket analysis, healthcare applications, knowledge discovery in healthcare databases, customer relationship management, healthcare systems

Procedia PDF Downloads 141
966 Algorithmic Generation of Carbon Nanochimneys

Authors: Sorin Muraru

Abstract:

Computational generation of carbon nanostructures is still a very demanding process. This work provides an alternative to manual molecular modeling through an algorithm meant to automate the design of such structures. Specifically, carbon nanochimneys are obtained through the bonding of a carbon nanotube with the smaller edge of an open carbon nanocone. The methods of connection rely on mathematical, geometrical and chemical properties. Non-hexagonal rings are used in order to perform the correct bonding of dangling bonds. Once obtained, they are useful for thermal transport, gas storage or other applications such as gas separation. The carbon nanochimneys are meant to produce a less steep connection between structures such as the carbon nanotube and graphene sheet, as in the pillared graphene, but can also provide functionality on its own. The method relies on connecting dangling bonds at the edges of the two carbon nanostructures, employing the use of two different types of auxiliary structures on a case-by-case basis. The code is implemented in Python 3.7 and generates an output file in the .pdb format containing all the system’s coordinates. Acknowledgment: This work was supported by a grant of the Executive Agency for Higher Education, Research, Development and innovation funding (UEFISCDI), project number PN-III-P1-1.1-TE-2016-24-2, contract TE 122/2018.

Keywords: carbon nanochimneys, computational, carbon nanotube, carbon nanocone, molecular modeling, carbon nanostructures

Procedia PDF Downloads 142
965 Investigation of Extreme Gradient Boosting Model Prediction of Soil Strain-Shear Modulus

Authors: Ehsan Mehryaar, Reza Bushehri

Abstract:

One of the principal parameters defining the clay soil dynamic response is the strain-shear modulus relation. Predicting the strain and, subsequently, shear modulus reduction of the soil is essential for performance analysis of structures exposed to earthquake and dynamic loadings. Many soil properties affect soil’s dynamic behavior. In order to capture those effects, in this study, a database containing 1193 data points consists of maximum shear modulus, strain, moisture content, initial void ratio, plastic limit, liquid limit, initial confining pressure resulting from dynamic laboratory testing of 21 clays is collected for predicting the shear modulus vs. strain curve of soil. A model based on an extreme gradient boosting technique is proposed. A tree-structured parzan estimator hyper-parameter tuning algorithm is utilized simultaneously to find the best hyper-parameters for the model. The performance of the model is compared to the existing empirical equations using the coefficient of correlation and root mean square error.

Keywords: XGBoost, hyper-parameter tuning, soil shear modulus, dynamic response

Procedia PDF Downloads 170
964 Skills Needed Amongst Secondary School Students for Artificial Intelligence Development in Southeast Nigeria

Authors: Chukwuma Mgboji

Abstract:

Since the advent of Artificial Intelligence, robots have become a major stay in developing societies. Robots are deployed in Education, Health, Food and in other spheres of life. Nigeria a country in West Africa has a very low profile in the advancement of Artificial Intelligence especially in the grass roots. The benefits of Artificial intelligence are not fully maximised and harnessed. Advances in artificial intelligence are perceived as impossible or observed as irrelevant. This study seeks to ascertain the needed skills for the development of artificialintelligence amongst secondary schools in Nigeria. The study focused on South East Nigeria with Five states namely Imo, Abia, Ebonyi, Anambra and Enugu. The sample size is 1000 students drawn from Five Government owned Universities offering Computer Science, Computer Education, Electronics Engineering across the Five South East states. Survey method was used to solicit responses from respondents. The findings from the study identified mathematical skills, analytical skills, problem solving skills, computing skills, programming skills, algorithm skills amongst others. The result of this study to the best of the author’s knowledge will be highly beneficial to all stakeholders involved in the advancements and development of artificial intelligence.

Keywords: artificial intelligence, secondary school, robotics, skills

Procedia PDF Downloads 121
963 Modelling and Simulation of a Commercial Thermophilic Biogas Plant

Authors: Jeremiah L. Chukwuneke, Obiora E. Anisiji, Chinonso H. Achebe, Paul C. Okolie

Abstract:

This paper developed a mathematical model of a commercial biogas plant for urban area clean energy requirement. It identified biodegradable waste materials like domestic/city refuse as economically viable alternative source of energy. The mathematical formulation of the proposed gas plant follows the fundamental principles of thermodynamics, and further analyses were accomplished to develop an algorithm for evaluating the plant performance preferably in terms of daily production capacity. In addition, the capacity of the plant is equally estimated for a given cycle of operation and presented in time histories. A nominal 1500 m3 power gas plant was studied characteristically and its performance efficiency evaluated. It was observed that the rate of bio gas production is essentially a function of the reactor temperature, pH, substrate concentration, rate of degradation of the biomass, and the accumulation of matter in the system due to bacteria growth. The results of this study conform to a very large extent with reported empirical data of some existing plant and further model validations were conducted in line with classical records found in literature.

Keywords: energy and mass conservation, specific growth rate, thermophilic bacteria, temperature, rate of bio gas production

Procedia PDF Downloads 416
962 Application of Machine Learning Models to Predict Couchsurfers on Free Homestay Platform Couchsurfing

Authors: Yuanxiang Miao

Abstract:

Couchsurfing is a free homestay and social networking service accessible via the website and mobile app. Couchsurfers can directly request free accommodations from others and receive offers from each other. However, it is typically difficult for people to make a decision that accepts or declines a request when they receive it from Couchsurfers because they do not know each other at all. People are expected to meet up with some Couchsurfers who are kind, generous, and interesting while it is unavoidable to meet up with someone unfriendly. This paper utilized classification algorithms of Machine Learning to help people to find out the Good Couchsurfers and Not Good Couchsurfers on the Couchsurfing website. By knowing the prior experience, like Couchsurfer’s profiles, the latest references, and other factors, it became possible to recognize what kind of the Couchsurfers, and furthermore, it helps people to make a decision that whether to host the Couchsurfers or not. The value of this research lies in a case study in Kyoto, Japan in where the author has hosted 54 Couchsurfers, and the author collected relevant data from the 54 Couchsurfers, finally build a model based on classification algorithms for people to predict Couchsurfers. Lastly, the author offered some feasible suggestions for future research.

Keywords: Couchsurfing, Couchsurfers prediction, classification algorithm, hospitality tourism platform, hospitality sciences, machine learning

Procedia PDF Downloads 102
961 Machine Learning Assisted Performance Optimization in Memory Tiering

Authors: Derssie Mebratu

Abstract:

As a large variety of micro services, web services, social graphic applications, and media applications are continuously developed, it is substantially vital to design and build a reliable, efficient, and faster memory tiering system. Despite limited design, implementation, and deployment in the last few years, several techniques are currently developed to improve a memory tiering system in a cloud. Some of these techniques are to develop an optimal scanning frequency; improve and track pages movement; identify pages that recently accessed; store pages across each tiering, and then identify pages as a hot, warm, and cold so that hot pages can store in the first tiering Dynamic Random Access Memory (DRAM) and warm pages store in the second tiering Compute Express Link(CXL) and cold pages store in the third tiering Non-Volatile Memory (NVM). Apart from the current proposal and implementation, we also develop a new technique based on a machine learning algorithm in that the throughput produced 25% improved performance compared to the performance produced by the baseline as well as the latency produced 95% improved performance compared to the performance produced by the baseline.

Keywords: machine learning, bayesian optimization, memory tiering, CXL, DRAM

Procedia PDF Downloads 72
960 Use of DNA Barcoding and UPLC-MS to Authenticate Agathosma spp. in South African Herbal Products

Authors: E. Pretorius, A. M. Viljoen, M. van der Bank

Abstract:

Introduction: The phytochemistry of Agathosma crenulata and A. betulina has been studied extensively, while their molecular analysis through DNA barcoding remains virtually unexplored. This technique can confirm the identity of plant species included in a herbal product, thereby ensuring the efficacy of the herbal product and the accuracy of its label. Materials and methods: Authentic Agathosma reference material of A. betulina (n=16) and A. crenulata (n=10) were obtained. Thirteen commercial products were purchased from various health shops around Johannesburg, South Africa, using the search term “Agathosma” or “Buchu.” The plastid regions matK and ycf1 were used to barcode the Buchu products, and BRONX analysis confirmed the taxonomic identity of the samples. UPLC-MS analyses were also performed. Results: Only (30/60) 60% of the traded samples tested from 13 suppliers contained A. betulina in their herbal products. Similar results were also obtained for the UPLC-MS analysis. Conclusion: In this study, we demonstrate the application of DNA barcoding in combination with phytochemical analysis to authenticate herbal products claiming to contain Agathosma plants as an ingredient in their products. This supports manufacturing efforts to ensure that herbal products that are safe for the consumer.

Keywords: Buchu, substitution, barcoding, BRONX algorithm, matK, ycf1, UPLC-MS

Procedia PDF Downloads 107
959 Manufacturing Anomaly Detection Using a Combination of Gated Recurrent Unit Network and Random Forest Algorithm

Authors: Atinkut Atinafu Yilma, Eyob Messele Sefene

Abstract:

Anomaly detection is one of the essential mechanisms to control and reduce production loss, especially in today's smart manufacturing. Quick anomaly detection aids in reducing the cost of production by minimizing the possibility of producing defective products. However, developing an anomaly detection model that can rapidly detect a production change is challenging. This paper proposes Gated Recurrent Unit (GRU) combined with Random Forest (RF) to detect anomalies in the production process in real-time quickly. The GRU is used as a feature detector, and RF as a classifier using the input features from GRU. The model was tested using various synthesis and real-world datasets against benchmark methods. The results show that the proposed GRU-RF outperforms the benchmark methods with the shortest time taken to detect anomalies in the production process. Based on the investigation from the study, this proposed model can eliminate or reduce unnecessary production costs and bring a competitive advantage to manufacturing industries.

Keywords: anomaly detection, multivariate time series data, smart manufacturing, gated recurrent unit network, random forest

Procedia PDF Downloads 81
958 Comparison of Deep Convolutional Neural Networks Models for Plant Disease Identification

Authors: Megha Gupta, Nupur Prakash

Abstract:

Identification of plant diseases has been performed using machine learning and deep learning models on the datasets containing images of healthy and diseased plant leaves. The current study carries out an evaluation of some of the deep learning models based on convolutional neural network (CNN) architectures for identification of plant diseases. For this purpose, the publicly available New Plant Diseases Dataset, an augmented version of PlantVillage dataset, available on Kaggle platform, containing 87,900 images has been used. The dataset contained images of 26 diseases of 14 different plants and images of 12 healthy plants. The CNN models selected for the study presented in this paper are AlexNet, ZFNet, VGGNet (four models), GoogLeNet, and ResNet (three models). The selected models are trained using PyTorch, an open-source machine learning library, on Google Colaboratory. A comparative study has been carried out to analyze the high degree of accuracy achieved using these models. The highest test accuracy and F1-score of 99.59% and 0.996, respectively, were achieved by using GoogLeNet with Mini-batch momentum based gradient descent learning algorithm.

Keywords: comparative analysis, convolutional neural networks, deep learning, plant disease identification

Procedia PDF Downloads 168
957 End-to-End Pyramid Based Method for Magnetic Resonance Imaging Reconstruction

Authors: Omer Cahana, Ofer Levi, Maya Herman

Abstract:

Magnetic Resonance Imaging (MRI) is a lengthy medical scan that stems from a long acquisition time. Its length is mainly due to the traditional sampling theorem, which defines a lower boundary for sampling. However, it is still possible to accelerate the scan by using a different approach such as Compress Sensing (CS) or Parallel Imaging (PI). These two complementary methods can be combined to achieve a faster scan with high-fidelity imaging. To achieve that, two conditions must be satisfied: i) the signal must be sparse under a known transform domain, and ii) the sampling method must be incoherent. In addition, a nonlinear reconstruction algorithm must be applied to recover the signal. While the rapid advances in Deep Learning (DL) have had tremendous successes in various computer vision tasks, the field of MRI reconstruction is still in its early stages. In this paper, we present an end-to-end method for MRI reconstruction from k-space to image. Our method contains two parts. The first is sensitivity map estimation (SME), which is a small yet effective network that can easily be extended to a variable number of coils. The second is reconstruction, which is a top-down architecture with lateral connections developed for building high-level refinement at all scales. Our method holds the state-of-art fastMRI benchmark, which is the largest, most diverse benchmark for MRI reconstruction.

Keywords: magnetic resonance imaging, image reconstruction, pyramid network, deep learning

Procedia PDF Downloads 66