Search results for: container based virtualization
26965 Model-Based Control for Piezoelectric-Actuated Systems Using Inverse Prandtl-Ishlinskii Model and Particle Swarm Optimization
Authors: Jin-Wei Liang, Hung-Yi Chen, Lung Lin
Abstract:
In this paper feedforward controller is designed to eliminate nonlinear hysteresis behaviors of a piezoelectric stack actuator (PSA) driven system. The control design is based on inverse Prandtl-Ishlinskii (P-I) hysteresis model identified using particle swarm optimization (PSO) technique. Based on the identified P-I model, both the inverse P-I hysteresis model and feedforward controller can be determined. Experimental results obtained using the inverse P-I feedforward control are compared with their counterparts using hysteresis estimates obtained from the identified Bouc-Wen model. Effectiveness of the proposed feedforward control scheme is demonstrated. To improve control performance feedback compensation using traditional PID scheme is adopted to integrate with the feedforward controller.Keywords: the Bouc-Wen hysteresis model, particle swarm optimization, Prandtl-Ishlinskii model, automation engineering
Procedia PDF Downloads 51426964 Agroforestry Practices on Soil Microbial Biomass Carbon and Organic Carbon in Southern Ethiopia
Authors: Nebiyou Masebo
Abstract:
The rapid conversion of an old aged agroforestry (AF) based agricultural system to monocropping farming system in southern Ethiopia is increasing. The consequence of this, combined with climate change, has been impaired biodiversity, soil microbial biomass carbon (MBC), and soil organic carbon (SOC). The AF system could curb such problems due it is an ecologically and economically sustainable strategies. This study was aimed to investigate different agroforestry practices (AFPs) on MBC and SOC in southern Ethiopia. Soil samples were collected from homegarden based agroforestry practice (HAFP), crop land based agroforestry practice (ClAFP), woodlot based agroforestry practice (WlAFP), and trees on soil and water conservation based agroforestry practice (TSWAFP) using two depth layer (0-30 & 30-60 cm) by systematic sampling. Moreover, woody species inventorywas also collected. The chloroform fumigation extraction method was employed to determine MBC from different AFP types. In this study, the value of MBC and SOC decreased significantly with soil depth (p< 0.05). Besides, AFP type, soil depth, woody species diversity, and key soil properties also strongly influenced MBC and SOC (p< 0.05). In this study, the MBC was the highest (786 mg kg⁻¹ soil) in HAFP, followed by WlAFP (592 mg kg⁻¹ soil), TSWAFP (421 mg kg⁻¹ soil), and ClAFP (357 mg kg⁻¹ soil). The highest mean value of SOC (43.5Mg C ha⁻¹) was recorded in HAFP, followed by WlAFP (35.1Mg C ha⁻¹), TSWAFP (22.3 Mg C ha⁻¹), while the lowest (21.8 Mg C ha⁻¹) was recorded in ClAFP. The HAFP had high woody species diversity, and the lowest was recorded in ClAFP. The finding indicated that SOC and MBC were significantly affected by land management practices, and HAFP has the potential to improve MBC and SOC through good management practices of AFP.Keywords: agroforestry practices, microbial biomass carbon, soil carbon, rapid conversion
Procedia PDF Downloads 10226963 Instructional Coaches' Perceptions of Professional Development: An Exploration of the School-Based Support Program
Authors: Youmen Chaaban, Abdallah Abu-Tineh
Abstract:
This article examines the development of a professional development (PD) model for educator growth and learning that is embedded into the school context. The School based Support Program (SBSP), designed for the Qatari context, targets the practices, knowledge, and skills of both school leadership and teachers in an attempt to improve students’ learning outcomes. Key aspects of the model include the development of learning communities among teachers, strong leadership that supports school improvement activities, and the use of research-based PD to improve teacher practices and student achievement. This paper further presents the results of a qualitative study examining the perceptions of nineteen instructional coaches about the strengths of the PD program, the challenges they face in their day-to-day implementation of the program, and their suggestions for the betterment of the program’s implementation and outcomes. Data were collected from the instructional coaches through open-ended surveys followed by focus group interviews. The instructional coaches reported several strengths, which were compatible with the literature on effective PD. However, the challenges they faced were deeply rooted within the structure of the program, in addition to external factors operating at the school and Ministry of Education levels. Thus, a general consensus on the way the program should ultimately develop was reached.Keywords: situated professional development, school reform, instructional coach, school based support program
Procedia PDF Downloads 35526962 A Weighted Approach to Unconstrained Iris Recognition
Authors: Yao-Hong Tsai
Abstract:
This paper presents a weighted approach to unconstrained iris recognition. Nowadays, commercial systems are usually characterized by strong acquisition constraints based on the subject’s cooperation. However, it is not always achievable for real scenarios in our daily life. Researchers have been focused on reducing these constraints and maintaining the performance of the system by new techniques at the same time. With large variation in the environment, there are two main improvements to develop the proposed iris recognition system. For solving extremely uneven lighting condition, statistic based illumination normalization is first used on eye region to increase the accuracy of iris feature. The detection of the iris image is based on Adaboost algorithm. Secondly, the weighted approach is designed by Gaussian functions according to the distance to the center of the iris. Furthermore, local binary pattern (LBP) histogram is then applied to texture classification with the weight. Experiment showed that the proposed system provided users a more flexible and feasible way to interact with the verification system through iris recognition.Keywords: authentication, iris recognition, adaboost, local binary pattern
Procedia PDF Downloads 22426961 Construction Quality Perception of Construction Professionals and Their Expectations from a Quality Improvement Technique in Pakistan
Authors: Muhammad Yousaf Sadiq
Abstract:
The complexity arises in defining the construction quality due to its perception, based on inherent market conditions and their requirements, the diversified stakeholders itself and their desired output. An quantitative survey based approach was adopted in this constructive study. A questionnaire-based survey was conducted for the assessment of construction Quality perception and expectations in the context of quality improvement technique. The survey feedback of professionals of the leading construction organizations/companies of Pakistan construction industry were analyzed. The financial capacity, organizational structure, and construction experience of the construction firms formed basis for their selection. The quality perception was found to be project-scope-oriented and considered as an excess cost for a construction project. Any quality improvement technique was expected to maximize the profit for the employer, by improving the productivity in a construction project. The study is beneficial for the construction professionals to assess the prevailing construction quality perception and the expectations from implementation of any quality improvement technique in construction projects.Keywords: construction quality, expectation, improvement, perception
Procedia PDF Downloads 47626960 Network Analysis and Sex Prediction based on a full Human Brain Connectome
Authors: Oleg Vlasovets, Fabian Schaipp, Christian L. Mueller
Abstract:
we conduct a network analysis and predict the sex of 1000 participants based on ”connectome” - pairwise Pearson’s correlation across 436 brain parcels. We solve the non-smooth convex optimization problem, known under the name of Graphical Lasso, where the solution includes a low-rank component. With this solution and machine learning model for a sex prediction, we explain the brain parcels-sex connectivity patterns.Keywords: network analysis, neuroscience, machine learning, optimization
Procedia PDF Downloads 14726959 Approach for Demonstrating Reliability Targets for Rail Transport during Low Mileage Accumulation in the Field: Methodology and Case Study
Authors: Nipun Manirajan, Heeralal Gargama, Sushil Guhe, Manoj Prabhakaran
Abstract:
In railway industry, train sets are designed based on contractual requirements (mission profile), where reliability targets are measured in terms of mean distance between failures (MDBF). However, during the beginning of revenue services, trains do not achieve the designed mission profile distance (mileage) within the timeframe due to infrastructure constraints, scarcity of commuters or other operational challenges thereby not respecting the original design inputs. Since trains do not run sufficiently and do not achieve the designed mileage within the specified time, car builder has a risk of not achieving the contractual MDBF target. This paper proposes a constant failure rate based model to deal with the situations where mileage accumulation is not a part of the design mission profile. The model provides appropriate MDBF target to be demonstrated based on actual accumulated mileage. A case study of rolling stock running in the field is undertaken to analyze the failure data and MDBF target demonstration during low mileage accumulation. The results of case study prove that with the proposed method, reliability targets are achieved under low mileage accumulation.Keywords: mean distance between failures, mileage-based reliability, reliability target appropriations, rolling stock reliability
Procedia PDF Downloads 26626958 Fault-Tolerant Predictive Control for Polytopic LPV Systems Subject to Sensor Faults
Authors: Sofiane Bououden, Ilyes Boulkaibet
Abstract:
In this paper, a robust fault-tolerant predictive control (FTPC) strategy is proposed for systems with linear parameter varying (LPV) models and input constraints subject to sensor faults. Generally, virtual observers are used for improving the observation precision and reduce the impacts of sensor faults and uncertainties in the system. However, this type of observer lacks certain system measurements which substantially reduce its accuracy. To deal with this issue, a real observer is then designed based on the virtual observer, and consequently a real observer-based robust predictive control is designed for polytopic LPV systems. Moreover, the proposed observer can entirely assure that all system states and sensor faults are estimated. As a result, and based on both observers, a robust fault-tolerant predictive control is then established via the Lyapunov method where sufficient conditions are proposed, for stability analysis and control purposes, in linear matrix inequalities (LMIs) form. Finally, simulation results are given to show the effectiveness of the proposed approach.Keywords: linear parameter varying systems, fault-tolerant predictive control, observer-based control, sensor faults, input constraints, linear matrix inequalities
Procedia PDF Downloads 20026957 Predicting the Human Impact of Natural Onset Disasters Using Pattern Recognition Techniques and Rule Based Clustering
Authors: Sara Hasani
Abstract:
This research focuses on natural sudden onset disasters characterised as ‘occurring with little or no warning and often cause excessive injuries far surpassing the national response capacities’. Based on the panel analysis of the historic record of 4,252 natural onset disasters between 1980 to 2015, a predictive method was developed to predict the human impact of the disaster (fatality, injured, homeless) with less than 3% of errors. The geographical dispersion of the disasters includes every country where the data were available and cross-examined from various humanitarian sources. The records were then filtered into 4252 records of the disasters where the five predictive variables (disaster type, HDI, DRI, population, and population density) were clearly stated. The procedure was designed based on a combination of pattern recognition techniques and rule-based clustering for prediction and discrimination analysis to validate the results further. The result indicates that there is a relationship between the disaster human impact and the five socio-economic characteristics of the affected country mentioned above. As a result, a framework was put forward, which could predict the disaster’s human impact based on their severity rank in the early hours of disaster strike. The predictions in this model were outlined in two worst and best-case scenarios, which respectively inform the lower range and higher range of the prediction. A necessity to develop the predictive framework can be highlighted by noticing that despite the existing research in literature, a framework for predicting the human impact and estimating the needs at the time of the disaster is yet to be developed. This can further be used to allocate the resources at the response phase of the disaster where the data is scarce.Keywords: disaster management, natural disaster, pattern recognition, prediction
Procedia PDF Downloads 15326956 Gene Prediction in DNA Sequences Using an Ensemble Algorithm Based on Goertzel Algorithm and Anti-Notch Filter
Authors: Hamidreza Saberkari, Mousa Shamsi, Hossein Ahmadi, Saeed Vaali, , MohammadHossein Sedaaghi
Abstract:
In the recent years, using signal processing tools for accurate identification of the protein coding regions has become a challenge in bioinformatics. Most of the genomic signal processing methods is based on the period-3 characteristics of the nucleoids in DNA strands and consequently, spectral analysis is applied to the numerical sequences of DNA to find the location of periodical components. In this paper, a novel ensemble algorithm for gene selection in DNA sequences has been presented which is based on the combination of Goertzel algorithm and anti-notch filter (ANF). The proposed algorithm has many advantages when compared to other conventional methods. Firstly, it leads to identify the coding protein regions more accurate due to using the Goertzel algorithm which is tuned at the desired frequency. Secondly, faster detection time is achieved. The proposed algorithm is applied on several genes, including genes available in databases BG570 and HMR195 and their results are compared to other methods based on the nucleotide level evaluation criteria. Implementation results show the excellent performance of the proposed algorithm in identifying protein coding regions, specifically in identification of small-scale gene areas.Keywords: protein coding regions, period-3, anti-notch filter, Goertzel algorithm
Procedia PDF Downloads 38726955 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Model
Authors: Alam Ali, Ashok Kumar Pathak
Abstract:
Path analysis is a statistical technique used to evaluate the strength of the direct and indirect effects of variables. One or more structural regression equations are used to estimate a series of parameters in order to find the better fit of data. Sometimes, exogenous variables do not show a significant strength of their direct and indirect effect when the assumption of classical regression (ordinary least squares (OLS)) are violated by the nature of the data. The main motive of this article is to investigate the efficacy of the copula-based regression approach over the classical regression approach and calculate the direct and indirect effects of variables when data violates the OLS assumption and variables are linked through an elliptical copula. We perform this study using a well-organized numerical scheme. Finally, a real data application is also presented to demonstrate the performance of the superiority of the copula approach.Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique
Procedia PDF Downloads 7226954 Reed: An Approach Towards Quickly Bootstrapping Multilingual Acoustic Models
Authors: Bipasha Sen, Aditya Agarwal
Abstract:
Multilingual automatic speech recognition (ASR) system is a single entity capable of transcribing multiple languages sharing a common phone space. Performance of such a system is highly dependent on the compatibility of the languages. State of the art speech recognition systems are built using sequential architectures based on recurrent neural networks (RNN) limiting the computational parallelization in training. This poses a significant challenge in terms of time taken to bootstrap and validate the compatibility of multiple languages for building a robust multilingual system. Complex architectural choices based on self-attention networks are made to improve the parallelization thereby reducing the training time. In this work, we propose Reed, a simple system based on 1D convolutions which uses very short context to improve the training time. To improve the performance of our system, we use raw time-domain speech signals directly as input. This enables the convolutional layers to learn feature representations rather than relying on handcrafted features such as MFCC. We report improvement on training and inference times by atleast a factor of 4x and 7.4x respectively with comparable WERs against standard RNN based baseline systems on SpeechOcean's multilingual low resource dataset.Keywords: convolutional neural networks, language compatibility, low resource languages, multilingual automatic speech recognition
Procedia PDF Downloads 12326953 Cryptography Based Authentication Methods
Authors: Mohammad A. Alia, Abdelfatah Aref Tamimi, Omaima N. A. Al-Allaf
Abstract:
This paper reviews a comparison study on the most common used authentication methods. Some of these methods are actually based on cryptography. In this study, we show the main cryptographic services. Also, this study presents a specific discussion about authentication service, since the authentication service is classified into several categorizes according to their methods. However, this study gives more about the real life example for each of the authentication methods. It talks about the simplest authentication methods as well about the available biometric authentication methods such as voice, iris, fingerprint, and face authentication.Keywords: information security, cryptography, system access control, authentication, network security
Procedia PDF Downloads 47126952 Configuration Design and Optimization of the Movable Leg-Foot Lunar Soft-Landing Device
Authors: Shan Jia, Jinbao Chen, Jinhua Zhou, Jiacheng Qian
Abstract:
Lunar exploration is a necessary foundation for deep-space exploration. For the functional limitations of the fixed landers which are widely used currently and are to expand the detection range by the use of wheeled rovers with unavoidable path-repeatability, a movable lunar soft-landing device based on cantilever type buffer mechanism and leg-foot type walking mechanism is presented. Firstly, a 20 DoFs quadruped configuration based on pushrod is proposed. The configuration is of the bionic characteristics such as hip, knee and ankle joints, and can make the kinematics of the whole mechanism unchanged before and after buffering. Secondly, the multi-function main/auxiliary buffers based on crumple-energy absorption and screw-nut mechanism, as well as the telescopic device which could be used to protect the plantar force sensors during the buffer process are designed. Finally, the kinematic model of the whole mechanism is established, and the configuration optimization of the whole mechanism is completed based on the performance requirements of slope adaptation and obstacle crossing. This research can provide a technical solution integrating soft-landing, large-scale inspection and material-transfer for future lunar exploration and even mars exploration, and can also serve as the technical basis for developing the reusable landers.Keywords: configuration design, lunar soft-landing device, movable, optimization
Procedia PDF Downloads 15826951 Streamlines: Paths of Fluid Flow through Sandstone Samples Based on Computed Microtomography
Authors: Ł. Kaczmarek, T. Wejrzanowski, M. Maksimczuk
Abstract:
The study presents the use of the numerical calculations based on high-resolution computed microtomography in analysis of fluid flow through Miocene sandstones. Therefore, the permeability studies of rocks were performed. Miocene samples were taken from well S-3, located in the eastern part of the Carpathian Foredeep. For aforementioned analysis, two series of X-ray irradiation were performed. The first set of samples was selected to obtain the spatial distribution of grains and pores. At this stage of the study length of voxel side amounted 27 microns. The next set of X-ray irradation enabled recognition of microstructural components as well as petrophysical features. The length of voxel side in this stage was up to 2 µm. Based on this study, the samples were broken down into two distinct groups. The first one represents conventional reservoir deposits, in opposite to second one - unconventional type. Appropriate identification of petrophysical parameters such as porosity and permeability of the formation is a key element for optimization of the reservoir development.Keywords: grains, permeability, pores, pressure distribution
Procedia PDF Downloads 25326950 Detection of Nanotoxic Material Using DNA Based QCM
Authors: Juneseok You, Chanho Park, Kuehwan Jang, Sungsoo Na
Abstract:
Sensing of nanotoxic materials is strongly important, as their engineering applications are growing recently and results in that nanotoxic material can harmfully influence human health and environment. In current study we report the quartz crystal microbalance (QCM)-based, in situ and real-time sensing of nanotoxic-material by frequency shift. We propose the in situ detection of nanotoxic material of zinc oxice by using QCM functionalized with a taget-specific DNA. Since the mass of a target material is comparable to that of an atom, the mass change caused by target binding to DNA on the quartz electrode is so small that it is practically difficult to detect the ions at low concentrations. In our study, we have demonstrated the in-situ and fast detection of zinc oxide using the quartz crystal microbalance (QCM). The detection was derived from the DNA hybridization between the DNA on the quartz electrode. The results suggest that QCM-based detection opens a new avenue for the development of a practical water-testing sensor.Keywords: nanotoxic material, qcm, frequency, in situ sensing
Procedia PDF Downloads 42226949 Development of a Congestion Controller of Computer Network Using Artificial Intelligence Algorithm
Authors: Mary Anne Roa
Abstract:
Congestion in network occurs due to exceed in aggregate demand as compared to the accessible capacity of the resources. Network congestion will increase as network speed increases and new effective congestion control methods are needed, especially for today’s very high speed networks. To address this undeniably global issue, the study focuses on the development of a fuzzy-based congestion control model concerned with allocating the resources of a computer network such that the system can operate at an adequate performance level when the demand exceeds or is near the capacity of the resources. Fuzzy logic based models have proven capable of accurately representing a wide variety of processes. The model built is based on bandwidth, the aggregate incoming traffic and the waiting time. The theoretical analysis and simulation results show that the proposed algorithm provides not only good utilization but also low packet loss.Keywords: congestion control, queue management, computer networks, fuzzy logic
Procedia PDF Downloads 39726948 The Estimation of Human Vital Signs Complexity
Authors: L. Bikulciene, E. Venskaityte, G. Jarusevicius
Abstract:
Non-stationary and nonlinear signals generated by living complex systems defy traditional mechanistic approaches, which are based on homeostasis. Previous our studies have shown that the evaluation of the interactions of physiological signals by using special analysis methods is suitable for observation of physiological processes. It is demonstrated the possibility of using deep physiological model, based interpretation of the changes of the human body’s functional states combined with an application of the analytical method based on matrix theory for the physiological signals analysis, which was applied on high risk cardiac patients. It is shown that evaluation of cardiac signals interactions show peculiar for each individual functional changes at the onset of hemodynamic restoration procedure. Therefore we suggest that the alterations of functional state of the body, after patients overcome surgery can be complemented by the data received from the suggested approach of the evaluation of functional variables interactions.Keywords: cardiac diseases, complex systems theory, ECG analysis, matrix analysis
Procedia PDF Downloads 34426947 H.263 Based Video Transceiver for Wireless Camera System
Authors: Won-Ho Kim
Abstract:
In this paper, a design of H.263 based wireless video transceiver is presented for wireless camera system. It uses standard WIFI transceiver and the covering area is up to 100m. Furthermore the standard H.263 video encoding technique is used for video compression since wireless video transmitter is unable to transmit high capacity raw data in real time and the implemented system is capable of streaming at speed of less than 1Mbps using NTSC 720x480 video.Keywords: wireless video transceiver, video surveillance camera, H.263 video encoding digital signal processing
Procedia PDF Downloads 36426946 Frame Camera and Event Camera in Stereo Pair for High-Resolution Sensing
Authors: Khen Cohen, Daniel Yankelevich, David Mendlovic, Dan Raviv
Abstract:
We present a 3D stereo system for high-resolution sensing in both the spatial and the temporal domains by combining a frame-based camera and an event-based camera. We establish a method to merge both devices into one unite system and introduce a calibration process, followed by a correspondence technique and interpolation algorithm for 3D reconstruction. We further provide quantitative analysis about our system in terms of depth resolution and additional parameter analysis. We show experimentally how our system performs temporal super-resolution up to effectively 1ms and can detect fast-moving objects and human micro-movements that can be used for micro-expression analysis. We also demonstrate how our method can extract colored events for an event-based camera without any degradation in the spatial resolution, compared to a colored filter array.Keywords: DVS-CIS stereo vision, micro-movements, temporal super-resolution, 3D reconstruction
Procedia PDF Downloads 29726945 A Modified NSGA-II Algorithm for Solving Multi-Objective Flexible Job Shop Scheduling Problem
Authors: Aydin Teymourifar, Gurkan Ozturk, Ozan Bahadir
Abstract:
NSGA-II is one of the most well-known and most widely used evolutionary algorithms. In addition to its new versions, such as NSGA-III, there are several modified types of this algorithm in the literature. In this paper, a hybrid NSGA-II algorithm has been suggested for solving the multi-objective flexible job shop scheduling problem. For a better search, new neighborhood-based crossover and mutation operators are defined. To create new generations, the neighbors of the selected individuals by the tournament selection are constructed. Also, at the end of each iteration, before sorting, neighbors of a certain number of good solutions are derived, except for solutions protected by elitism. The neighbors are generated using a constraint-based neural network that uses various constructs. The non-dominated sorting and crowding distance operators are same as the classic NSGA-II. A comparison based on some multi-objective benchmarks from the literature shows the efficiency of the algorithm.Keywords: flexible job shop scheduling problem, multi-objective optimization, NSGA-II algorithm, neighborhood structures
Procedia PDF Downloads 22926944 Viral Metagenomics Revealed a Novel Cardiovirus in Feces of Wild Rats
Authors: Asif Mahmood, Shama Shama, Hao Ni, Hao Wang, Yu Ling, Hui Xu, Shixing Yang, Qais Ahmad Naseer, Wen Zhang
Abstract:
Cardiovirus is a genus of viruses belonging to the family Picornaviridae. Here, we used viral metagenomic techniques to detect the viral nucleic acid in the fecal samples from wild rats in Zhenjiang city in China. Fecal samples were collected from 20 wild rats and pooled into four sample pools and then subjected to libraries construction which were then sequenced on Illumina MiSeq platform. The sequenced reads were analyzed using viral metagenomic analysis pipeline. A novel cardiovirus from feces of a wild rat was identified, named amzj-2018, of which the complete genome was acquired. Phylogenetic analysis based on the complete amino acid sequence of polyprotein revealed that amzj-2018 formed a separate branch located between clusters of Saffold virus and Rat Theilovirus 1 (RTV-1). Phylogenetic analysis based on different regions of the polyproteins, including P1, P2, P3, and P2+P3, respectively, showed discordant trees, where the tree based on P3 region indicated that amzj-2018 clustered separately between Theiler's murine encephalomyelitis virus and RTV-1. The complete genome of a cardiovirus was determined from the feces of wild rats which belonged to a novel type of cardiovirus based on phylogenetic analysis. Whether it is associated with disease needs further investigation.Keywords: cardiovirus, viral metagenomics, genomic organization, phylogenetic analysis
Procedia PDF Downloads 1826943 Development of Open Source Geospatial Certification Model Based on Geospatial Technology Competency Model
Authors: Tanzeel Ur Rehman Khan, Franz Josef Behr, Phillip Davis
Abstract:
Open source geospatial certifications are needed in geospatial technology education and industry sector. In parallel with proprietary software, free and open source software solutions become important in geospatial technology research and play an important role for the growth of the geospatial industry. ESRI, GISCI (GIS Certification Institute), ASPRS (American Society of Photogrammetry and remote sensing), and Meta spatial are offering certifications on proprietary and open source software. These are portfolio and competency based certifications depending on GIS Body of Knowledge (Bok). The analysis of these certification approaches might lead to the discovery of some gaps in them and will open a new way to develop certifications related to the geospatial open source (OS). This new certification will investigate the different geospatial competencies according to open source tools that help to identify geospatial professionals and strengthen the geospatial academic content. The goal of this research is to introduce a geospatial certification model based on geospatial technology competency model (GTCM).The developed certification will not only incorporate the importance of geospatial education and production of the geospatial competency-based workforce in universities and companies (private or public) as well as describe open source solutions with tools and technology. Job analysis, market analysis, survey analysis of this certification opens a new horizon for business as well.Keywords: geospatial certification, open source, geospatial technology competency model, geoscience
Procedia PDF Downloads 56626942 Improvement of Camera Calibration Based on the Relationship between Focal Length and Aberration Coefficient
Authors: Guorong Sui, Xingwei Jia, Chenhui Yin, Xiumin Gao
Abstract:
In the processing of camera-based high precision and non-contact measurement, the geometric-optical aberration is always inevitably disturbing the measuring system. Moreover, the aberration is different with the different focal length, which will increase the difficulties of the system’s calibration. Therefore, to understand the relationship between the focal length as a function of aberration properties is a very important issue to the calibration of the measuring systems. In this study, we propose a new mathematics model, which is based on the plane calibration method by Zhang Zhengyou, and establish a relationship between the focal length and aberration coefficient. By using the mathematics model and carefully modified compensation templates, the calibration precision of the system can be dramatically improved. The experiment results show that the relative error is less than 1%. It is important for optoelectronic imaging systems that apply to measure, track and position by changing the camera’s focal length.Keywords: camera calibration, aberration coefficient, vision measurement, focal length, mathematics model
Procedia PDF Downloads 36426941 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization
Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman
Abstract:
In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization
Procedia PDF Downloads 24026940 Architecture for QoS Based Service Selection Using Local Approach
Authors: Gopinath Ganapathy, Chellammal Surianarayanan
Abstract:
Services are growing rapidly and generally they are aggregated into a composite service to accomplish complex business processes. There may be several services that offer the same required function of a particular task in a composite service. Hence a choice has to be made for selecting suitable services from alternative functionally similar services. Quality of Service (QoS)plays as a discriminating factor in selecting which component services should be selected to satisfy the quality requirements of a user during service composition. There are two categories of approaches for QoS based service selection, namely global and local approaches. Global approaches are known to be Non-Polynomial (NP) hard in time and offer poor scalability in large scale composition. As an alternative to global methods, local selection methods which reduce the search space by breaking up the large/complex problem of selecting services for the workflow into independent sub problems of selecting services for individual tasks are coming up. In this paper, distributed architecture for selecting services based on QoS using local selection is presented with an overview of local selection methodology. The architecture describes the core components, namely, selection manager and QoS manager needed to implement the local approach and their functions. Selection manager consists of two components namely constraint decomposer which decomposes the given global or workflow level constraints in local or task level constraints and service selector which selects appropriate service for each task with maximum utility, satisfying the corresponding local constraints. QoS manager manages the QoS information at two levels namely, service class level and individual service level. The architecture serves as an implementation model for local selection.Keywords: architecture of service selection, local method for service selection, QoS based service selection, approaches for QoS based service selection
Procedia PDF Downloads 42626939 Assessing the Leadership Succession Plan in Faith-Based Senior High Schools in Ghana and Its Associated Challenges
Authors: J. E. Cobbinah
Abstract:
One of the most challenging issues confronting schools is good leadership succession planning. Experts argue that, although the idea of leadership succession planning is one of the strategies or practices that can help sustain improvement and promote continuity of good leadership, seem to have been neglected in many schools over the years. Appointment of head teachers in senior high schools is based on long service or one’s ability to demonstrate his/her competence in a leadership selection interview. There is no clear and well-structured leadership succession plan, before leadership position is filled, while school leadership succession planning seem to be an issue that nobody talks about. In faith-based schools the issue is even worse, because religious groups impose whoever they consider strong in the faith on schools as leaders, irrespective of the individual competence, ability to take up challenges associated with individuals’ preparedness to take up leadership position. Therefore, the present study examined the nature (including type) of leadership succession plans in faith-based senior high schools and its associated challenges. Convergent mixed method design was employed to effectively achieve the objectives of the study. The data collection strategies involved the use of interviews, questionnaires, and reviews of secondary data. The data was gathered from students, school leaders (head teachers, deputy heads, and head of departments), selected parents teachers associated members, school management committee members and members from school governors. The results show that governors of faith-based schools are making efforts to enhance education quality, by making school leadership accountable, the absence and the neglect of clear, and well-structured leadership succession plan has some negative outcomes. Unsustainable students’ academic performance, lack of support from existing staffs and senior leaders and lack of support in the implementation of school improvement plan. It would be concluded that, faith-based schools should focus on leadership competence and abilities in the selection process of potential school leaders to achieve a good succession plan rather than appointing leaders who are affiliates of one’s faith.Keywords: school leadership, succession planning, faith-based schools, school governors
Procedia PDF Downloads 44526938 Research on Air pollution Spatiotemporal Forecast Model Based on LSTM
Authors: JingWei Yu, Hong Yang Yu
Abstract:
At present, the increasingly serious air pollution in various cities of China has made people pay more attention to the air quality index(hereinafter referred to as AQI) of their living areas. To face this situation, it is of great significance to predict air pollution in heavily polluted areas. In this paper, based on the time series model of LSTM, a spatiotemporal prediction model of PM2.5 concentration in Mianyang, Sichuan Province, is established. The model fully considers the temporal variability and spatial distribution characteristics of PM2.5 concentration. The spatial correlation of air quality at different locations is based on the Air quality status of other nearby monitoring stations, including AQI and meteorological data to predict the air quality of a monitoring station. The experimental results show that the method has good prediction accuracy that the fitting degree with the actual measured data reaches more than 0.7, which can be applied to the modeling and prediction of the spatial and temporal distribution of regional PM2.5 concentration.Keywords: LSTM, PM2.5, neural networks, spatio-temporal prediction
Procedia PDF Downloads 13426937 Distribution and Characterization of Thermal Springs in Northern Oman
Authors: Fahad Al Shidi, Reginald Victor
Abstract:
This study was conducted in Northern Oman to assess the physical and chemical characteristics of 40 thermal springs distributed in Al Hajar Mountains in northern Oman. Physical measurements of water samples were carried out in two main seasons in Oman (winter and summer 2019). Studied springs were classified into three groups based on water temperature, four groups based on water pH values and two groups based on conductivity. Ten thermal alkaline springs that originated in Ophiolite (Samail Napp) were dominated by high pH (> 11), elevated concentration of Cl- and Na+ ions, relatively low temperature and discharge ratio. Other springs in the Hajar Super Group massif recorded high concentrations of Ca2+ and SO2-4 ions controlled by rock dominance, geochemistry processes, and mineralization. There was only one spring which has brackish water with very high conductivity (5500 µs/cm) and Total Dissolved Solids and it is not suitable for irrigation purposes because of the high abundance of Na+, Cl−, and Ca2+ ions.Keywords: alkaline springs, geothermal, HSG, ophiolite
Procedia PDF Downloads 14326936 Developing a Knowledge-Based Lean Six Sigma Model to Improve Healthcare Leadership Performance
Authors: Yousuf N. Al Khamisi, Eduardo M. Hernandez, Khurshid M. Khan
Abstract:
Purpose: This paper presents a model of a Knowledge-Based (KB) using Lean Six Sigma (L6σ) principles to enhance the performance of healthcare leadership. Design/methodology/approach: Using L6σ principles to enhance healthcare leaders’ performance needs a pre-assessment of the healthcare organisation’s capabilities. The model will be developed using a rule-based approach of KB system. Thus, KB system embeds Gauging Absence of Pre-requisite (GAP) for benchmarking and Analytical Hierarchy Process (AHP) for prioritization. A comprehensive literature review will be covered for the main contents of the model with a typical output of GAP analysis and AHP. Findings: The proposed KB system benchmarks the current position of healthcare leadership with the ideal benchmark one (resulting from extensive evaluation by the KB/GAP/AHP system of international leadership concepts in healthcare environments). Research limitations/implications: Future work includes validating the implementation model in healthcare environments around the world. Originality/value: This paper presents a novel application of a hybrid KB combines of GAP and AHP methodology. It implements L6σ principles to enhance healthcare performance. This approach assists healthcare leaders’ decision making to reach performance improvement against a best practice benchmark.Keywords: Lean Six Sigma (L6σ), Knowledge-Based System (KBS), healthcare leadership, Gauge Absence Prerequisites (GAP), Analytical Hierarchy Process (AHP)
Procedia PDF Downloads 166