Search results for: error estimates
2427 Heteroscedastic Parametric and Semiparametric Smooth Coefficient Stochastic Frontier Application to Technical Efficiency Measurement
Authors: Rebecca Owusu Coffie, Atakelty Hailu
Abstract:
Variants of production frontier models have emerged, however, only a limited number of them are applied in empirical research. Hence the effects of these alternative frontier models are not well understood, particularly within sub-Saharan Africa. In this paper, we apply recent advances in the production frontier to examine levels of technical efficiency and efficiency drivers. Specifically, we compare the heteroscedastic parametric and the semiparametric stochastic smooth coefficient (SPSC) models. Using rice production data from Ghana, our empirical estimates reveal that alternative specification of efficiency estimators results in either downward or upward bias in the technical efficiency estimates. Methodologically, we find that the SPSC model is more suitable and generates high-efficiency estimates. Within the parametric framework, we find that parameterization of both the mean and variance of the pre-truncated function is the best model. For the drivers of technical efficiency, we observed that longer farm distances increase inefficiency through a reduction in labor productivity. High soil quality, however, increases productivity through increased land productivity.Keywords: pre-truncated, rice production, smooth coefficient, technical efficiency
Procedia PDF Downloads 4462426 Influence of Error Correction Codes on the Quality of Optical Broadband Connections
Authors: Mouna Hemdi, Jamel bel Hadj Tahar
Abstract:
The increasing development of multimedia applications requiring the simultaneous transport of several different services contributes to the evolution of the need for very high-speed network. In this paper, we propose an effective solution to achieve the very high speed while retaining elements of the optical transmission channel. So our study focuses on error correcting codes that aim for quality improvement on duty. We present a comparison of the quality of service for single channels and integrating the code BCH, RS and LDPC in order to find the best code in the different conditions of the transmission.Keywords: code error correction, high speed broadband, optical transmission, information systems security
Procedia PDF Downloads 3932425 A Short-Baseline Dual-Antenna BDS/MEMS-IMU Integrated Navigation System
Authors: Tijing Cai, Qimeng Xu, Daijin Zhou
Abstract:
This paper puts forward a short-baseline dual-antenna BDS/MEMS-IMU integrated navigation, constructs the carrier phase double difference model of BDS (BeiDou Navigation Satellite System), and presents a 2-position initial orientation method on BDS. The Extended Kalman-filter has been introduced for the integrated navigation system. The differences between MEMS-IMU and BDS position, velocity and carrier phase indications are used as measurements. To show the performance of the short-baseline dual-antenna BDS/MEMS-IMU integrated navigation system, the experiment results show that the position error is less than 1m, the pitch angle error and roll angle error are less than 0.1°, and the heading angle error is about 1°.Keywords: MEMS-IMU (Micro-Electro-Mechanical System Inertial Measurement Unit), BDS (BeiDou Navigation Satellite System), dual-antenna, integrated navigation
Procedia PDF Downloads 1932424 Weighted G2 Multi-Degree Reduction of Bezier Curves
Authors: Salisu ibrahim, Abdalla Rababah
Abstract:
In this research, we use Weighted G2-Multi-degree reduction of Bezier curve of degree n to a Bezier curve of degree m, m < n. The degree reduction of Bezier curves is used to represent a given Bezier curve of n by a Bezier curve of degree m, m < n. Exact degree reduction is not possible, and degree reduction is approximate process in nature. We derive a weighted degree reducing method that is geometrically continuous at the end points. Different norms will be considered, several error minimizations will be given. The proposed methods produce error function that are less than the errors of existing methods.Keywords: Bezier curves, multiple degree reduction, geometric continuity, error function
Procedia PDF Downloads 4822423 Does sustainability disclosure improve analysts’ forecast accuracy Evidence from European banks
Authors: Albert Acheampong, Tamer Elshandidy
Abstract:
We investigate the extent to which sustainability disclosure from the narrative section of European banks’ annual reports improves analyst forecast accuracy. We capture sustainability disclosure using a machine learning approach and use forecast error to proxy analyst forecast accuracy. Our results suggest that sustainability disclosure significantly improves analyst forecast accuracy by reducing the forecast error. In a further analysis, we also find that the induction of Directive 2014/95/European Union (EU) is associated with increased disclosure content, which then reduces forecast error. Collectively, our results suggest that sustainability disclosure improves forecast accuracy, and the induction of the new EU directive strengthens this improvement. These results hold after several further and robustness analyses. Our findings have implications for market participants and policymakers.Keywords: sustainability disclosure, machine learning, analyst forecast accuracy, forecast error, European banks, EU directive
Procedia PDF Downloads 772422 Estimating Cyclone Intensity Using INSAT-3D IR Images Based on Convolution Neural Network Model
Authors: Divvela Vishnu Sai Kumar, Deepak Arora, Sheenu Rizvi
Abstract:
Forecasting a cyclone through satellite images consists of the estimation of the intensity of the cyclone and predicting it before a cyclone comes. This research work can help people to take safety measures before the cyclone comes. The prediction of the intensity of a cyclone is very important to save lives and minimize the damage caused by cyclones. These cyclones are very costliest natural disasters that cause a lot of damage globally due to a lot of hazards. Authors have proposed five different CNN (Convolutional Neural Network) models that estimate the intensity of cyclones through INSAT-3D IR images. There are a lot of techniques that are used to estimate the intensity; the best model proposed by authors estimates intensity with a root mean squared error (RMSE) of 10.02 kts.Keywords: estimating cyclone intensity, deep learning, convolution neural network, prediction models
Procedia PDF Downloads 1292421 Support Vector Regression with Weighted Least Absolute Deviations
Authors: Kang-Mo Jung
Abstract:
Least squares support vector machine (LS-SVM) is a penalized regression which considers both fitting and generalization ability of a model. However, the squared loss function is very sensitive to even single outlier. We proposed a weighted absolute deviation loss function for the robustness of the estimates in least absolute deviation support vector machine. The proposed estimates can be obtained by a quadratic programming algorithm. Numerical experiments on simulated datasets show that the proposed algorithm is competitive in view of robustness to outliers.Keywords: least absolute deviation, quadratic programming, robustness, support vector machine, weight
Procedia PDF Downloads 5272420 A Variant of Newton's Method with Free Second-Order Derivative
Authors: Young Hee Geum
Abstract:
In this paper, we present the iterative method and determine the control parameters to converge cubically for solving nonlinear equations. In addition, we derive the asymptotic error constant.Keywords: asymptotic error constant, iterative method, multiple root, root-finding, order of convergent
Procedia PDF Downloads 2942419 Robust Variogram Fitting Using Non-Linear Rank-Based Estimators
Authors: Hazem M. Al-Mofleh, John E. Daniels, Joseph W. McKean
Abstract:
In this paper numerous robust fitting procedures are considered in estimating spatial variograms. In spatial statistics, the conventional variogram fitting procedure (non-linear weighted least squares) suffers from the same outlier problem that has plagued this method from its inception. Even a 3-parameter model, like the variogram, can be adversely affected by a single outlier. This paper uses the Hogg-Type adaptive procedures to select an optimal score function for a rank-based estimator for these non-linear models. Numeric examples and simulation studies will demonstrate the robustness, utility, efficiency, and validity of these estimates.Keywords: asymptotic relative efficiency, non-linear rank-based, rank estimates, variogram
Procedia PDF Downloads 4312418 Cuckoo Search Optimization for Black Scholes Option Pricing
Authors: Manas Shah
Abstract:
Black Scholes option pricing model is one of the most important concepts in modern world of computational finance. However, its practical use can be challenging as one of the input parameters must be estimated; implied volatility of the underlying security. The more precisely these values are estimated, the more accurate their corresponding estimates of theoretical option prices would be. Here, we present a novel model based on Cuckoo Search Optimization (CS) which finds more precise estimates of implied volatility than Particle Swarm Optimization (PSO) and Genetic Algorithm (GA).Keywords: black scholes model, cuckoo search optimization, particle swarm optimization, genetic algorithm
Procedia PDF Downloads 4532417 Channel Estimation/Equalization with Adaptive Modulation and Coding over Multipath Faded Channels for WiMAX
Authors: B. Siva Kumar Reddy, B. Lakshmi
Abstract:
WiMAX has adopted an Adaptive Modulation and Coding (AMC) in OFDM to endure higher data rates and error free transmission. AMC schemes employ the Channel State Information (CSI) to efficiently utilize the channel and maximize the throughput and for better spectral efficiency. This CSI has given to the transmitter by the channel estimators. In this paper, LSE (Least Square Error) and MMSE (Minimum Mean square Error) estimators are suggested and BER (Bit Error Rate) performance has been analyzed. Channel equalization is also integrated with with AMC-OFDM system and presented with Constant Modulus Algorithm (CMA) and Least Mean Square (LMS) algorithms with convergence rates analysis. Simulation results proved that increment in modulation scheme size causes to improvement in throughput along with BER value. There is a trade-off among modulation size, throughput, BER value and spectral efficiency. Results also reported the requirement of channel estimation and equalization in high data rate systems.Keywords: AMC, CSI, CMA, OFDM, OFDMA, WiMAX
Procedia PDF Downloads 3932416 Developing a Machine Learning-based Cost Prediction Model for Construction Projects using Particle Swarm Optimization
Authors: Soheila Sadeghi
Abstract:
Accurate cost prediction is essential for effective project management and decision-making in the construction industry. This study aims to develop a cost prediction model for construction projects using Machine Learning techniques and Particle Swarm Optimization (PSO). The research utilizes a comprehensive dataset containing project cost estimates, actual costs, resource details, and project performance metrics from a road reconstruction project. The methodology involves data preprocessing, feature selection, and the development of an Artificial Neural Network (ANN) model optimized using PSO. The study investigates the impact of various input features, including cost estimates, resource allocation, and project progress, on the accuracy of cost predictions. The performance of the optimized ANN model is evaluated using metrics such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and R-squared. The results demonstrate the effectiveness of the proposed approach in predicting project costs, outperforming traditional benchmark models. The feature selection process identifies the most influential variables contributing to cost variations, providing valuable insights for project managers. However, this study has several limitations. Firstly, the model's performance may be influenced by the quality and quantity of the dataset used. A larger and more diverse dataset covering different types of construction projects would enhance the model's generalizability. Secondly, the study focuses on a specific optimization technique (PSO) and a single Machine Learning algorithm (ANN). Exploring other optimization methods and comparing the performance of various ML algorithms could provide a more comprehensive understanding of the cost prediction problem. Future research should focus on several key areas. Firstly, expanding the dataset to include a wider range of construction projects, such as residential buildings, commercial complexes, and infrastructure projects, would improve the model's applicability. Secondly, investigating the integration of additional data sources, such as economic indicators, weather data, and supplier information, could enhance the predictive power of the model. Thirdly, exploring the potential of ensemble learning techniques, which combine multiple ML algorithms, may further improve cost prediction accuracy. Additionally, developing user-friendly interfaces and tools to facilitate the adoption of the proposed cost prediction model in real-world construction projects would be a valuable contribution to the industry. The findings of this study have significant implications for construction project management, enabling proactive cost estimation, resource allocation, budget planning, and risk assessment, ultimately leading to improved project performance and cost control. This research contributes to the advancement of cost prediction techniques in the construction industry and highlights the potential of Machine Learning and PSO in addressing this critical challenge. However, further research is needed to address the limitations and explore the identified future research directions to fully realize the potential of ML-based cost prediction models in the construction domain.Keywords: cost prediction, construction projects, machine learning, artificial neural networks, particle swarm optimization, project management, feature selection, road reconstruction
Procedia PDF Downloads 602415 A Sectional Control Method to Decrease the Accumulated Survey Error of Tunnel Installation Control Network
Authors: Yinggang Guo, Zongchun Li
Abstract:
In order to decrease the accumulated survey error of tunnel installation control network of particle accelerator, a sectional control method is proposed. Firstly, the accumulation rule of positional error with the length of the control network is obtained by simulation calculation according to the shape of the tunnel installation-control-network. Then, the RMS of horizontal positional precision of tunnel backbone control network is taken as the threshold. When the accumulated error is bigger than the threshold, the tunnel installation control network should be divided into subsections reasonably. On each segment, the middle survey station is taken as the datum for independent adjustment calculation. Finally, by taking the backbone control points as faint datums, the weighted partial parameters adjustment is performed with the adjustment results of each segment and the coordinates of backbone control points. The subsections are jointed and unified into the global coordinate system in the adjustment process. An installation control network of the linac with a length of 1.6 km is simulated. The RMS of positional deviation of the proposed method is 2.583 mm, and the RMS of the difference of positional deviation between adjacent points reaches 0.035 mm. Experimental results show that the proposed sectional control method can not only effectively decrease the accumulated survey error but also guarantee the relative positional precision of the installation control network. So it can be applied in the data processing of tunnel installation control networks, especially for large particle accelerators.Keywords: alignment, tunnel installation control network, accumulated survey error, sectional control method, datum
Procedia PDF Downloads 1912414 Maximum Power Point Tracking Based on Estimated Power for PV Energy Conversion System
Authors: Zainab Almukhtar, Adel Merabet
Abstract:
In this paper, a method for maximum power point tracking of a photovoltaic energy conversion system is presented. This method is based on using the difference between the power from the solar panel and an estimated power value to control the DC-DC converter of the photovoltaic system. The difference is continuously compared with a preset error permitted value. If the power difference is more than the error, the estimated power is multiplied by a factor and the operation is repeated until the difference is less or equal to the threshold error. The difference in power will be used to trigger a DC-DC boost converter in order to raise the voltage to where the maximum power point is achieved. The proposed method was experimentally verified through a PV energy conversion system driven by the OPAL-RT real time controller. The method was tested on varying radiation conditions and load requirements, and the Photovoltaic Panel was operated at its maximum power in different conditions of irradiation.Keywords: control system, error, solar panel, MPPT tracking
Procedia PDF Downloads 2842413 Role of Vision Centers in Eliminating Avoidable Blindness Caused Due to Uncorrected Refractive Error in Rural South India
Authors: Ranitha Guna Selvi D, Ramakrishnan R, Mohideen Abdul Kader
Abstract:
Purpose: To study the role of Vision centers in managing preventable blindness through refractive error correction in Rural South India. Methods: A retrospective analysis of patients attending 15 Vision centers in Rural South India from a period of January 2021 to December 2021 was done. Medical records of 10,85,81 patients both new and reviewed, 79,562 newly registered patients and 29,019 review patient’s from15 Vision centers were included for data analysis. All the patients registered at the vision center underwent basic eye examination, including visual acuity, IOP measurement, Slit-lamp examination, retinoscopy, Fundus examination etc. Results: A total of 1,08,581 patients were included in the study. Of the total 1,08,581 patients, 79,562 were newly registered patients at Vision center and 29,019 were review patients. Males were 52,201(48.1%) and Females were 56,308(51.9) among them. The mean age of all examined patients was 41.03 ± 20.9 years (Standard deviation) and ranged from 01 – 113 years. Presenting mean visual acuity was 0.31 ± 0.5 in the right eye and 0.31 ± 0.4 in the left eye. Of the 1,08,581 patients 22,770 patients had refractive error in right eye and 22,721 patients had uncorrected refractive error in left eye. Glass prescription was given to 17,178 (15.8%) patients. 8,109 (7.5%) patients were referred to the base hospital for specialty clinic expert opinion or for cataract surgery. Conclusion: Vision center utilizing teleconsultation for comprehensive eye screening unit is a very effective tool in reducing the avoidable visual impairment caused due to uncorrected refractive error. Vision Centre model is believed to be efficient as it facilitates early detection and management of uncorrected refractive errors.Keywords: refractive error, uncorrected refractive error, vision center, vision technician, teleconsultation
Procedia PDF Downloads 1422412 Malpractice Makes Perfect: A Thematic Analysis on How Doctors Handle Medical Errors
Authors: Kathleen Joy Hingan, Jessiraye Luienne Catubigan, Carlo Mercado, Janisse RañEses
Abstract:
In this research, the researchers wanted to explore how specialists and resident doctors in the fields of surgery, and obstetrics and gynecology handle their medical errors. They are interested in understanding the factors that contributed to the disclosure of medical error, the feelings after the occurrence of an error, and the way they coped with it given the power relations in place. The researchers conducted semi-structured interviews, transcribed the recordings, and analyzed the transcripts using thematic analysis. They found that doctors disclosed to their superiors and co-residents to cope with and to learn from the errors. In terms of disclosure to patients, the participants told them about the adverse event, but not about the error because of fear for themselves, their colleagues, their institution, and their patient. Doctors also performed compensatory actions to make up for the error and the nondisclosure of its occurrence. These actions functioned as a form of damage control too. Resident doctors and specialists receive different sanctions because of the power structures in the system.Keywords: coping, disclosure, doctors, interviews, medical errors, thematic analysis
Procedia PDF Downloads 2902411 Genetic-Environment Influences on the Cognitive Abilities of 6-to-8 Years Old Twins
Authors: Annu Panghal, Bimla Dhanda
Abstract:
This research paper aims to determine the genetic-environment influences on the cognitive abilities of twins. Using the 100 pairs of twins from two districts, namely: Bhiwani (N = 90) and Hisar (N = 110) of Haryana State, genetic and environmental influences were assessed in twin study design. The cognitive abilities of twins were measured using the Wechsler Intelligence Scale for Children (WISC-R). Home Observation for Measurement of the Environment (HOME) Inventory was taken to examine the home environment of twins. Heritability estimate was used to analyze the genes contributing to shape the cognitive abilities of twins. The heritability estimates for cognitive abilities of 6-7 years old twins in Hisar district were 74% and in Bhiwani District 76%. Further the heritability estimates were 64% in the twins of Hisar district and 60 in Bhiwani district % in the age group of 7-8 years. The remaining variations in the cognitive abilities of twins were due to environmental factors namely: provision for Active Stimulation, paternal involvement, safe physical environment. The findings provide robust evidence that the cognitive abilities were more influenced by genes than the environmental factors and also revealed that the influence of genetic was more in the age group 6-7 years than the age group 7-8 years. The conclusion of the heritability estimates indicates that the genetic influence was more in the age group of 6-7 years than the age group of 7-8 years. As the age increases the genetic influence decreases and environment influence increases. Mother education was strongly associated with the cognitive abilities of twins.Keywords: genetics, heritability, twins, environment, cognitive abilities
Procedia PDF Downloads 1392410 Taylor’s Law and Relationship between Life Expectancy at Birth and Variance in Age at Death in Period Life Table
Authors: David A. Swanson, Lucky M. Tedrow
Abstract:
Taylor’s Law is a widely observed empirical pattern that relates variances to means in sets of non-negative measurements via an approximate power function, which has found application to human mortality. This study adds to this research by showing that Taylor’s Law leads to a model that reasonably describes the relationship between life expectancy at birth (e0, which also is equal to mean age at death in a life table) and variance at age of death in seven World Bank regional life tables measured at two points in time, 1970 and 2000. Using as a benchmark a non-random sample of four Japanese female life tables covering the period from 1950 to 2004, the study finds that the simple linear model provides reasonably accurate estimates of variance in age at death in a life table from e0, where the latter range from 60.9 to 85.59 years. Employing 2017 life tables from the Human Mortality Database, the simple linear model is used to provide estimates of variance at age in death for six countries, three of which have high e0 values and three of which have lower e0 values. The paper provides a substantive interpretation of Taylor’s Law relative to e0 and concludes by arguing that reasonably accurate estimates of variance in age at death in a period life table can be calculated using this approach, which also can be used where e0 itself is estimated rather than generated through the construction of a life table, a useful feature of the model.Keywords: empirical pattern, mean age at death in a life table, mean age of a stationary population, stationary population
Procedia PDF Downloads 3302409 Macroeconomic Effects and Dynamics of Natural Disaster Damages: Evidence from SETX on the Resiliency Hypothesis
Authors: Agim Kukelii, Gevorg Sargsyan
Abstract:
This study, focusing on the base regional area (county level), estimates the effect of natural disaster damages on aggregate personal income, aggregate wages, wages per worker, aggregate employment, and aggregate income transfer. The study further estimates the dynamics of personal income, employment, and wages under natural disaster shocks. Southeast Texas, located at the center of Golf Coast, is hit by meteorological and hydrological caused natural disasters yearly. On average, there are more than four natural disasters per year that cane an estimated damage average of 2.2% of real personal income. The study uses the panel data method to estimate the average effect of natural disasters on the area’s economy (personal income, wages, employment, and income transfer). It also uses Panel Vector Autoregressive (PVAR) model to study the dynamics of macroeconomic variables under natural disaster shocks. The study finds that the average effect of natural disasters is positive for personal income and income transfer and is negative for wages and employment. The PVAR and the impulse response function estimates reveal that natural disaster shocks cause a decrease in personal income, employment, and wages. However, the economy’s variables bounce back after three years. The novelty of this study rests on several aspects. First, this is the first study to investigate the effects of natural disasters on macroeconomic variables at a regional level. Second, the study uses direct measures of natural disaster damages. Third, the study estimates that the time that the local economy takes to absorb the natural disaster damages shocks is three years. This is a relatively good reaction to the local economy, therefore, adding to the “resiliency” hypothesis. The study has several implications for policymakers, businesses, and households. First, this study serves to increase the awareness of local stakeholders that natural disaster damages do worsen, macroeconomic variables, such as personal income, employment, and wages beyond the immediate damages to residential and commercial properties, physical infrastructure, and discomfort in daily lives. Second, the study estimates that these effects linger on the economy on average for three years, which would require policymakers to factor in the time area need to be on focus.Keywords: natural disaster damages, macroeconomics effects, PVAR, panel data
Procedia PDF Downloads 882408 Sensor Registration in Multi-Static Sonar Fusion Detection
Authors: Longxiang Guo, Haoyan Hao, Xueli Sheng, Hanjun Yu, Jingwei Yin
Abstract:
In order to prevent target splitting and ensure the accuracy of fusion, system error registration is an important step in multi-static sonar fusion detection system. To eliminate the inherent system errors including distance error and angle error of each sonar in detection, this paper uses offline estimation method for error registration. Suppose several sonars from different platforms work together to detect a target. The target position detected by each sonar is based on each sonar’s own reference coordinate system. Based on the two-dimensional stereo projection method, this paper uses real-time quality control (RTQC) method and least squares (LS) method to estimate sensor biases. The RTQC method takes the average value of each sonar’s data as the observation value and the LS method makes the least square processing of each sonar’s data to get the observation value. In the underwater acoustic environment, matlab simulation is carried out and the simulation results show that both algorithms can estimate the distance and angle error of sonar system. The performance of the two algorithms is also compared through the root mean square error and the influence of measurement noise on registration accuracy is explored by simulation. The system error convergence of RTQC method is rapid, but the distribution of targets has a serious impact on its performance. LS method can not be affected by target distribution, but the increase of random noise will slow down the convergence rate. LS method is an improvement of RTQC method, which is widely used in two-dimensional registration. The improved method can be used for underwater multi-target detection registration.Keywords: data fusion, multi-static sonar detection, offline estimation, sensor registration problem
Procedia PDF Downloads 1692407 Monocular Visual Odometry for Three Different View Angles by Intel Realsense T265 with the Measurement of Remote
Authors: Heru Syah Putra, Aji Tri Pamungkas Nurcahyo, Chuang-Jan Chang
Abstract:
MOIL-SDK method refers to the spatial angle that forms a view with a different perspective from the Fisheye image. Visual Odometry forms a trusted application for extending projects by tracking using image sequences. A real-time, precise, and persistent approach that is able to contribute to the work when taking datasets and generate ground truth as a reference for the estimates of each image using the FAST Algorithm method in finding Keypoints that are evaluated during the tracking process with the 5-point Algorithm with RANSAC, as well as produce accurate estimates the camera trajectory for each rotational, translational movement on the X, Y, and Z axes.Keywords: MOIL-SDK, intel realsense T265, Fisheye image, monocular visual odometry
Procedia PDF Downloads 1352406 Estimation of Heritability and Repeatability for Pre-Weaning Body Weights of Domestic Rabbits Raised in Derived Savanna Zone of Nigeria
Authors: Adewale I. Adeolu, Vivian U. Oleforuh-Okoleh, Sylvester N. Ibe
Abstract:
Heritability and repeatability estimates are needed for the genetic evaluation of livestock populations and consequently for the purpose of upgrading or improvement. Pooled data on 604 progeny from three consecutive parities of purebred rabbit breeds (Chinchilla, Dutch and New Zealand white) raised in Derived Savanna Zone of Nigeria were used to estimate heritability and repeatability for pre-weaning body weights between 1st and 8th week of age. Traits studied include Individual kit weight at birth (IKWB), 2nd week (IK2W), 4th week (IK4W), 6th week (IK6W) and 8th week (IK8W). Nested random effects analysis of (Co)variances as described by Statistical Analysis System (SAS) were employed in the estimation. Respective heritability estimates from the sire component (h2s) and repeatability (R) as intra-class correlations of repeated measurements from the three parties for IKWB, IK2W, IK4W and IK8W are 0.59±0.24, 0.55±0.24, 0.93±0.31, 0.28±0.17, 0.64±0.26 and 0.12±0.14, 0.05±0.14, 0.58±0.02, 0.60±0.11, 0.20±0.14. Heritability and repeatability (except R for IKWB and IK2W) estimates are moderate to high. In conclusion, since pre-weaning body weights in the present study tended to be moderately to highly heritable and repeatable, improvement of rabbits raised in derived savanna zone can be realized through genetic selection criterions.Keywords: heritability, nested design, parity, pooled data, repeatability
Procedia PDF Downloads 1472405 Groundwater Flow Assessment Based on Numerical Simulation at Omdurman Area, Khartoum State, Sudan
Authors: Adil Balla Elkrail
Abstract:
Visual MODFLOW computer codes were selected to simulate head distribution, calculate the groundwater budgets of the area, and evaluate the effect of external stresses on the groundwater head and to demonstrate how the groundwater model can be used as a comparative technique in order to optimize utilization of the groundwater resource. A conceptual model of the study area, aquifer parameters, boundary, and initial conditions were used to simulate the flow model. The trial-and-error technique was used to calibrate the model. The most important criteria used to check the calibrated model were Root Mean Square error (RMS), Mean Absolute error (AM), Normalized Root Mean Square error (NRMS) and mass balance. The maps of the simulated heads elaborated acceptable model calibration compared to observed heads map. A time length of eight years and the observed heads of the year 2004 were used for model prediction. The predictive simulation showed that the continuation of pumping will cause relatively high changes in head distribution and components of groundwater budget whereas, the low deficit computed (7122 m3/d) between inflows and outflows cannot create a significant drawdown of the potentiometric level. Hence, the area under consideration may represent a high permeability and productive zone and strongly recommended for further groundwater development.Keywords: aquifers, model simulation, groundwater, calibrations, trail-and- error, prediction
Procedia PDF Downloads 2422404 A Methodology of Using Fuzzy Logics and Data Analytics to Estimate the Life Cycle Indicators of Solar Photovoltaics
Authors: Thor Alexis Sazon, Alexander Guzman-Urbina, Yasuhiro Fukushima
Abstract:
This study outlines the method of how to develop a surrogate life cycle model based on fuzzy logic using three fuzzy inference methods: (1) the conventional Fuzzy Inference System (FIS), (2) the hybrid system of Data Analytics and Fuzzy Inference (DAFIS), which uses data clustering for defining the membership functions, and (3) the Adaptive-Neuro Fuzzy Inference System (ANFIS), a combination of fuzzy inference and artificial neural network. These methods were demonstrated with a case study where the Global Warming Potential (GWP) and the Levelized Cost of Energy (LCOE) of solar photovoltaic (PV) were estimated using Solar Irradiation, Module Efficiency, and Performance Ratio as inputs. The effects of using different fuzzy inference types, either Sugeno- or Mamdani-type, and of changing the number of input membership functions to the error between the calibration data and the model-generated outputs were also illustrated. The solution spaces of the three methods were consequently examined with a sensitivity analysis. ANFIS exhibited the lowest error while DAFIS gave slightly lower errors compared to FIS. Increasing the number of input membership functions helped with error reduction in some cases but, at times, resulted in the opposite. Sugeno-type models gave errors that are slightly lower than those of the Mamdani-type. While ANFIS is superior in terms of error minimization, it could generate solutions that are questionable, i.e. the negative GWP values of the Solar PV system when the inputs were all at the upper end of their range. This shows that the applicability of the ANFIS models highly depends on the range of cases at which it was calibrated. FIS and DAFIS generated more intuitive trends in the sensitivity runs. DAFIS demonstrated an optimal design point wherein increasing the input values does not improve the GWP and LCOE anymore. In the absence of data that could be used for calibration, conventional FIS presents a knowledge-based model that could be used for prediction. In the PV case study, conventional FIS generated errors that are just slightly higher than those of DAFIS. The inherent complexity of a Life Cycle study often hinders its widespread use in the industry and policy-making sectors. While the methodology does not guarantee a more accurate result compared to those generated by the Life Cycle Methodology, it does provide a relatively simpler way of generating knowledge- and data-based estimates that could be used during the initial design of a system.Keywords: solar photovoltaic, fuzzy logic, inference system, artificial neural networks
Procedia PDF Downloads 1652403 Model Estimation and Error Level for Okike’s Merged Irregular Transposition Cipher
Authors: Okike Benjamin, Garba E. J. D.
Abstract:
The researcher has developed a new encryption technique known as Merged Irregular Transposition Cipher. In this cipher method of encryption, a message to be encrypted is split into parts and each part encrypted separately. Before the encrypted message is transmitted to the recipient(s), the positions of the split in the encrypted messages could be swapped to ensure more security. This work seeks to develop a model by considering the split number, S and the average number of characters per split, L as the message under consideration is split from 2 through 10. Again, after developing the model, the error level in the model would be determined.Keywords: merged irregular transposition, error level, model estimation, message splitting
Procedia PDF Downloads 3142402 Combined Automatic Speech Recognition and Machine Translation in Business Correspondence Domain for English-Croatian
Authors: Sanja Seljan, Ivan Dunđer
Abstract:
The paper presents combined automatic speech recognition (ASR) for English and machine translation (MT) for English and Croatian in the domain of business correspondence. The first part presents results of training the ASR commercial system on two English data sets, enriched by error analysis. The second part presents results of machine translation performed by online tool Google Translate for English and Croatian and Croatian-English language pairs. Human evaluation in terms of usability is conducted and internal consistency calculated by Cronbach's alpha coefficient, enriched by error analysis. Automatic evaluation is performed by WER (Word Error Rate) and PER (Position-independent word Error Rate) metrics, followed by investigation of Pearson’s correlation with human evaluation.Keywords: automatic machine translation, integrated language technologies, quality evaluation, speech recognition
Procedia PDF Downloads 4842401 Adjustment and Compensation Techniques for the Rotary Axes of Five-axis CNC Machine Tools
Authors: Tung-Hui Hsu, Wen-Yuh Jywe
Abstract:
Five-axis computer numerical control (CNC) machine tools (three linear and two rotary axes) are ideally suited to the fabrication of complex work pieces, such as dies, turbo blades, and cams. The locations of the axis average line and centerline of the rotary axes strongly influence the performance of these machines; however, techniques to compensate for eccentric error in the rotary axes remain weak. This paper proposes optical (Non-Bar) techniques capable of calibrating five-axis CNC machine tools and compensating for eccentric error in the rotary axes. This approach employs the measurement path in ISO/CD 10791-6 to determine the eccentric error in two rotary axes, for which compensatory measures can be implemented. Experimental results demonstrate that the proposed techniques can improve the performance of various five-axis CNC machine tools by more than 90%. Finally, a result of the cutting test using a B-type five-axis CNC machine tool confirmed to the usefulness of this proposed compensation technique.Keywords: calibration, compensation, rotary axis, five-axis computer numerical control (CNC) machine tools, eccentric error, optical calibration system, ISO/CD 10791-6
Procedia PDF Downloads 3832400 Formulation of a Stress Management Program for Human Error Prevention in Nuclear Power Plants
Authors: Hyeon-Kyo Lim, Tong-il Jang, Yong-Hee Lee
Abstract:
As for any nuclear power plant, human error is one of the most dreaded factors that may result in unexpected accidents. Thus, for accident prevention, it is quite indispensable to analyze and to manage the influence of any factor which may raise the possibility of human errors. Among lots factors, stress has been reported to have significant influence on human performance. Stress level of a person may fluctuate over time. To handle the possibility over time, robust stress management program is required, especially in nuclear power plants. Therefore, to overcome the possibility of human errors, this study aimed to develop a stress management program as a part of Fitness-for-Duty (FFD) Program for the workers in nuclear power plants. The meaning of FFD might be somewhat different by research objectives, appropriate definition of FFD was accomplished in this study with special reference to human error prevention, and diverse stress factors were elicited for management of human error susceptibility. In addition, with consideration of conventional FFD management programs, appropriate tests and interventions were introduced over the whole employment cycle including selection and screening of workers, job allocation, job rotation, and disemployment as well as Employee-Assistance-Program (EAP). The results showed that most tools mainly concentrated their weights on common organizational factors such as Demands, Supports, and Relationships in sequence, which were referred as major stress factors.Keywords: human error, accident prevention, work performance, stress, fatigue
Procedia PDF Downloads 3262399 Analysis of an Error Estimate for the Asymptotic Solution of the Heat Conduction Problem in a Dilated Pipe
Authors: E. Marušić-Paloka, I. Pažanin, M. Prša
Abstract:
Subject of this study is the stationary heat conduction problem through a pipe filled with incompressible viscous fluid. In previous work, we observed the existence and uniqueness theorems for the corresponding boundary-value problem and within we have taken into account the effects of the pipe's dilatation due to the temperature of the fluid inside of the pipe. The main difficulty comes from the fact that flow domain changes depending on the solution of the observed heat equation leading to a non-standard coupled governing problem. The goal of this work is to find solution estimate since the exact solution of the studied problem is not possible to determine. We use an asymptotic expansion in order of a small parameter which is presented as a heat expansion coefficient of the pipe's material. Furthermore, an error estimate is provided for the mentioned asymptotic approximation of the solution for inner area of the pipe. Close to the boundary, problem becomes more complex so different approaches are observed, mainly Theory of Perturbations and Separations of Variables. In view of that, error estimate for the whole approximation will be provided with additional software simulations of gotten situation.Keywords: asymptotic analysis, dilated pipe, error estimate, heat conduction
Procedia PDF Downloads 2362398 Investigating the Causes of Human Error-Induced Incidents in the Maintenance Operations of Petrochemical Industry by Using Human Factors Analysis and Classification System
Authors: Omid Kalatpour, Mohammadreza Ajdari
Abstract:
This article studied the possible causes of human error-induced incidents in the petrochemical industry maintenance activities by using Human Factors Analysis and Classification System (HFACS). The purpose of the study was anticipating and identifying these causes and proposing corrective and preventive actions. Maintenance department in a petrochemical company was selected for research. A checklist of human error-induced incidents was developed based on four HFACS main levels and nineteen sub-groups. Hierarchical task analysis (HTA) technique was used to identify maintenance activities and tasks. The main causes of possible incidents were identified by checklist and recorded. Corrective and preventive actions were defined depending on priority. Analyzing the worksheets of 444 activities in four levels of HFACS showed 37.6% of the causes were at the level of unsafe actions, 27.5% at the level of unsafe supervision, 20.9% at the level of preconditions for unsafe acts and 14% of the causes were at the level of organizational effects. The HFACS sub-groups showed errors (24.36%) inadequate supervision (14.89%) and violations (13.26%) with the most frequency. According to findings of this study, increasing the training effectiveness of operators and supervision improvement respectively are the most important measures in decreasing the human error-induced incidents in petrochemical industry maintenance.Keywords: human error, petrochemical industry, maintenance, HFACS
Procedia PDF Downloads 243