Search results for: bit error ratio (BER)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6205

Search results for: bit error ratio (BER)

5845 Design and Test a Robust Bearing-Only Target Motion Analysis Algorithm Based on Modified Gain Extended Kalman Filter

Authors: Mohammad Tarek Al Muallim, Ozhan Duzenli, Ceyhun Ilguy

Abstract:

Passive sonar is a method for detecting acoustic signals in the ocean. It detects the acoustic signals emanating from external sources. With passive sonar, we can determine the bearing of the target only, no information about the range of the target. Target Motion Analysis (TMA) is a process to estimate the position and speed of a target using passive sonar information. Since bearing is the only available information, the TMA technique called Bearing-only TMA. Many TMA techniques have been developed. However, until now, there is not a very effective method that could be used to always track an unknown target and extract its moving trace. In this work, a design of effective Bearing-only TMA Algorithm is done. The measured bearing angles are very noisy. Moreover, for multi-beam sonar, the measurements is quantized due to the sonar beam width. To deal with this, modified gain extended Kalman filter algorithm is used. The algorithm is fine-tuned, and many modules are added to improve the performance. A special validation gate module is used to insure stability of the algorithm. Many indicators of the performance and confidence level measurement are designed and tested. A new method to detect if the target is maneuvering is proposed. Moreover, a reactive optimal observer maneuver based on bearing measurements is proposed, which insure converging to the right solution all of the times. To test the performance of the proposed TMA algorithm a simulation is done with a MATLAB program. The simulator program tries to model a discrete scenario for an observer and a target. The simulator takes into consideration all the practical aspects of the problem such as a smooth transition in the speed, a circular turn of the ship, noisy measurements, and a quantized bearing measurement come for multi-beam sonar. The tests are done for a lot of given test scenarios. For all the tests, full tracking is achieved within 10 minutes with very little error. The range estimation error was less than 5%, speed error less than 5% and heading error less than 2 degree. For the online performance estimator, it is mostly aligned with the real performance. The range estimation confidence level gives a value equal to 90% when the range error less than 10%. The experiments show that the proposed TMA algorithm is very robust and has low estimation error. However, the converging time of the algorithm is needed to be improved.

Keywords: target motion analysis, Kalman filter, passive sonar, bearing-only tracking

Procedia PDF Downloads 379
5844 Numerical Analysis of a Reaction Diffusion System of Lambda-Omega Type

Authors: Hassan J. Al Salman, Ahmed A. Al Ghafli

Abstract:

In this study, we consider a nonlinear in time finite element approximation of a reaction diffusion system of lambda-omega type. We use a fixed-point theorem to prove existence of the approximations at each time level. Then, we derive some essential stability estimates and discuss the uniqueness of the approximations. In addition, we employ Nochetto mathematical framework to prove an optimal error bound in time for d= 1, 2 and 3 space dimensions. Finally, we present some numerical experiments to verify the obtained theoretical results.

Keywords: reaction diffusion system, finite element approximation, stability estimates, error bound

Procedia PDF Downloads 412
5843 Comparison of the Distillation Curve Obtained Experimentally with the Curve Extrapolated by a Commercial Simulator

Authors: Lívia B. Meirelles, Erika C. A. N. Chrisman, Flávia B. de Andrade, Lilian C. M. de Oliveira

Abstract:

True Boiling Point distillation (TBP) is one of the most common experimental techniques for the determination of petroleum properties. This curve provides information about the performance of petroleum in terms of its cuts. The experiment is performed in a few days. Techniques are used to determine the properties faster with a software that calculates the distillation curve when a little information about crude oil is known. In order to evaluate the accuracy of distillation curve prediction, eight points of the TBP curve and specific gravity curve (348 K and 523 K) were inserted into the HYSYS Oil Manager, and the extended curve was evaluated up to 748 K. The methods were able to predict the curve with the accuracy of 0.6%-9.2% error (Software X ASTM), 0.2%-5.1% error (Software X Spaltrohr).

Keywords: distillation curve, petroleum distillation, simulation, true boiling point curve

Procedia PDF Downloads 425
5842 The Investigation of Correlation between Body Composition and Physical Activity in University Students

Authors: Ferruh Taspinar, Gulce K. Seyyar, Gamze Kurt, Eda O. Okur, Emrah Afsar, Ismail Saracoglu, Betul Taspinar

Abstract:

Alterations of physical activity can effect body composition (especially body fat ratio); however body mass index may not sufficient to indicate these minimal differences. The aim of this study was to evaluate the relationship between body composition and physical activity in university students. In this study, 132 university students (mean age; 21.21±1.51) were included. Tanita BC-418 and International Physical Activity Questionnaire (IPAQ) were used to evaluate participants. The correlation between the parameters was analysed via Spearman correlation analysis. Significance level in statistical analyses was accepted is 0.05. The results showed that there was no correlation between body mass index and physical activity (p>0.05). There was a positive correlation between body muscle ratio and physical activity, whereas a negative correlation between body fat ratio and physical activity (p<0.05). This study showed that body fat and muscle ratio affects the level of physical activity in healthy university students. Therefore, we thought that physical activity might reduce effects of the diseases caused by disturbed body composition. Further studies are required to support this idea.

Keywords: body composition, body mass index, physical activity, university student

Procedia PDF Downloads 343
5841 A High Compression Ratio for a Losseless Image Compression Based on the Arithmetic Coding with the Sorted Run Length Coding: Meteosat Second Generation Image Compression

Authors: Cherifi Mehdi, Lahdir Mourad, Ameur Soltane

Abstract:

Image compression is the heart of several multimedia techniques. It is used to reduce the number of bits required to represent an image. Meteosat Second Generation (MSG) satellite allows the acquisition of 12 image files every 15 minutes and that results in a large databases sizes. In this paper, a novel image compression method based on the arithmetic coding with the sorted Run Length Coding (SRLC) for MSG images is proposed. The SRLC allows us to find the occurrence of the consecutive pixels of the original image to create a sorted run. The arithmetic coding allows the encoding of the sorted data of the previous stage to retrieve a unique code word that represents a binary code stream in the sorted order to boost the compression ratio. Through this article, we show that our method can perform the best results concerning compression ratio and bit rate unlike the method based on the Run Length Coding (RLC) and the arithmetic coding. Evaluation criteria like the compression ratio and the bit rate allow the confirmation of the efficiency of our method of image compression.

Keywords: image compression, arithmetic coding, Run Length Coding, RLC, Sorted Run Length Coding, SRLC, Meteosat Second Generation, MSG

Procedia PDF Downloads 336
5840 Mathematical and Numerical Analysis of a Reaction Diffusion System of Lambda-Omega Type

Authors: Hassan Al Salman, Ahmed Al Ghafli

Abstract:

In this study we consider a nonlinear in time finite element approximation of a reaction diffusion system of lambda-omega type. We use a fixed point theorem to prove existence of the approximations. Then, we derive some essential stability estimates and discuss the uniqueness of the approximations. Also, we prove an optimal error bound in time for d=1, 2 and 3 space dimensions. Finally, we present some numerical experiments to verify the theoretical results.

Keywords: reaction diffusion system, finite element approximation, fixed point theorem, an optimal error bound

Procedia PDF Downloads 516
5839 Development and Structural Performance Evaluation on Slit Circular Shear Panel Damper

Authors: Daniel Y. Abebe, Jaehyouk Choi

Abstract:

There are several types of metal-based devices conceived as dampers for the seismic energy absorber whereby damages to the major structural components could be minimized for both new and existing structures. This paper aimed to develop and evaluate structural performance of slit circular shear panel damper for passive seismic energy protection by inelastic deformation. Structural evaluation was done using commercially available nonlinear FE simulation program. The main parameters considered are: diameter-to-thickness (D/t) ratio and slit length-to-width ratio (l/w). Depending on these parameters three different buckling modes and hysteretic behaviors were found: yielding prior to buckling without strength degradation, yielding prior to buckling with strength degradation, and yielding with buckling and strength degradation which forms pinching at initial displacement. The susceptible location at which the possible crack is initiated is also identified for selected specimens using rupture index.

Keywords: slit circular shear panel damper, hysteresis characteristics, slip length-to-width ratio, D/t ratio, FE analysis

Procedia PDF Downloads 383
5838 A Deep Learning Model with Greedy Layer-Wise Pretraining Approach for Optimal Syngas Production by Dry Reforming of Methane

Authors: Maryam Zarabian, Hector Guzman, Pedro Pereira-Almao, Abraham Fapojuwo

Abstract:

Dry reforming of methane (DRM) has sparked significant industrial and scientific interest not only as a viable alternative for addressing the environmental concerns of two main contributors of the greenhouse effect, i.e., carbon dioxide (CO₂) and methane (CH₄), but also produces syngas, i.e., a mixture of hydrogen (H₂) and carbon monoxide (CO) utilized by a wide range of downstream processes as a feedstock for other chemical productions. In this study, we develop an AI-enable syngas production model to tackle the problem of achieving an equivalent H₂/CO ratio [1:1] with respect to the most efficient conversion. Firstly, the unsupervised density-based spatial clustering of applications with noise (DBSAN) algorithm removes outlier data points from the original experimental dataset. Then, random forest (RF) and deep neural network (DNN) models employ the error-free dataset to predict the DRM results. DNN models inherently would not be able to obtain accurate predictions without a huge dataset. To cope with this limitation, we employ reusing pre-trained layers’ approaches such as transfer learning and greedy layer-wise pretraining. Compared to the other deep models (i.e., pure deep model and transferred deep model), the greedy layer-wise pre-trained deep model provides the most accurate prediction as well as similar accuracy to the RF model with R² values 1.00, 0.999, 0.999, 0.999, 0.999, and 0.999 for the total outlet flow, H₂/CO ratio, H₂ yield, CO yield, CH₄ conversion, and CO₂ conversion outputs, respectively.

Keywords: artificial intelligence, dry reforming of methane, artificial neural network, deep learning, machine learning, transfer learning, greedy layer-wise pretraining

Procedia PDF Downloads 70
5837 Improvement of Parallel Compressor Model in Dealing Outlet Unequal Pressure Distribution

Authors: Kewei Xu, Jens Friedrich, Kevin Dwinger, Wei Fan, Xijin Zhang

Abstract:

Parallel Compressor Model (PCM) is a simplified approach to predict compressor performance with inlet distortions. In PCM calculation, it is assumed that the sub-compressors’ outlet static pressure is uniform and therefore simplifies PCM calculation procedure. However, if the compressor’s outlet duct is not long and straight, such assumption frequently induces error ranging from 10% to 15%. This paper provides a revised calculation method of PCM that can correct the error. The revised method employs energy equation, momentum equation and continuity equation to acquire needed parameters and replace the equal static pressure assumption. Based on the revised method, PCM is applied on two compression system with different blades types. The predictions of their performance in non-uniform inlet conditions are yielded through the revised calculation method and are employed to evaluate the method’s efficiency. Validating the results by experimental data, it is found that although little deviation occurs, calculated result agrees well with experiment data whose error ranges from 0.1% to 3%. Therefore, this proves the revised calculation method of PCM possesses great advantages in predicting the performance of the distorted compressor with limited exhaust duct.

Keywords: parallel compressor model (pcm), revised calculation method, inlet distortion, outlet unequal pressure distribution

Procedia PDF Downloads 317
5836 Aggregate Supply Response of Some Livestock Commodities in Algeria: Cointegration- Vector Error Correction Model Approach

Authors: Amine M. Benmehaia, Amine Oulmane

Abstract:

The supply response of agricultural commodities to changes in price incentives is an important issue for the success of any policy reform in the agricultural sector. This study aims to quantify the responsiveness of producers of some livestock commodities to price incentives in Algerian context. Time series analysis is used on annual data for a period of 52 years (1966-2018). Both co-integration and vector error correction model (VECM) are used through the Nerlove model of partial adjustment. The study attempts to determine the long-run and short-run relationships along with the magnitudes of disequilibria in the selected commodities. Results show that the short-run price elasticities are low in cow and sheep meat sectors (8.7 and 8% respectively), while their respective long-run elasticities are 16.5 and 10.5, whereas eggs and milk have very high short-run price elasticities (82 and 90% respectively) with long-run elasticities of 40 and 46 respectively. The error correction coefficient, reflecting the speed of adjustment towards the long-run equilibrium, is statistically significant and have the expected negative sign. Its estimates are 12.7 for cow meat, 33.5 for sheep meat, 46.7 for eggs and 8.4 for milk. It seems that cow meat and milk producers have a weak feedback of about 12.7% and 8.4% respectively of the previous year's disequilibrium from the long-run price elasticity, whereas sheep meat and eggs producers adjust to correct long run disequilibrium with a high speed of adjustment (33.5% and 46.7 % respectively). The implication of this is that much more in-depth research is needed to identify those factors that affect agricultural supply and to describe the effect of factors that shift supply in response to price incentives. This could provide valuable information for government in the use of appropriate policy measures.

Keywords: Algeria, cointegration, livestock, supply response, vector error correction model

Procedia PDF Downloads 118
5835 Effect of Packing Ratio on Fire Spread across Discrete Fuel Beds: An Experimental Analysis

Authors: Qianqian He, Naian Liu, Xiaodong Xie, Linhe Zhang, Yang Zhang, Weidong Yan

Abstract:

In the wild, the vegetation layer with exceptionally complex fuel composition and heterogeneous spatial distribution strongly affects the rate of fire spread (ROS) and fire intensity. Clarifying the influence of fuel bed structure on fire spread behavior is of great significance to wildland fire management and prediction. The packing ratio is one of the key physical parameters describing the property of the fuel bed. There is a threshold value of the packing ratio for ROS, but little is known about the controlling mechanism. In this study, to address this deficiency, a series of fire spread experiments were performed across a discrete fuel bed composed of some regularly arranged laser-cut cardboards, with constant wind speed and different packing ratios (0.0125-0.0375). The experiment aims to explore the relative importance of the internal and surface heat transfer with packing ratio. The dependence of the measured ROS on the packing ratio was almost consistent with the previous researches. The data of the radiative and total heat fluxes show that the internal heat transfer and surface heat transfer are both enhanced with increasing packing ratio (referred to as ‘Stage 1’). The trend agrees well with the variation of the flame length. The results extracted from the video show that the flame length markedly increases with increasing packing ratio in Stage 1. Combustion intensity is suggested to be increased, which, in turn, enhances the heat radiation. The heat flux data shows that the surface heat transfer appears to be more important than the internal heat transfer (fuel preheating inside the fuel bed) in Stage 1. On the contrary, the internal heat transfer dominates the fuel preheating mechanism when the packing ratio further increases (referred to as ‘Stage 2’) because the surface heat flux keeps almost stable with the packing ratio in Stage 2. As for the heat convection, the flow velocity was measured using Pitot tubes both inside and on the upper surface of the fuel bed during the fire spread. Based on the gas velocity distribution ahead of the flame front, it is found that the airflow inside the fuel bed is restricted in Stage 2, which can reduce the internal heat convection in theory. However, the analysis indicates not the influence of inside flow on convection and combustion, but the decreased internal radiation of per unit fuel is responsible for the decrease of ROS.

Keywords: discrete fuel bed, fire spread, packing ratio, wildfire

Procedia PDF Downloads 126
5834 Impact of Weather Conditions on Generalized Frequency Division Multiplexing over Gamma Gamma Channel

Authors: Muhammad Sameer Ahmed, Piotr Remlein, Tansal Gucluoglu

Abstract:

The technique called as Generalized frequency division multiplexing (GFDM) used in the free space optical channel can be a good option for implementation free space optical communication systems. This technique has several strengths e.g. good spectral efficiency, low peak-to-average power ratio (PAPR), adaptability and low co-channel interference. In this paper, the impact of weather conditions such as haze, rain and fog on GFDM over the gamma-gamma channel model is discussed. A Trade off between link distance and system performance under intense weather conditions is also analysed. The symbol error probability (SEP) of GFDM over the gamma-gamma turbulence channel is derived and verified with the computer simulations.

Keywords: free space optics, generalized frequency division multiplexing, weather conditions, gamma gamma distribution

Procedia PDF Downloads 156
5833 Cellular Traffic Prediction through Multi-Layer Hybrid Network

Authors: Supriya H. S., Chandrakala B. M.

Abstract:

Deep learning based models have been recently successful adoption for network traffic prediction. However, training a deep learning model for various prediction tasks is considered one of the critical tasks due to various reasons. This research work develops Multi-Layer Hybrid Network (MLHN) for network traffic prediction and analysis; MLHN comprises the three distinctive networks for handling the different inputs for custom feature extraction. Furthermore, an optimized and efficient parameter-tuning algorithm is introduced to enhance parameter learning. MLHN is evaluated considering the “Big Data Challenge” dataset considering the Mean Absolute Error, Root Mean Square Error and R^2as metrics; furthermore, MLHN efficiency is proved through comparison with a state-of-art approach.

Keywords: MLHN, network traffic prediction

Procedia PDF Downloads 70
5832 Methanol Steam Reforming with Heat Recovery for Hydrogen-Rich Gas Production

Authors: Horng-Wen Wu, Yi Chao, Rong-Fang Horng

Abstract:

This study is to develop a methanol steam reformer with a heat recovery zone, which recovers heat from exhaust gas of a diesel engine, and to investigate waste heat recovery ratio at the required reaction temperature. The operation conditions of the reformer are reaction temperature (200 °C, 250 °C, and 300 °C), steam to carbonate (S/C) ratio (0.9, 1.1, and 1.3), and N2 volume flow rate (40 cm3/min, 70 cm3/min, and 100 cm3/min). Finally, the hydrogen concentration, the CO, CO2, and N2 concentrations are measured and recorded to calculate methanol conversion efficiency, hydrogen flow rate, and assisting combustion gas and impeding combustion gas ratio. The heat source of this reformer comes from electric heater and waste heat of exhaust gas from diesel engines. The objective is to recover waste heat from the engine and to make more uniform temperature distribution within the reformer. It is beneficial for the reformer to enhance the methanol conversion efficiency and hydrogen-rich gas production. Experimental results show that the highest hydrogen flow rate exists at N2 of the volume rate 40 cm3/min and reforming reaction temperature of 300 °C and the value is 19.6 l/min. With the electric heater and heat recovery from exhaust gas, the maximum heat recovery ratio is 13.18 % occurring at water-methanol (S/C) ratio of 1.3 and the reforming reaction temperature of 300 °C.

Keywords: heat recovery, hydrogen-rich production, methanol steam reformer, methanol conversion efficiency

Procedia PDF Downloads 450
5831 Alumina Supported Copper-Manganese-Cobalt Catalysts for CO and VOCs Oxidation

Authors: Elitsa Kolentsova, Dimitar Dimitrov, Vasko Idakiev, Tatyana Tabakova, Krasimir Ivanov

Abstract:

Formaldehyde production by selective oxidation of methanol is an important industrial process. The main by-products in the waste gas are CO and dimethyl ether (DME). The idea of this study is to combine the advantages of both Cu-Mn and Cu-Co catalytic systems by obtaining a new mixed Cu-Mn-Co catalyst with high activity and selectivity at the simultaneous oxidation of CO, methanol, and DME. Two basic Cu-Mn samples with high activity were selected for further investigation: (i) manganese-rich Cu-Mn/γ–Al2O3 catalyst with Cu/Mn molar ratio 1:5 and (ii) copper-rich Cu-Mn/γ-Al2O3 catalyst with Cu/Mn molar ratio 2:1. Manganese in these samples was replaced by cobalt in the whole concentration region, and catalytic properties were determined. The results show a general trend of decreasing the activity toward DME oxidation and increasing the activity toward CO and methanol oxidation with the increase of cobalt up to 60% for both groups of catalyst. This general trend, however, contains specific features, depending on the composition of the catalyst and the nature of the oxidized gas. The catalytic activity of the sample with Cu/(Mn+Co) molar ratio of 2:1 is gradually changed with increasing the cobalt content. The activity of the sample with Cu/(Mn+Co) molar ratio of 1: 5 passes through a maximum at 60% manganese replacement by cobalt, probably due to the formation of highly dispersed Co-based spinel structures (Co3O4 and/or MnCo2O4). In conclusion, the present study demonstrates that the Cu-Mn-Co/γ–alumina supported catalysts have enhanced activity toward CO, methanol and DME oxidation. Cu/(Mn+Co) molar ratio 1:5 and Co/Mn molar ratio 1.5 in the active component can ensure successful oxidation of CO, CH3OH and DME. The active component of the mixed Cu-Mn-Co/γ–alumina catalysts consists of at least six compounds - CuO, Co3O4, MnO2, Cu1.5Mn1.5O4, MnCo2O4 and CuCo2O4, depending on the Cu/Mn/Co molar ratio. Chemical composition strongly influences catalytic properties, this effect being quite variable with regards to the different processes.

Keywords: Cu-Mn-Co catalysts, oxidation, carbon oxide, VOCs

Procedia PDF Downloads 208
5830 An Evaluation of the Relationship between the Anthropometric Measurements and Blood Lipid Profiles in Adolescents

Authors: Nalan Hakime Nogay

Abstract:

Childhood obesity is a significant health issue that is currently on the rise all over the world. In recent years, the relationship between childhood obesity and cardiovascular disease risk has been pointed out. The purpose of this study is to evaluate the relationship between some of the anthropometric indicators and blood lipid levels in adolescents. The present study has been conducted on a total of 252 adolescents -200 girls and 52 boys- within an age group of 12 to 18 years. Blood was drawn from each participant in the morning -after having fasted for 10 hours from the day before- to analyze their total cholesterol, HDL, LDL and triglyceride levels. Their body weight, height, waist circumference, subscapular skinfold thicknesses and triceps skinfold thicknesses measurements were taken and their individual waist/height ratios, BMI and body fat ratios were calculated. The blood lipid levels of the participants were categorized as acceptable, borderline and high in accordance with the 2011 Expert Panel Integrated Guidelines. The body fat ratios, total blood cholesterol and HDL levels of the girls were significantly higher than the boys whereas their waist circumference values were lower. The triglyceride levels, total cholesterol/HDL, LDL/HDL, triglyceride/HDL ratios of the group with the BMI ≥ 95 percentile ratio (the obese group) were higher than the groups that were considered to be overweight and normal weight as per their respective BMI values, while the HDL level of the obese group was lower; a fact that was found to be statistically significant. No significant relationship could be established, however, between the total blood cholesterol and LDL levels with their anthropometric measurements. The BMI, waist circumference, waist/height ratio, body fat ratio and triglyceride level of the group with the higher triglyceride level ( ≥ 130mg/dl) were found to be significantly higher compared to borderline (90-129 mg/dl) and the normal group (< 90 mg/dl). The BMI, waist circumference, waist/height ratio values of the group with the lower HDL level ( < 40 mg/dl) were significantly higher than the normal ( > 45 mg/dl) and borderline (40-45 mg/dl) groups. All of the anthropometric measurements of the group with the higher triglyceride/HDL ratio ( ≥ 3) were found to be significantly higher than that of the group with the lower ratio (< 3). Having a high BMI, waist/height ratio and waist circumference is related to low HDL and high blood triglyceride and triglyceride/HDL ratio. A high body fat ratio, on the other hand, is associated with a low HDL and high triglyceride/HDL ratio. Tackling childhood and adolescent obesity are important in terms of preventing cardiovascular diseases.

Keywords: adolescent, body fat, body mass index, lipid profile

Procedia PDF Downloads 250
5829 Comparison of Agree Method and Shortest Path Method for Determining the Flow Direction in Basin Morphometric Analysis: Case Study of Lower Tapi Basin, Western India

Authors: Jaypalsinh Parmar, Pintu Nakrani, Bhaumik Shah

Abstract:

Digital Elevation Model (DEM) is elevation data of the virtual grid on the ground. DEM can be used in application in GIS such as hydrological modelling, flood forecasting, morphometrical analysis and surveying etc.. For morphometrical analysis the stream flow network plays a very important role. DEM lacks accuracy and cannot match field data as it should for accurate results of morphometrical analysis. The present study focuses on comparing the Agree method and the conventional Shortest path method for finding out morphometric parameters in the flat region of the Lower Tapi Basin which is located in the western India. For the present study, open source SRTM (Shuttle Radar Topography Mission with 1 arc resolution) and toposheets issued by Survey of India (SOI) were used to determine the morphometric linear aspect such as stream order, number of stream, stream length, bifurcation ratio, mean stream length, mean bifurcation ratio, stream length ratio, length of overland flow, constant of channel maintenance and aerial aspect such as drainage density, stream frequency, drainage texture, form factor, circularity ratio, elongation ratio, shape factor and relief aspect such as relief ratio, gradient ratio and basin relief for 53 catchments of Lower Tapi Basin. Stream network was digitized from the available toposheets. Agree DEM was created by using the SRTM and stream network from the toposheets. The results obtained were used to demonstrate a comparison between the two methods in the flat areas.

Keywords: agree method, morphometric analysis, lower Tapi basin, shortest path method

Procedia PDF Downloads 223
5828 Profitability Assessment of Granite Aggregate Production and the Development of a Profit Assessment Model

Authors: Melodi Mbuyi Mata, Blessing Olamide Taiwo, Afolabi Ayodele David

Abstract:

The purpose of this research is to create empirical models for assessing the profitability of granite aggregate production in Akure, Ondo state aggregate quarries. In addition, an artificial neural network (ANN) model and multivariate predicting models for granite profitability were developed in the study. A formal survey questionnaire was used to collect data for the study. The data extracted from the case study mine for this study includes granite marketing operations, royalty, production costs, and mine production information. The following methods were used to achieve the goal of this study: descriptive statistics, MATLAB 2017, and SPSS16.0 software in analyzing and modeling the data collected from granite traders in the study areas. The ANN and Multi Variant Regression models' prediction accuracy was compared using a coefficient of determination (R²), Root mean square error (RMSE), and mean square error (MSE). Due to the high prediction error, the model evaluation indices revealed that the ANN model was suitable for predicting generated profit in a typical quarry. More quarries in Nigeria's southwest region and other geopolitical zones should be considered to improve ANN prediction accuracy.

Keywords: national development, granite, profitability assessment, ANN models

Procedia PDF Downloads 82
5827 Improved Performance Scheme for Joint Transmission in Downlink Coordinated Multi-Point Transmission

Authors: Young-Su Ryu, Su-Hyun Jung, Myoung-Jin Kim, Hyoung-Kyu Song

Abstract:

In this paper, improved performance scheme for joint transmission is proposed in downlink (DL) coordinated multi-point(CoMP) in case of constraint transmission power. This scheme is that serving transmission point (TP) request a joint transmission to inter-TP and selects one pre-coding technique according to channel state information(CSI) from user equipment(UE). The simulation results show that the bit error rate(BER) and throughput performances of the proposed scheme provide high spectral efficiency and reliable data at the cell edge.

Keywords: CoMP, joint transmission, minimum mean square error, zero-forcing, zero-forcing dirty paper coding

Procedia PDF Downloads 536
5826 Identification of Architectural Design Error Risk Factors in Construction Projects Using IDEF0 Technique

Authors: Sahar Tabarroki, Ahad Nazari

Abstract:

The design process is one of the most key project processes in the construction industry. Although architects have the responsibility to produce complete, accurate, and coordinated documents, architectural design is accompanied by many errors. A design error occurs when the constraints and requirements of the design are not satisfied. Errors are potentially costly and time-consuming to correct if not caught early during the design phase, and they become expensive in either construction documents or in the construction phase. The aim of this research is to identify the risk factors of architectural design errors, so identification of risks is necessary. First, a literature review in the design process was conducted and then a questionnaire was designed to identify the risks and risk factors. The questions in the form of the questionnaire were based on the “similar service description of study and supervision of architectural works” published by “Vice Presidency of Strategic Planning & Supervision of I.R. Iran” as the base of architects’ tasks. Second, the top 10 risks of architectural activities were identified. To determine the positions of possible causes of risks with respect to architectural activities, these activities were located in a design process modeled by the IDEF0 technique. The research was carried out by choosing a case study, checking the design drawings, interviewing its architect and client, and providing a checklist in order to identify the concrete examples of architectural design errors. The results revealed that activities such as “defining the current and future requirements of the project”, “studies and space planning,” and “time and cost estimation of suggested solution” has a higher error risk than others. Moreover, the most important causes include “unclear goals of a client”, “time force by a client”, and “lack of knowledge of architects about the requirements of end-users”. For error detecting in the case study, lack of criteria, standards and design criteria, and lack of coordination among them, was a barrier, anyway, “lack of coordination between architectural design and electrical and mechanical facility”, “violation of the standard dimensions and sizes in space designing”, “design omissions” were identified as the most important design errors.

Keywords: architectural design, design error, risk management, risk factor

Procedia PDF Downloads 115
5825 Feature Location Restoration for Under-Sampled Photoplethysmogram Using Spline Interpolation

Authors: Hangsik Shin

Abstract:

The purpose of this research is to restore the feature location of under-sampled photoplethysmogram using spline interpolation and to investigate feasibility for feature shape restoration. We obtained 10 kHz-sampled photoplethysmogram and decimated it to generate under-sampled dataset. Decimated dataset has 5 kHz, 2.5 k Hz, 1 kHz, 500 Hz, 250 Hz, 25 Hz and 10 Hz sampling frequency. To investigate the restoration performance, we interpolated under-sampled signals with 10 kHz, then compared feature locations with feature locations of 10 kHz sampled photoplethysmogram. Features were upper and lower peak of photplethysmography waveform. Result showed that time differences were dramatically decreased by interpolation. Location error was lesser than 1 ms in both feature types. In 10 Hz sampled cases, location error was also deceased a lot, however, they were still over 10 ms.

Keywords: peak detection, photoplethysmography, sampling, signal reconstruction

Procedia PDF Downloads 351
5824 An Initial Evaluation of Newly Proposed Biomarker of Zinc Status in Humans: The Erythrocyte Linoleic Acid: Dihomo-γ-Linolenic Acid (LA:DGLA) Ratio

Authors: Marija Knez, James C.R. Stangoulis, Manja Zec, Zoran Pavlovic, Jasmina D. Martacic, Mirjana Gurinovic, Maria Glibetic

Abstract:

Background: Zinc is an essential micronutrient for humans with important physiological functions. A sensitive and specific biomarker for assessing Zn status is still needed. Objective: The major aim of this study was to examine if the changes in the content of plasma phospholipid LA, DGLA and LA: DGLA ratio can be used to efficiently predict the dietary Zn intake and plasma Zn status of humans. Methods: The study was performed on apparently healthy human volunteers. The dietary Zn intake was assessed using 24h recall questionnaires. Plasma phospholipid fatty acid analysis was done by gas chromatography and plasma analysis of minerals by atomic absorption spectrometry. Biochemical, anthropometrical and hematological parameters were assessed. Results: No significant relationship was found between the dietary and plasma zinc status (r=0.07; p=0.6). There is a statistically significant correlation between DGLA and plasma Zn (r=0.39, p=0.00). No relationship was observed between the linoleic acid and plasma Zn, while there was a significant negative correlation between LA: DGLA ratio and plasma Zn status (r=-0.35, p=0.01). Similarly, there were statistically significant difference in DGLA status (p=0.004) and LA: DGLA ratio (p=0.042) between the Zn formed groups. Conclusions: This study is an initial step in evaluating LA: DGLA ratio as a biomarker of Zn status in humans. The results are encouraging as they show that concentration of DGLA is decreased and LA: DGLA ratio increased in people with lower dietary Zn intake. However, additional studies are needed to fully examine the sensitivity of this biomarker.

Keywords: dietary Zn intake Zinc, fatty acid composition, LA: DGLA, healthy population, plasma Zn status, Zn biomarker

Procedia PDF Downloads 256
5823 The Effect of Velocity Increment by Blockage Factor on Savonius Hydrokinetic Turbine Performance

Authors: Thochi Seb Rengma, Mahendra Kumar Gupta, P. M. V. Subbarao

Abstract:

Hydrokinetic turbines can be used to produce power in inaccessible villages located near rivers. The hydrokinetic turbine uses the kinetic energy of the water and maybe put it directly into the natural flow of water without dams. For off-grid power production, the Savonius-type vertical axis turbine is the easiest to design and manufacture. This proposal uses three-dimensional computational fluid dynamics (CFD) simulations to measure the considerable interaction and complexity of turbine blades. Savonius hydrokinetic turbine (SHKT) performance is affected by a blockage in the river, canals, and waterways. Putting a large object in a water channel causes water obstruction and raises local free stream velocity. The blockage correction factor or velocity increment measures the impact of velocity on the performance. SHKT performance is evaluated by comparing power coefficient (Cp) with tip-speed ratio (TSR) at various blockage ratios. The maximum Cp was obtained at a TSR of 1.1 with a blockage ratio of 45%, whereas TSR of 0.8 yielded the highest Cp without blockage. The greatest Cp of 0.29 was obtained with a 45% blockage ratio compared to a Cp max of 0.18 without a blockage.

Keywords: savonius hydrokinetic turbine, blockage ratio, vertical axis turbine, power coefficient

Procedia PDF Downloads 112
5822 Assessing Effective Parameters on the Extraction of Copper from Pregnant Leach Solution Using Chemorex CP-150

Authors: Kimia Kiaei, Mohammad Hasan Golpayegani

Abstract:

The extraction of copper from a pregnant leach solution obtained through leaching was investigated in this study. Chemorex CP-150 was utilized as an organic extractant, while kerosene served as a diluent. The study focused on determining the optimal ratios of extractant to diluent, as well as the pH of the aqueous phase. Isotherm curves of extraction were generated, and Mc. Cabe-Thiele diagrams were constructed separately for an optimized experimental pH of 3.17 and a typical industrial pH of 2. Additionally, the sulfuric acid-to-PLS ratio and concentrations of interfering ions comprising Mn²⁺ and Fe³⁺ in the strip solution were evaluated during the stripping stage. The results indicated that the optimized values for the extractant-to-diluent ratio and pH were 5% and 3.17, respectively. The Mc. Cabe-Thiele diagrams revealed that at an aqueous-to-organic ratio of 1, the theoretical stages of solvent extraction at pH levels of 3.17 and 2 were one and two, respectively. Moreover, a sulfuric acid-to-PLS ratio of 20% was employed in the stripping stage, and it was observed that the concentrations of interfering ions fell within the acceptable range.

Keywords: copper, solvent extraction, heap leaching, Chemorex CP-150, pregnant leach solution

Procedia PDF Downloads 51
5821 Frequency Recognition Models for Steady State Visual Evoked Potential Based Brain Computer Interfaces (BCIs)

Authors: Zeki Oralhan, Mahmut Tokmakçı

Abstract:

SSVEP based brain computer interface (BCI) systems have been preferred, because of high information transfer rate (ITR) and practical use. ITR is the parameter of BCI overall performance. For high ITR value, one of specification BCI system is that has high accuracy. In this study, we investigated to recognize SSVEP with shorter time and lower error rate. In the experiment, there were 8 flickers on light crystal display (LCD). Participants gazed to flicker which had 12 Hz frequency and 50% duty cycle ratio on the LCD during 10 seconds. During the experiment, EEG signals were acquired via EEG device. The EEG data was filtered in preprocessing session. After that Canonical Correlation Analysis (CCA), Multiset CCA (MsetCCA), phase constrained CCA (PCCA), and Multiway CCA (MwayCCA) methods were applied on data. The highest average accuracy value was reached when MsetCCA was applied.

Keywords: brain computer interface, canonical correlation analysis, human computer interaction, SSVEP

Procedia PDF Downloads 251
5820 Maximum Initial Input Allowed to Iterative Learning Control Set-up Using Singular Values

Authors: Naser Alajmi, Ali Alobaidly, Mubarak Alhajri, Salem Salamah, Muhammad Alsubaie

Abstract:

Iterative Learning Control (ILC) known to be a controlling tool to overcome periodic disturbances for repetitive systems. This technique is required to let the error signal tends to zero as the number of operation increases. The learning process that lies within this context is strongly dependent on the initial input which if selected properly tends to let the learning process be more effective compared to the case where a system starts from blind. ILC uses previous recorded execution data to update the following execution/trial input such that a reference trajectory is followed to a high accuracy. Error convergence in ILC is generally highly dependent on the input applied to a plant for trial $1$, thus a good choice of initial starting input signal would make learning faster and as a consequence the error tends to zero faster as well. In the work presented within, an upper limit based on the Singular Values Principle (SV) is derived for the initial input signal applied at trial $1$ such that the system follow the reference in less number of trials without responding aggressively or exceeding the working envelope where a system is required to move within in a robot arm, for example. Simulation results presented illustrate the theory introduced within this paper.

Keywords: initial input, iterative learning control, maximum input, singular values

Procedia PDF Downloads 226
5819 The Non-Existence of Perfect 2-Error Correcting Lee Codes of Word Length 7 over Z

Authors: Catarina Cruz, Ana Breda

Abstract:

Tiling problems have been capturing the attention of many mathematicians due to their real-life applications. In this study, we deal with tilings of Zⁿ by Lee spheres, where n is a positive integer number, being these tilings related with error correcting codes on the transmission of information over a noisy channel. We focus our attention on the question ‘for what values of n and r does the n-dimensional Lee sphere of radius r tile Zⁿ?’. It seems that the n-dimensional Lee sphere of radius r does not tile Zⁿ for n ≥ 3 and r ≥ 2. Here, we prove that is not possible to tile Z⁷ with Lee spheres of radius 2 presenting a proof based on a combinatorial method and faithful to the geometric idea of the problem. The non-existence of such tilings has been studied by several authors being considered the most difficult cases those in which the radius of the Lee spheres is equal to 2. The relation between these tilings and error correcting codes is established considering the center of a Lee sphere as a codeword and the other elements of the sphere as words which are decoded by the central codeword. When the Lee spheres of radius r centered at elements of a set M ⊂ Zⁿ tile Zⁿ, M is a perfect r-error correcting Lee code of word length n over Z, denoted by PL(n, r). Our strategy to prove the non-existence of PL(7, 2) codes are based on the assumption of the existence of such code M. Without loss of generality, we suppose that O ∈ M, where O = (0, ..., 0). In this sense and taking into account that we are dealing with Lee spheres of radius 2, O covers all words which are distant two or fewer units from it. By the definition of PL(7, 2) code, each word which is distant three units from O must be covered by a unique codeword of M. These words have to be covered by codewords which dist five units from O. We prove the non-existence of PL(7, 2) codes showing that it is not possible to cover all the referred words without superposition of Lee spheres whose centers are distant five units from O, contradicting the definition of PL(7, 2) code. We achieve this contradiction by combining the cardinality of particular subsets of codewords which are distant five units from O. There exists an extensive literature on codes in the Lee metric. Here, we present a new approach to prove the non-existence of PL(7, 2) codes.

Keywords: Golomb-Welch conjecture, Lee metric, perfect Lee codes, tilings

Procedia PDF Downloads 145
5818 Assessment of Time-variant Work Stress for Human Error Prevention

Authors: Hyeon-Kyo Lim, Tong-Il Jang, Yong-Hee Lee

Abstract:

For an operator in a nuclear power plant, human error is one of the most dreaded factors that may result in unexpected accidents. The possibility of human errors may be low, but the risk of them would be unimaginably enormous. Thus, for accident prevention, it is quite indispensable to analyze the influence of any factors which may raise the possibility of human errors. During the past decades, not a few research results showed that performance of human operators may vary over time due to lots of factors. Among them, stress is known to be an indirect factor that may cause human errors and result in mental illness. Until now, not a few assessment tools have been developed to assess stress level of human workers. However, it still is questionable to utilize them for human performance anticipation which is related with human error possibility, because they were mainly developed from the viewpoint of mental health rather than industrial safety. Stress level of a person may go up or down with work time. In that sense, if they would be applicable in the safety aspect, they should be able to assess the variation resulted from work time at least. Therefore, this study aimed to compare their applicability for safety purpose. More than 10 kinds of work stress tools were analyzed with reference to assessment items, assessment and analysis methods, and follow-up measures which are known to close related factors with work stress. The results showed that most tools mainly focused their weights on some common organizational factors such as demands, supports, and relationships, in sequence. Their weights were broadly similar. However, they failed to recommend practical solutions. Instead, they merely advised to set up overall counterplans in PDCA cycle or risk management activities which would be far from practical human error prevention. Thus, it was concluded that application of stress assessment tools mainly developed for mental health seemed to be impractical for safety purpose with respect to human performance anticipation, and that development of a new assessment tools would be inevitable if anyone wants to assess stress level in the aspect of human performance variation and accident prevention. As a consequence, as practical counterplans, this study proposed a new scheme for assessment of work stress level of a human operator that may vary over work time which is closely related with the possibility of human errors.

Keywords: human error, human performance, work stress, assessment tool, time-variant, accident prevention

Procedia PDF Downloads 657
5817 Banking Sector Development and Economic Growth: Evidence from the State of Qatar

Authors: Fekri Shawtari

Abstract:

The banking sector plays a very crucial role in the economic development of the country. As a financial intermediary, it has assigned a great role in the economic growth and stability. This paper aims to examine the empirically the relationship between banking industry and economic growth in state of Qatar. We adopt the VAR vector error correction model (VECM) along with Granger causality to address the issue over the long-run and short-run between the banking sector and economic growth. It is expected that the results will give policy directions to the policymakers to make strategies that are conducive toward boosting development to achieve the targeted economic growth in current situation.

Keywords: economic growth, banking sector, Qatar, vector error correction model, VECM

Procedia PDF Downloads 154
5816 Virtual Assessment of Measurement Error in the Fractional Flow Reserve

Authors: Keltoum Chahour, Mickael Binois

Abstract:

Due to a lack of standardization during the invasive fractional flow reserve (FFR) procedure, the index is subject to many sources of uncertainties. In this paper, we investigate -through simulation- the effect of the (FFR) device position and configuration on the obtained value of the (FFR) fraction. For this purpose, we use computational fluid dynamics (CFD) in a 3D domain corresponding to a diseased arterial portion. The (FFR) pressure captor is introduced inside it with a given length and coefficient of bending to capture the (FFR) value. To get over the computational limitations, basically, the time of the simulation is about 2h 15min for one (FFR) value; we generate a Gaussian Process (GP) model for (FFR) prediction. The (GP) model indicates good accuracy and demonstrates the effective error in the measurement created by the random configuration of the pressure captor.

Keywords: fractional flow reserve, Gaussian processes, computational fluid dynamics, drift

Procedia PDF Downloads 111