Search results for: Fuzzy random variables
1357 Analysis of Education Faculty Students’ Attitudes towards E-Learning According to Different Variables
Authors: Eyup Yurt, Ahmet Kurnaz, Ismail Sahin
Abstract:
The purpose of the study is to investigate the education faculty students’ attitudes towards e-learning according to different variables. In current study, the data were collected from 393 students of an education faculty in Turkey. In this study, theattitude towards e‐learning scale and the demographic information form were used to collect data. The collected data were analyzed by t-test, ANOVA and Pearson correlation coefficient. It was found that there is a significant difference in students’ tendency towards e-learning and avoidance from e-learning based on gender. Male students have more positive attitudes towards e-learning than female students. Also, the students who used the internet lesshave higher levels of avoidance from e-learning. Additionally, it is found that there is a positive and significant relationship between the number of personal mobile learning devices and tendency towards e-learning. On the other hand, there is a negative and significant relationship between the number of personal mobile learning devices and avoidance from e-learning. Also, suggestions were presented according to findings.
Keywords: Education faculty students, attitude towards e-learning, gender, daily Internet usage time, m-learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21641356 Land Suitability Prediction Modelling for Agricultural Crops Using Machine Learning Approach: A Case Study of Khuzestan Province, Iran
Authors: Saba Gachpaz, Hamid Reza Heidari
Abstract:
The sharp increase in population growth leads to more pressure on agricultural areas to satisfy the food supply. This necessitates increased resource consumption and underscores the importance of addressing sustainable agriculture development along with other environmental considerations. Land-use management is a crucial factor in obtaining optimum productivity. Machine learning is a widely used technique in the agricultural sector, from yield prediction to customer behavior. This method focuses on learning and provides patterns and correlations from our data set. In this study, nine physical control factors, namely, soil classification, electrical conductivity, normalized difference water index (NDWI), groundwater level, elevation, annual precipitation, pH of water, annual mean temperature, and slope in the alluvial plain in Khuzestan (an agricultural hotspot in Iran) are used to decide the best agricultural land use for both rainfed and irrigated agriculture for 10 different crops. For this purpose, each variable was imported into Arc GIS, and a raster layer was obtained. In the next level, by using training samples, all layers were imported into the python environment. A random forest model was applied, and the weight of each variable was specified. In the final step, results were visualized using a digital elevation model, and the importance of all factors for each one of the crops was obtained. Our results show that despite 62% of the study area being allocated to agricultural purposes, only 42.9% of these areas can be defined as a suitable class for cultivation purposes.
Keywords: Land suitability, machine learning, random forest, sustainable agriculture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2831355 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison
Authors: Xiangtuo Chen, Paul-Henry Cournéde
Abstract:
Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.Keywords: Crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11751354 Identifying Factors Contributing to the Spread of Lyme Disease: A Regression Analysis of Virginia’s Data
Authors: Fatemeh Valizadeh Gamchi, Edward L. Boone
Abstract:
This research focuses on Lyme disease, a widespread infectious condition in the United States caused by the bacterium Borrelia burgdorferi sensu stricto. It is critical to identify environmental and economic elements that are contributing to the spread of the disease. This study examined data from Virginia to identify a subset of explanatory variables significant for Lyme disease case numbers. To identify relevant variables and avoid overfitting, linear poisson, and regularization regression methods such as ridge, lasso, and elastic net penalty were employed. Cross-validation was performed to acquire tuning parameters. The methods proposed can automatically identify relevant disease count covariates. The efficacy of the techniques was assessed using four criteria on three simulated datasets. Finally, using the Virginia Department of Health’s Lyme disease dataset, the study successfully identified key factors, and the results were consistent with previous studies.
Keywords: Lyme disease, Poisson generalized linear model, Ridge regression, Lasso Regression, elastic net regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1221353 LTE Performance Analysis in the City of Bogota Northern Zone for Two Different Mobile Broadband Operators over Qualipoc
Authors: Víctor D. Rodríguez, Edith P. Estupiñán, Juan C. Martínez
Abstract:
The evolution in mobile broadband technologies has allowed to increase the download rates in users considering the current services. The evaluation of technical parameters at the link level is of vital importance to validate the quality and veracity of the connection, thus avoiding large losses of data, time and productivity. Some of these failures may occur between the eNodeB (Evolved Node B) and the user equipment (UE), so the link between the end device and the base station can be observed. LTE (Long Term Evolution) is considered one of the IP-oriented mobile broadband technologies that work stably for data and VoIP (Voice Over IP) for those devices that have that feature. This research presents a technical analysis of the connection and channeling processes between UE and eNodeB with the TAC (Tracking Area Code) variables, and analysis of performance variables (Throughput, Signal to Interference and Noise Ratio (SINR)). Three measurement scenarios were proposed in the city of Bogotá using QualiPoc, where two operators were evaluated (Operator 1 and Operator 2). Once the data were obtained, an analysis of the variables was performed determining that the data obtained in transmission modes vary depending on the parameters BLER (Block Error Rate), performance and SNR (Signal-to-Noise Ratio). In the case of both operators, differences in transmission modes are detected and this is reflected in the quality of the signal. In addition, due to the fact that both operators work in different frequencies, it can be seen that Operator 1, despite having spectrum in Band 7 (2600 MHz), together with Operator 2, is reassigning to another frequency, a lower band, which is AWS (1700 MHz), but the difference in signal quality with respect to the establishment with data by the provider Operator 2 and the difference found in the transmission modes determined by the eNodeB in Operator 1 is remarkable.
Keywords: BLER, LTE, Network, Qualipoc, SNR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5351352 Design and Development of Real-Time Optimal Energy Management System for Hybrid Electric Vehicles
Authors: Masood Roohi, Amir Taghavipour
Abstract:
This paper describes a strategy to develop an energy management system (EMS) for a charge-sustaining power-split hybrid electric vehicle. This kind of hybrid electric vehicles (HEVs) benefit from the advantages of both parallel and series architecture. However, it gets relatively more complicated to manage power flow between the battery and the engine optimally. The applied strategy in this paper is based on nonlinear model predictive control approach. First of all, an appropriate control-oriented model which was accurate enough and simple was derived. Towards utilization of this controller in real-time, the problem was solved off-line for a vast area of reference signals and initial conditions and stored the computed manipulated variables inside look-up tables. Look-up tables take a little amount of memory. Also, the computational load dramatically decreased, because to find required manipulated variables the controller just needed a simple interpolation between tables.Keywords: Hybrid electric vehicles, energy management system, nonlinear model predictive control, real-time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13871351 Modality and Redundancy Effects on Music Theory Learning Among Pupils of Different Anxiety Levels
Authors: Soon Fook Fong, Aldalalah, M. Osamah
Abstract:
The purpose of this study was to investigate effects of modality and redundancy principles on music theory learning among pupils of different anxiety levels. The lesson of music theory was developed in three different modes, audio and image (AI), text with image (TI) and audio with image and text (AIT). The independent variables were the three modes of courseware. The moderator variable was the anxiety level, while the dependent variable was the post test score. The study sample consisted of 405 third-grade pupils. Descriptive and inferential statistics were conducted to analyze the collected data. Analyses of covariance (ANCOVA) and Post hoc were carried out to examine the main effects as well as the interaction effects of the independent variables on the dependent variable. The findings of this study showed that medium anxiety pupils performed significantly better than low and high anxiety pupils in all the three treatment modes. The AI mode was found to help pupils with high anxiety significantly more than the TI and AIT modes.Keywords: Modality, Redundancy, Music theory, Cognitivetheory of multimedia learning, Cognitive load theory, Anxiety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21361350 Improved Segmentation of Speckled Images Using an Arithmetic-to-Geometric Mean Ratio Kernel
Abstract:
In this work, we improve a previously developed segmentation scheme aimed at extracting edge information from speckled images using a maximum likelihood edge detector. The scheme was based on finding a threshold for the probability density function of a new kernel defined as the arithmetic mean-to-geometric mean ratio field over a circular neighborhood set and, in a general context, is founded on a likelihood random field model (LRFM). The segmentation algorithm was applied to discriminated speckle areas obtained using simple elliptic discriminant functions based on measures of the signal-to-noise ratio with fractional order moments. A rigorous stochastic analysis was used to derive an exact expression for the cumulative density function of the probability density function of the random field. Based on this, an accurate probability of error was derived and the performance of the scheme was analysed. The improved segmentation scheme performed well for both simulated and real images and showed superior results to those previously obtained using the original LRFM scheme and standard edge detection methods. In particular, the false alarm probability was markedly lower than that of the original LRFM method with oversegmentation artifacts virtually eliminated. The importance of this work lies in the development of a stochastic-based segmentation, allowing an accurate quantification of the probability of false detection. Non visual quantification and misclassification in medical ultrasound speckled images is relatively new and is of interest to clinicians.Keywords: Discriminant function, false alarm, segmentation, signal-to-noise ratio, skewness, speckle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16551349 Customer Churn Prediction Using Four Machine Learning Algorithms Integrating Feature Selection and Normalization in the Telecom Sector
Authors: Alanoud Moraya Aldalan, Abdulaziz Almaleh
Abstract:
A crucial part of maintaining a customer-oriented business in the telecommunications industry is understanding the reasons and factors that lead to customer churn. Competition between telecom companies has greatly increased in recent years, which has made it more important to understand customers’ needs in this strong market. For those who are looking to turn over their service providers, understanding their needs is especially important. Predictive churn is now a mandatory requirement for retaining customers in the telecommunications industry. Machine learning can be used to accomplish this. Churn Prediction has become a very important topic in terms of machine learning classification in the telecommunications industry. Understanding the factors of customer churn and how they behave is very important to building an effective churn prediction model. This paper aims to predict churn and identify factors of customers’ churn based on their past service usage history. Aiming at this objective, the study makes use of feature selection, normalization, and feature engineering. Then, this study compared the performance of four different machine learning algorithms on the Orange dataset: Logistic Regression, Random Forest, Decision Tree, and Gradient Boosting. Evaluation of the performance was conducted by using the F1 score and ROC-AUC. Comparing the results of this study with existing models has proven to produce better results. The results showed the Gradients Boosting with feature selection technique outperformed in this study by achieving a 99% F1-score and 99% AUC, and all other experiments achieved good results as well.
Keywords: Machine Learning, Gradient Boosting, Logistic Regression, Churn, Random Forest, Decision Tree, ROC, AUC, F1-score.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4081348 A Study of the Variables in the Optimisation of a Platinum Precipitation Process
Authors: Tebogo Phetla, Edison Muzenda, M Belaid
Abstract:
This study investigated possible ways to improve the efficiency of the platinum precipitation process using ammonium chloride by reducing the platinum content reporting to the effluent. The ore treated consist of five platinum group metals namely, ruthenium, rhodium, iridium, platinum, palladium and a precious metal gold. Gold, ruthenium, rhodium and iridium were extracted prior the platinum precipitation process. Temperature, reducing agent, flow rate and potential difference were the variables controlled to determine the operation conditions for optimum platinum precipitation efficiency. Hydrogen peroxide was added as the oxidizing agent at the temperature of 85-90oC and potential difference of 700-850mV was the variable used to check the oxidizing state of platinum. The platinum was further purified at temperature between 60-65oC, potential difference above 700 mV, ammonium chloride of 200 l, and at these conditions the platinum content reporting to the effluent was reduced to less than 300ppm, resulting in optimum platinum precipitation efficiency and purity of 99.9%.Keywords: Platinum Group Metals (PGM), Potential difference, Precipitation, Redox reactions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47891347 Causal Factors Affecting on Trustworthiness and Success of the National Press Council of Thailand in Regulating Professional Ethics in Views of Newspaper Journalists
Authors: Bubpha Makesrithongkum
Abstract:
The objectives of this research were 1) to study the opinions of newspaper journalists about their trustworthiness in the National Press Council of Thailand (NPCT) and the NPCT-s success in regulating the professional ethics; and 2) to study the differences among mean vectors of the variables of trustworthiness in the NPCT and opinions on the NPCT-s success in regulating professional ethics among samples working at different work positions and from different affiliation of newspaper organizations. The results showed that 1) Interaction effects between the variables of work positions and affiliation were not statistically significant at the confidence level of 0.05. 2) There was a statistically significant difference (p<0.05) in the views of journalists (reporters, heads of news desks and editors) at newspapers in the Bangkok metropolis and at local newspapers in other regions regarding their level of trustworthiness in the NPCT-s fulfillment of its duty to regulate professional ethics.Keywords: National Press Council of Thailand, newspaper journalists, regulation of newspaper professional ethics, trustworthiness and success in fulfilling duties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15791346 Diversity and Structure of Trichoptera Communities and Water Quality Variables in Streams, Northern Thailand
Authors: T. Prommi, P. Thamsenanupap
Abstract:
The influence of physicochemical water quality parameters on the abundance and diversity of caddisfly larvae was studied in seven sampling stations in Mae Tao and Mae Ku watersheds, Mae Sot District, Tak Province, northern Thailand. The streams: MK2 and MK8 as reference site, and impacted streams (MT1-MT5) were sampled bi-monthly during July 2011 to May 2012. A total of 4,584 individual of caddisfly larvae belonging to 10 family and 17 genera were found. The larvae of family Hydropsychidae were the most abundance, followed by Philopotamidae, Odontoceridae, and Leptoceridae, respectively. The genus Cheumatopsyche, Hydropsyche, and Chimarra were the most abundance genera in this study. Results of CCA ordination showed the total dissolved solids, sulfate, water temperature, dissolved oxygen and pH were the most important physicochemical factors to affect distribution of caddisflies communities. Changes in the caddisfly fauna may indicate changes in physicochemical factors owing to agricultural pollution, urbanization, or other human activities. Results revealed that the order Trichoptera, identified to species or genus, can be potentially used to assess environmental water quality status in freshwater ecosystems.Keywords: Caddisfly larvae, environmental variables, diversity, streams.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21201345 Quantification of Aerodynamic Variables Using Analytical Technique and Computational Fluid Dynamics
Authors: Adil Loya, Kamran Maqsood, Muhammad Duraid
Abstract:
Aerodynamic stability coefficients are necessary to be known before any unmanned aircraft flight is performed. This requires expertise on aerodynamics and stability control of the aircraft. To enable efficacious performance of aircraft requires that a well-defined flight path and aerodynamics should be defined beforehand. This paper presents a study on the aerodynamics of an unmanned aero vehicle (UAV) during flight conditions. Current research holds comparative studies of different parameters for flight aerodynamic, measured using two different open source analytical software programs. These software packages are DATCOM and XLRF5, which help in depicting the flight aerodynamic variables. Computational fluid dynamics (CFD) was also used to perform aerodynamic analysis for which Star CCM+ was used. Output trends of the study demonstrate high accuracies between the two software programs with that of CFD. It can be seen that the Coefficient of Lift (CL) obtained from DATCOM and XFLR is similar to CL of CFD simulation. In the similar manner, other potential aerodynamic stability parameters obtained from analytical software are in good agreement with CFD.
Keywords: XFLR5, DATCOM, computational fluid dynamic, unmanned aero vehicle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8761344 Buckling Optimization of Radially-Graded, Thin-Walled, Long Cylinders under External Pressure
Authors: Karam Y. Maalawi
Abstract:
This paper presents a generalized formulation for the problem of buckling optimization of anisotropic, radially graded, thin-walled, long cylinders subject to external hydrostatic pressure. The main structure to be analyzed is built of multi-angle fibrous laminated composite lay-ups having different volume fractions of the constituent materials within the individual plies. This yield to a piecewise grading of the material in the radial direction; that is the physical and mechanical properties of the composite material are allowed to vary radially. The objective function is measured by maximizing the critical buckling pressure while preserving the total structural mass at a constant value equals to that of a baseline reference design. In the selection of the significant optimization variables, the fiber volume fractions adjoin the standard design variables including fiber orientation angles and ply thicknesses. The mathematical formulation employs the classical lamination theory, where an analytical solution that accounts for the effective axial and flexural stiffness separately as well as the inclusion of the coupling stiffness terms is presented. The proposed model deals with dimensionless quantities in order to be valid for thin shells having arbitrary thickness-to-radius ratios. The critical buckling pressure level curves augmented with the mass equality constraint are given for several types of cylinders showing the functional dependence of the constrained objective function on the selected design variables. It was shown that material grading can have significant contribution to the whole optimization process in achieving the required structural designs with enhanced stability limits.Keywords: Buckling instability, structural optimization, functionally graded material, laminated cylindrical shells, externalhydrostatic pressure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23591343 Risk of Late Payment in the Malaysian Construction Industry
Authors: Kho Mei Ye, Hamzah Abdul Rahman
Abstract:
The purpose of this study is to identify the underlying causes of late payment from the contractors- perspective in the Malaysian construction industry and to recommend effective solutions to mitigate late payment problems. The target groups of respondents in this study were Grades G3, G5, G6 and G7 contractors with specialization in building works and civil engineering works registered with the Construction Industry Development Board (CIDB) in Malaysia. Results from this study were analyzed with Statistical Package for the Social Science (SPSS 15.0). From this study, it was found that respondents have highest ranked five significant variables out of a total of forty-one variables which can caused late payment problems: a) cash flow problems due to deficiencies in client-s management capacity (mean = 3.96); b) client-s ineffective utilization of funds (mean = 3.88); c) scarcity of capital to finance the project (mean = 3.81); d) clients failure to generate income from bank when sales of houses do not hit the targeted amount (mean=3.72); and e) poor cash flow because of lack of proper process implementation, delay in releasing of the retention monies to contractor and delay in the evaluation and certification of interim and final payment (mean = 3.66).Keywords: Underlying causes, late payment, constructionindustry, Malaysia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 71921342 Construction of Space-Filling Designs for Three Input Variables Computer Experiments
Authors: Kazeem A. Osuolale, Waheed B. Yahya, Babatunde L. Adeleke
Abstract:
Latin hypercube designs (LHDs) have been applied in many computer experiments among the space-filling designs found in the literature. A LHD can be randomly generated but a randomly chosen LHD may have bad properties and thus act poorly in estimation and prediction. There is a connection between Latin squares and orthogonal arrays (OAs). A Latin square of order s involves an arrangement of s symbols in s rows and s columns, such that every symbol occurs once in each row and once in each column and this exists for every non-negative integer s. In this paper, a computer program was written to construct orthogonal array-based Latin hypercube designs (OA-LHDs). Orthogonal arrays (OAs) were constructed from Latin square of order s and the OAs constructed were afterward used to construct the desired Latin hypercube designs for three input variables for use in computer experiments. The LHDs constructed have better space-filling properties and they can be used in computer experiments that involve only three input factors. MATLAB 2012a computer package (www.mathworks.com/) was used for the development of the program that constructs the designs.Keywords: Computer Experiments, Latin Squares, Latin Hypercube Designs, Orthogonal Array, Space-filling Designs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17231341 The Association between C-Reactive Protein and Hypertension of Different United States Participants Categorized by Ethnicity: Applying the National Health and Nutrition Examination Survey from 1999-2010
Authors: Ghada Abo-Zaid
Abstract:
Objectives: The main objective of this study was to examine the association between the elevated level of C-reactive protein (CRP) and incidence of hypertension before and after adjustments for age, BMI, gender, SES, smoking, diabetes, cholesterol LDL and cholesterol HDL, and to determine whether the association differs by race. Method: Cross sectional data for participants from aged 17 years to 74 years, included in The National Health and Nutrition Examination Survey (NHANES) from 1999 to 2010 were analyzed. The CRP level was classified into three categories (> 3 mg/L, between 1 mg/L and 3 mg/L, and < 3 mg/L). Blood pressure categorization was done using JNC 7 indicator. Hypertension is defined as either systolic blood pressure (SBP) of 140 mmHg or more and diastolic blood pressure (DBP) of 90 mmHg or more, otherwise a self-reported prior diagnosis by a physician. Pre-hypertension was defined as 139 ≥ SBP > 120 or 89 ≥ DBP >80. Multinominal regression model was undertaken to measure the association between CRP level and hypertension. Results: In univariable models, CRP concentrations > 3 mg/L were associated with a 73% greater risk of incident hypertension compared with CRP concentrations < 1 mg/L (Hypertension: odds ratio [OR] = 1.73; 95% confidence interval [CI], 1.50-1.99). Ethnic comparisons showed that American Mexicans had the highest risk of incident hypertension (OR = 2.39; 95% CI, 2.21-2.58). This risk was statistically insignificant after controlling by other variables (Hypertension: OR = 0.75; 95% CI, 0.52-1.08), or categorized by race [American Mexican: OR= 1.58; 95% CI, 0.58-4.26, Other Hispanic: OR = 0.87; 95% CI, 0.19-4.42, Non-Hispanic white: OR = 0.90; 95% CI, 0.50-1.59, Non-Hispanic Black: OR = 0.44; 95% CI, 0.22-0.87. The same results were found for pre-hypertension, and the Non-Hispanic black segment showed the highest significant risk for Pre-Hypertension (OR = 1.60; 95% CI, 1.26-2.03). When CRP concentrations were between 1.0 and 3.0 mg/L in unadjusted models, prehypertension was associated with higher likelihood of elevated CRP (OR = 1.37; 95% CI, 1.15-1.62). The same relationship was maintained in Non-Hispanic white, Non-Hispanic black, and other race (Non-Hispanic white: OR = 1.24; 95% CI, 1.03-1.48, Non-Hispanic black: OR = 1.60; 95% CI, 1.27-2.03, other race: OR = 2.50; 95% CI, 1.32-4.74) while the association was insignificant with American Mexican and other Hispanic. In the adjusted model, the relationship between CRP and prehypertension were no longer available. Contrary, hypertension was not independently associated with elevated CRP, and the results were the same after being grouped by race or adjustments for the possible confounder variables. The same results were obtained when SBP or DBP were on a continuous measure. Conclusions: This study confirmed the existence of an association between hypertension, prehypertension and elevated level of CRP, however this association was no longer available after adjusting by other variables. Ethic group differences were statistically significant at the univariable models, while it disappeared after controlling by other variables.Keywords: CRP, hypertension, ethnicity, NHANES, blood pressure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13601340 Reliability Levels of Reinforced Concrete Bridges Obtained by Mixing Approaches
Authors: Adrián D. García-Soto, Alejandro Hernández-Martínez, Jesús G. Valdés-Vázquez, Reyna A. Vizguerra-Alvarez
Abstract:
Reinforced concrete bridges designed by code are intended to achieve target reliability levels adequate for the geographical environment where the code is applicable. Several methods can be used to estimate such reliability levels. Many of them require the establishment of an explicit limit state function (LSF). When such LSF is not available as a close-form expression, the simulation techniques are often employed. The simulation methods are computing intensive and time consuming. Note that if the reliability of real bridges designed by code is of interest, numerical schemes, the finite element method (FEM) or computational mechanics could be required. In these cases, it can be quite difficult (or impossible) to establish a close-form of the LSF, and the simulation techniques may be necessary to compute reliability levels. To overcome the need for a large number of simulations when no explicit LSF is available, the point estimate method (PEM) could be considered as an alternative. It has the advantage that only the probabilistic moments of the random variables are required. However, in the PEM, fitting of the resulting moments of the LSF to a probability density function (PDF) is needed. In the present study, a very simple alternative which allows the assessment of the reliability levels when no explicit LSF is available and without the need of extensive simulations is employed. The alternative includes the use of the PEM, and its applicability is shown by assessing reliability levels of reinforced concrete bridges in Mexico when a numerical scheme is required. Comparisons with results by using the Monte Carlo simulation (MCS) technique are included. To overcome the problem of approximating the probabilistic moments from the PEM to a PDF, a well-known distribution is employed. The approach mixes the PEM and other classic reliability method (first order reliability method, FORM). The results in the present study are in good agreement whit those computed with the MCS. Therefore, the alternative of mixing the reliability methods is a very valuable option to determine reliability levels when no close form of the LSF is available, or if numerical schemes, the FEM or computational mechanics are employed.
Keywords: Structural reliability, reinforced concrete bridges, mixing approaches, point estimate method, Monte Carlo simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14121339 An Empirical Analysis of the Impact of Selected Macroeconomic Variables on Capital Formation in Libya (1970–2010)
Authors: Khaled Ramadan Elbeydi
Abstract:
This study is carried out to provide an insight into the analysis of the impact of selected macro-economic variables on gross fixed capital formation in Libya using annual data over the period (1970-2010). The importance of this study comes from the ability to show the relative important factors that impact the Libyan gross fixed capital formation. This understanding would give indications to decision makers on which policy they must focus to stimulate the economy. An Autoregressive Distributed Lag (ARDL) modeling process is employed to investigate the impact of the Gross Domestic Product, Monetary Base and Trade Openness on Gross Fixed Capital Formation in Libya. The results of this study reveal that there is an equilibrium relationship between capital formation and its determinants. The results also indicate that GDP and trade openness largely explain the pattern of capital formation in Libya. The findings and recommendations provide vital information relevant for policy formulation and implementation aimed to improve capital formation in Libya.
Keywords: ARDL, Bounds test, capital formation, Cointegration, Libya.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17281338 A Frame Work for the Development of a Suitable Method to Find Shoot Length at Maturity of Mustard Plant Using Soft Computing Model
Authors: Satyendra Nath Mandal, J. Pal Choudhury, Dilip De, S. R. Bhadra Chaudhuri
Abstract:
The production of a plant can be measured in terms of seeds. The generation of seeds plays a critical role in our social and daily life. The fruit production which generates seeds, depends on the various parameters of the plant, such as shoot length, leaf number, root length, root number, etc When the plant is growing, some leaves may be lost and some new leaves may appear. It is very difficult to use the number of leaves of the tree to calculate the growth of the plant.. It is also cumbersome to measure the number of roots and length of growth of root in several time instances continuously after certain initial period of time, because roots grow deeper and deeper under ground in course of time. On the contrary, the shoot length of the tree grows in course of time which can be measured in different time instances. So the growth of the plant can be measured using the data of shoot length which are measured at different time instances after plantation. The environmental parameters like temperature, rain fall, humidity and pollution are also play some role in production of yield. The soil, crop and distance management are taken care to produce maximum amount of yields of plant. The data of the growth of shoot length of some mustard plant at the initial stage (7,14,21 & 28 days after plantation) is available from the statistical survey by a group of scientists under the supervision of Prof. Dilip De. In this paper, initial shoot length of Ken( one type of mustard plant) has been used as an initial data. The statistical models, the methods of fuzzy logic and neural network have been tested on this mustard plant and based on error analysis (calculation of average error) that model with minimum error has been selected and can be used for the assessment of shoot length at maturity. Finally, all these methods have been tested with other type of mustard plants and the particular soft computing model with the minimum error of all types has been selected for calculating the predicted data of growth of shoot length. The shoot length at the stage of maturity of all types of mustard plants has been calculated using the statistical method on the predicted data of shoot length.Keywords: Fuzzy time series, neural network, forecasting error, average error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15911337 Analysis of the Diffusion Behavior of an Information and Communication Technology Platform for City Logistics
Authors: Giulio Mangano, Alberto De Marco, Giovanni Zenezini
Abstract:
The concept of City Logistics (CL) has emerged to improve the impacts of last mile freight distribution in urban areas. In this paper, a System Dynamics (SD) model exploring the dynamics of the diffusion of a ICT platform for CL management across different populations is proposed. For the development of the model two sources have been used. On the one hand, the major diffusion variables and feedback loops are derived from a literature review of existing diffusion models. On the other hand, the parameters are represented by the value propositions delivered by the platform as a response to some of the users’ needs. To extract the most important value propositions the Business Model Canvas approach has been used. Such approach in fact focuses on understanding how a company can create value for her target customers. These variables and parameters are thus translated into a SD diffusion model with three different populations namely municipalities, logistics service providers, and own account carriers. Results show that, the three populations under analysis fully adopt the platform within the simulation time frame, highlighting a strong demand by different stakeholders for CL projects aiming at carrying out more efficient urban logistics operations.
Keywords: City logistics, simulation, system dynamics, business model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10291336 Binary Decision Diagrams: An Improved Variable Ordering using Graph Representation of Boolean Functions
Authors: P.W. C. Prasad, A. Assi, A. Harb, V.C. Prasad
Abstract:
This paper presents an improved variable ordering method to obtain the minimum number of nodes in Reduced Ordered Binary Decision Diagrams (ROBDD). The proposed method uses the graph topology to find the best variable ordering. Therefore the input Boolean function is converted to a unidirectional graph. Three levels of graph parameters are used to increase the probability of having a good variable ordering. The initial level uses the total number of nodes (NN) in all the paths, the total number of paths (NP) and the maximum number of nodes among all paths (MNNAP). The second and third levels use two extra parameters: The shortest path among two variables (SP) and the sum of shortest path from one variable to all the other variables (SSP). A permutation of the graph parameters is performed at each level for each variable order and the number of nodes is recorded. Experimental results are promising; the proposed method is found to be more effective in finding the variable ordering for the majority of benchmark circuits.
Keywords: Binary decision diagrams, graph representation, Boolean functions representation, variable ordering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21171335 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm
Authors: Ameur Abdelkader, Abed Bouarfa Hafida
Abstract:
Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.
Keywords: Predictive analysis, big data, predictive analysis algorithms. CART algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10751334 Application of Rapidly Exploring Random Tree Star-Smart and G2 Quintic Pythagorean Hodograph Curves to the UAV Path Planning Problem
Authors: Luiz G. Véras, Felipe L. Medeiros, Lamartine F. Guimarães
Abstract:
This work approaches the automatic planning of paths for Unmanned Aerial Vehicles (UAVs) through the application of the Rapidly Exploring Random Tree Star-Smart (RRT*-Smart) algorithm. RRT*-Smart is a sampling process of positions of a navigation environment through a tree-type graph. The algorithm consists of randomly expanding a tree from an initial position (root node) until one of its branches reaches the final position of the path to be planned. The algorithm ensures the planning of the shortest path, considering the number of iterations tending to infinity. When a new node is inserted into the tree, each neighbor node of the new node is connected to it, if and only if the extension of the path between the root node and that neighbor node, with this new connection, is less than the current extension of the path between those two nodes. RRT*-smart uses an intelligent sampling strategy to plan less extensive routes by spending a smaller number of iterations. This strategy is based on the creation of samples/nodes near to the convex vertices of the navigation environment obstacles. The planned paths are smoothed through the application of the method called quintic pythagorean hodograph curves. The smoothing process converts a route into a dynamically-viable one based on the kinematic constraints of the vehicle. This smoothing method models the hodograph components of a curve with polynomials that obey the Pythagorean Theorem. Its advantage is that the obtained structure allows computation of the curve length in an exact way, without the need for quadratural techniques for the resolution of integrals.Keywords: Path planning, path smoothing, Pythagorean hodograph curve, RRT*-Smart.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8981333 A Study of Current Maintenance Strategies and the Reliability of Critical Medical Equipment in Hospitals in Relation to Patient Outcomes
Authors: Khelood A. Mkalaf, Peter Gibson, John Flanagan
Abstract:
This study investigates the relationship between the reliability of critical medical equipment (CME) and the effectiveness of CME maintenance management strategies in relation to patient outcomes in 84 public hospitals of a top 20 OECD country. The work has examined the effectiveness of CME maintenance management strategies used by the public hospital system of a large state run health organization. The conceptual framework was designed to examine the significance of the relationship between six variables: (1) types of maintenance management strategies, (2) maintenance services, (3) maintenance practice, (4) medical equipment reliability, (5) maintenance costs and (6) patient outcomes. The results provide interesting insights into the effectiveness of the maintenance strategies used. For example, there appears to be about a 1 in 10 000 probability of failure of anesthesia equipment, but these seem to be confined to specific maintenance situations. There are also some findings in relation to outsourcing of maintenance. For each of the variables listed, results are reported in relation to the various types of maintenance strategies and services. Decision-makers may use these results to evaluate more effective maintenance strategies for their CME and generate more effective patient outcomes.
Keywords: Critical medical equipment, maintenance strategy, patient outcomes, reliability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49151332 Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method
Authors: Prashant P. Bedekar, Sudhir R. Bhide, Vijay S. Kale
Abstract:
Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.Keywords: Constrained optimization, LPP, Overcurrent relaycoordination, Two-phase simplex method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30071331 Machine Learning Techniques for COVID-19 Detection: A Comparative Analysis
Authors: Abeer Aljohani
Abstract:
The COVID-19 virus spread has been one of the extreme pandemics across the globe. It is also referred as corona virus which is a contagious disease that continuously mutates into numerous variants. Currently, the B.1.1.529 variant labeled as Omicron is detected in South Africa. The huge spread of COVID-19 disease has affected several lives and has surged exceptional pressure on the healthcare systems worldwide. Also, everyday life and the global economy have been at stake. Numerous COVID-19 cases have produced a huge burden on hospitals as well as health workers. To reduce this burden, this paper predicts COVID-19 disease based on the symptoms and medical history of the patient. As machine learning is a widely accepted area and gives promising results for healthcare, this research presents an architecture for COVID-19 detection using ML techniques integrated with feature dimensionality reduction. This paper uses a standard University of California Irvine (UCI) dataset for predicting COVID-19 disease. This dataset comprises symptoms of 5434 patients. This paper also compares several supervised ML techniques on the presented architecture. The architecture has also utilized 10-fold cross validation process for generalization and Principal Component Analysis (PCA) technique for feature reduction. Standard parameters are used to evaluate the proposed architecture including F1-Score, precision, accuracy, recall, Receiver Operating Characteristic (ROC) and Area under Curve (AUC). The results depict that Decision tree, Random Forest and neural networks outperform all other state-of-the-art ML techniques. This result can be used to effectively identify COVID-19 infection cases.
Keywords: Supervised machine learning, COVID-19 prediction, healthcare analytics, Random Forest, Neural Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3841330 ALD HfO2 Based RRAM with Ti Capping
Authors: B. B. Weng, Z. Fang, Z. X. Chen, X. P. Wang, G. Q. Lo, D. L. Kwong
Abstract:
HfOx based Resistive Random Access Memory (RRAM) is one of the most widely studied material stack due to its promising performances as an emerging memory technology. In this work, we systematically investigated the effect of metal capping layer by preparing sample devices with varying thickness of Ti cap and comparing their operating parameters with the help of an Agilent-B1500A analyzer.
Keywords: HfOx, resistive switching, RRAM, metal capping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20401329 A Model of Market Segmentation for the Customers of Mellat Bank in Iran
Authors: Nader Gharibnavaz, Hossein Yazdi
Abstract:
If organizations like Mellat Bank want to identify its customer market completely to reach its specified goals, it can segment the market to offer the product package to the right segment. Our objective is to offer a segmentation model for Iran banking market in Mellat bank view. The methodology of this project is combined by “segmentation on the basis of four part-quality variables" and “segmentation on the basis of different in means". Required data are gathered from E-Systems and researcher personal observation. Finally, the research offers the organization that at first step form a four dimensional matrix with 756 segments using four variables named value-based, behavioral, activity style, and activity level, and at the second step calculate the means of profit for every cell of matrix in two distinguished work level (levels α1:normal condition and α2: high pressure condition) and compare the segments by checking two conditions that are 1- homogeneity every segment with its sub segment and 2- heterogeneity with other segments, and so it can do the necessary segmentation process. After all, the last offer (more explained by an operational example and feedback algorithm) is to test and update the model because of dynamic environment, technology, and banking system.Keywords: market segmentation model, banking system, Mellat bank
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32871328 Non-parametric Linear Technique for Measuring the Efficiency of Winter Road Maintenance in the Arctic Area
Authors: Mahshid Hatamzad, Geanette Polanco
Abstract:
Improving the performance of Winter Road Maintenance (WRM) can increase the traffic safety and reduce the cost as well as environmental impacts. This study evaluates the efficiency of WRM technique, named salting, in the Arctic area by using Data Envelopment Analysis (DEA), which is a non-parametric linear method to measure the efficiencies of decision-making units (DMUs) based on handling multiple inputs and multiple outputs at the same time that their associated weights are not known. Here, roads are considered as DMUs for which the efficiency must be determined. The three input variables considered are traffic flow, road area and WRM cost. In addition, the two output variables included are level of safety in the roads and environment impacts resulted from WRM, which is also considered as an uncontrollable factor in the second scenario. The results show the performance of DMUs from the most efficient WRM to the inefficient/least efficient one and this information provides decision makers with technical support and the required suggested improvements for inefficient WRM, in order to achieve a cost-effective WRM and a safe road transportation during wintertime in the Arctic areas.
Keywords: DEA, environmental impacts, risk and safety, WRM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 554