Search results for: total error rate
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17082

Search results for: total error rate

16752 Seasonal Influence on Environmental Indicators of Beach Waste

Authors: Marcus C. Garcia, Giselle C. Guimarães, Luciana H. Yamane, Renato R. Siman

Abstract:

The environmental indicators and the classification of beach waste are essential tools to diagnose the current situation and to indicate ways to improve the quality of this environment. The purpose of this paper was to perform a quali-quantitative analysis of the beach waste on the Curva da Jurema Beach (Espírito Santo - Brazil). Three transects were used with equidistant positioning over the total length of the beach for the solid waste collection. Solid wastes were later classified according to their use and primary raw material from the low and high summer season. During the low season, average values of 7.10 items.m-1, 18.22 g.m-1 and 0.91 g.m-2 were found for the whole beach, and transect 3 contributed the most waste, with the total sum of items equal to 999 (49%), a total mass of 5.62 kg and a total volume of 21.31 L. During the high summer season, average values of 8.22 items.m-1, 54.40 g.m-1 and 2.72 g.m-2 were found, with transect 2 contributing the most to the total sum with 1,212 items (53%), a total mass of 10.76 kg and a total volume of 51.99 L. Of the total collected, plastic materials represented 51.4% of the total number of items, 35.9% of the total mass and 68% of the total volume. The implementation of reactive and proactive measures is necessary so that the management of the solid wastes on Curva da Jurema Beach is in accordance with principles of sustainability.

Keywords: beach solid waste, environmental indicators, quali-quantitative analysis, waste management

Procedia PDF Downloads 306
16751 Effects of Palm Waste Ash Residues on Acidic Soil in Relation to Physiological Responses of Habanero Chili Pepper (Capsicum chinense jacq.)

Authors: Kalu Samuel Ukanwa, Kumar Patchigolla, Ruben Sakrabani

Abstract:

The use of biosolids from thermal conversion of palm waste for soil fertility enhancement was tested in acidic soil of Southern Nigeria for the growing of Habanero chili pepper (Capsicum chinense jacq.). Soil samples from the two sites, showed pH 4.8 and 4.8 for site A and B respectively, below 5.6-6.8 optimum range and other fertility parameters indicating a low threshold for pepper growth. Nursery planting was done at different weeks to determine the optimum planting period. Ash analysis showed that it contains 26% of total K, 20% of total Ca, 0.27% of total P, and pH 11. The two sites were laid for an experiment in randomized complete block design and setup with three replications side by side. Each plot measured 3 x 2 m and a total of 15 plots for each site, four treatments, and one control. Outlined as control, 2, 4, 6 and 8 tonnes/hectare of palm waste ash, the combined average for both sites with correspondent yield after six harvests in one season are; 0, 5.8, 6, 6, 14.5 tonnes/hectare respectively to treatments. Optimum nursery survival rate was high in July; the crop yield was linear to the ash application. Site A had 6% yield higher than site B. Fruit development, weight, and total yield in relation to the control plot showed that palm waste ash is effective for soil amendment, nutrient delivery, and exchange.

Keywords: ash, palm waste, pepper, soil amendment

Procedia PDF Downloads 133
16750 On Differential Growth Equation to Stochastic Growth Model Using Hyperbolic Sine Function in Height/Diameter Modeling of Pines

Authors: S. O. Oyamakin, A. U. Chukwu

Abstract:

Richard's growth equation being a generalized logistic growth equation was improved upon by introducing an allometric parameter using the hyperbolic sine function. The integral solution to this was called hyperbolic Richard's growth model having transformed the solution from deterministic to a stochastic growth model. Its ability in model prediction was compared with the classical Richard's growth model an approach which mimicked the natural variability of heights/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using the coefficient of determination (R2), Mean Absolute Error (MAE) and Mean Square Error (MSE) results. The Kolmogorov-Smirnov test and Shapiro-Wilk test was also used to test the behavior of the error term for possible violations. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic Richard's nonlinear growth models better than the classical Richard's growth model.

Keywords: height, Dbh, forest, Pinus caribaea, hyperbolic, Richard's, stochastic

Procedia PDF Downloads 480
16749 Comparisons of Individual and Group Replacement Policies for a Series Connection System with Two Machines

Authors: Wen Liang Chang, Mei Wei Wang, Ruey Huei Yeh

Abstract:

This paper studies the comparisons of individual and group replacement policies for a series connection system with two machines. Suppose that manufacturer’s production system is a series connection system which is combined by two machines. For two machines, when machines fail within the operating time, minimal repair is performed for machines by the manufacturer. The manufacturer plans to a preventive replacement for machines at a pre-specified time to maintain system normal operation. Under these maintenance policies, the maintenance cost rate models of individual and group replacement for a series connection system with two machines is derived and further, optimal preventive replacement time is obtained such that the expected total maintenance cost rate is minimized. Finally, some numerical examples are given to illustrate the influences of individual and group replacement policies to the maintenance cost rate.

Keywords: individual replacement, group replacement, replacement time, two machines, series connection system

Procedia PDF Downloads 488
16748 A Comparative Evaluation of the SIR and SEIZ Epidemiological Models to Describe the Diffusion Characteristics of COVID-19 Polarizing Viewpoints on Online

Authors: Maryam Maleki, Esther Mead, Mohammad Arani, Nitin Agarwal

Abstract:

This study is conducted to examine how opposing viewpoints related to COVID-19 were diffused on Twitter. To accomplish this, six datasets using two epidemiological models, SIR (Susceptible, Infected, Recovered) and SEIZ (Susceptible, Exposed, Infected, Skeptics), were analyzed. The six datasets were chosen because they represent opposing viewpoints on the COVID-19 pandemic. Three of the datasets contain anti-subject hashtags, while the other three contain pro-subject hashtags. The time frame for all datasets is three years, starting from January 2020 to December 2022. The findings revealed that while both models were effective in evaluating the propagation trends of these polarizing viewpoints, the SEIZ model was more accurate with a relatively lower error rate (6.7%) compared to the SIR model (17.3%). Additionally, the relative error for both models was lower for anti-subject hashtags compared to pro-subject hashtags. By leveraging epidemiological models, insights into the propagation trends of polarizing viewpoints on Twitter were gained. This study paves the way for the development of methods to prevent the spread of ideas that lack scientific evidence while promoting the dissemination of scientifically backed ideas.

Keywords: mathematical modeling, epidemiological model, seiz model, sir model, covid-19, twitter, social network analysis, social contagion

Procedia PDF Downloads 62
16747 Reproduction Characteristics of Saanen Goats Raised under Intensive Conditions in Konya Province

Authors: Vahdettin Sariyel, Birol Dag

Abstract:

In this research, it is aimed to determine the effects of several environmental factors on adaptation and some yield parameters of Saanen goats reared under intensive conditions at a private farm in Konya province. Gestation rate, twins rate and litter size were evaluated as reproductive traits. Gestation rate was determined as 93.8% and 90.5% for 2011 and 2012 years respectively. Twins rate was determined as 59.35 % and 70.00 % for 2011 and 2012 years respectively. Litter size was 1.49 and 1.46 for 2011 and 2012 years respectively. Survival rates of kids from birth to weaning at three months of age were found as 87.74 % and 98.54 % for 2011 and 2012 years respectively.

Keywords: gestation rate, reproduction, saanen, twins rate, vitality

Procedia PDF Downloads 590
16746 Sardine Oil as a Source of Lipid in the Diet of Giant Freshwater Prawn (Macrobrachium rosenbergii)

Authors: A. T. Ramachandra Naik, H. Shivananda Murthy, H. n. Anjanayappa

Abstract:

The freshwater prawn, Macrobrachium rosenbergii is a more popular crustacean cultured widely in monoculture system in India. It has got high nutritional value in the human diet. Hence, understanding its enzymatic and body composition is important in order to judge its flesh quality. Fish oil specially derived from Indian oil sardine is a good source of highly unsaturated fatty acid and lipid source in fish/prawn diet. A 35% crude protein diet with graded levels of Sardine oil as a source of fat was incorporated at four levels viz, 2.07, 4.07, 6.07 and 8.07% maintaining a total lipid level of feed at 8.11, 10.24, 12.28 and 14.33% respectively. Diet without sardine oil (6.05% total lipid) was served as basal treatment. The giant freshwater prawn, Macrobrachium rosenbergii was used as test animal and the experiment was lost for 112 days. Significantly, higher gain in weight of prawn was recorded in the treatment with 6.07% sardine oil incorporation followed by higher specific growth rate, food conversion rate and protein efficiency ratio. The 8.07% sardine oil diet produced the highest RNA: DNA ratio in the prawn muscle. Digestive enzyme analyses in the digestive tract and mid-gut gland showed the greatest activity in prawns fed the 8.07% diet.

Keywords: digestive enzyme, fish diet, Macrobrachium rosenbergii, sardine oil

Procedia PDF Downloads 329
16745 Hardware Error Analysis and Severity Characterization in Linux-Based Server Systems

Authors: Nikolaos Georgoulopoulos, Alkis Hatzopoulos, Konstantinos Karamitsios, Konstantinos Kotrotsios, Alexandros I. Metsai

Abstract:

In modern server systems, business critical applications run in different types of infrastructure, such as cloud systems, physical machines and virtualization. Often, due to high load and over time, various hardware faults occur in servers that translate to errors, resulting to malfunction or even server breakdown. CPU, RAM and hard drive (HDD) are the hardware parts that concern server administrators the most regarding errors. In this work, selected RAM, HDD and CPU errors, that have been observed or can be simulated in kernel ring buffer log files from two groups of Linux servers, are investigated. Moreover, a severity characterization is given for each error type. Better understanding of such errors can lead to more efficient analysis of kernel logs that are usually exploited for fault diagnosis and prediction. In addition, this work summarizes ways of simulating hardware errors in RAM and HDD, in order to test the error detection and correction mechanisms of a Linux server.

Keywords: hardware errors, Kernel logs, Linux servers, RAM, hard disk, CPU

Procedia PDF Downloads 154
16744 GPU Based High Speed Error Protection for Watermarked Medical Image Transmission

Authors: Md Shohidul Islam, Jongmyon Kim, Ui-pil Chong

Abstract:

Medical image is an integral part of e-health care and e-diagnosis system. Medical image watermarking is widely used to protect patients’ information from malicious alteration and manipulation. The watermarked medical images are transmitted over the internet among patients, primary and referred physicians. The images are highly prone to corruption in the wireless transmission medium due to various noises, deflection, and refractions. Distortion in the received images leads to faulty watermark detection and inappropriate disease diagnosis. To address the issue, this paper utilizes error correction code (ECC) with (8, 4) Hamming code in an existing watermarking system. In addition, we implement the high complex ECC on a graphics processing units (GPU) to accelerate and support real-time requirement. Experimental results show that GPU achieves considerable speedup over the sequential CPU implementation, while maintaining 100% ECC efficiency.

Keywords: medical image watermarking, e-health system, error correction, Hamming code, GPU

Procedia PDF Downloads 290
16743 An Alternative Richards’ Growth Model Based on Hyperbolic Sine Function

Authors: Samuel Oluwafemi Oyamakin, Angela Unna Chukwu

Abstract:

Richrads growth equation being a generalized logistic growth equation was improved upon by introducing an allometric parameter using the hyperbolic sine function. The integral solution to this was called hyperbolic Richards growth model having transformed the solution from deterministic to a stochastic growth model. Its ability in model prediction was compared with the classical Richards growth model an approach which mimicked the natural variability of heights/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using the coefficient of determination (R2), Mean Absolute Error (MAE) and Mean Square Error (MSE) results. The Kolmogorov-Smirnov test and Shapiro-Wilk test was also used to test the behavior of the error term for possible violations. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic Richards nonlinear growth models better than the classical Richards growth model.

Keywords: height, diameter at breast height, DBH, hyperbolic sine function, Pinus caribaea, Richards' growth model

Procedia PDF Downloads 392
16742 The Effect of Macroeconomic Policies on Cambodia's Economy: ARDL and VECM Model

Authors: Siphat Lim

Abstract:

This study used Autoregressive Distributed Lag (ARDL) approach to cointegration. In the long-run the general price level and exchange rate have a positively significant effect on domestic output. The estimated result further revealed that fiscal stimulus help stimulate domestic output in the long-run, but not in the short-run, while monetary expansion help to stimulate output in both short-run and long-run. The result is complied with the theory which is the macroeconomic policies, fiscal and monetary policy; help to stimulate domestic output in the long-run. The estimated result of the Vector Error Correction Model (VECM) has indicated more clearly that the consumer price index has a positive effect on output with highly statistically significant. Increasing in the general price level would increase the competitiveness among producers than increase in the output. However, the exchange rate also has a positive effect and highly significant on the gross domestic product. The exchange rate depreciation might increase export since the purchasing power of foreigners has increased. More importantly, fiscal stimulus would help stimulate the domestic output in the long-run since the coefficient of government expenditure is positive. In addition, monetary expansion would also help stimulate the output and the result is highly significant. Thus, fiscal stimulus and monetary expansionary would help stimulate the domestic output in the long-run in Cambodia.

Keywords: fiscal policy, monetary policy, ARDL, VECM

Procedia PDF Downloads 431
16741 Tax Evasion and Macroeconomic (In)stability

Authors: Wei-Neng Wang, Jhy-Yuan Shieh, Jhy-Hwa Chen, Juin-Jen Chang

Abstract:

This paper incorporate tax evasion into a one-sector real business cycle (RBC) model to explores the quantitative interrelations between income tax rate and equilibrium (in)determinacy, and income tax rate is endogenously determined in order to balance the government budget. We find that the level of the effective income tax rate is key factor for equilibrium (in)determinacy, instead of the level of income tax rate in a tax evasion economy. Under an economy with tax evasion, the higher income tax rate is not sufficiently to lead to equilibrium indeterminate, it must combine with a necessary condition which is the lower fraction of tax evasion and that can result in agents' optimistic expectations to become self-fulfilling and sunspot fluctuation more likely to occur. On the other hand, an economy with tax evasion can see its macroeconomy become more stabilize, and a higher fraction of income tax evasion may has a stronger stabilizing effect.

Keywords: tax evasion, balanced-budget rule, equlibirium (in)determinacy, effective income tax rate

Procedia PDF Downloads 63
16740 The Maximum Throughput Analysis of UAV Datalink 802.11b Protocol

Authors: Inkyu Kim, SangMan Moon

Abstract:

This IEEE 802.11b protocol provides up to 11Mbps data rate, whereas aerospace industry wants to seek higher data rate COTS data link system in the UAV. The Total Maximum Throughput (TMT) and delay time are studied on many researchers in the past years This paper provides theoretical data throughput performance of UAV formation flight data link using the existing 802.11b performance theory. We operate the UAV formation flight with more than 30 quad copters with 802.11b protocol. We may be predicting that UAV formation flight numbers have to bound data link protocol performance limitations.

Keywords: UAV datalink, UAV formation flight datalink, UAV WLAN datalink application, UAV IEEE 802.11b datalink application

Procedia PDF Downloads 392
16739 Total Lipid of Mutant Synechococcus sp. PCC 7002

Authors: Azlin S Azmi, Mus’ab Zainal, Sarina Sulaiman, Azura Amid, Zaki Zainudin

Abstract:

Microalgae lipid is a promising feedstock for biodiesel production. The objective of this work was to study growth factors affecting marine mutant Synechococcus sp. (PCC 7002) for high lipid production. Four growth factors were investigated; nitrogen-phosporus-potassium (NPK) concentration, light intensity, temperature and NaNO3 concentration on mutant strain growth and lipid production were studied. Design Expert v8.0 was used to design the experimental and analyze the data. The experimental design selected was Min-Run Res IV which consists of 12 runs and the response surfaces measured were specific growth rate and lipid concentration. The extraction of lipid was conducted by chloroform/methanol solvents system. Based on the study, mutant Synechococcus sp. PCC 7002 gave the highest specific growth rate of 0.0014 h-1 at 0% NPK, 2500 lux, 40oC and 0% NaNO3. On the other hand, the highest lipid concentration was obtained at 0% NPK, 3500 lux, 30°C and 1% NaNO3.

Keywords: Cyanobacteria, lipid, mutant, marine Synechococcus sp. (PCC 7002), specific growth rate

Procedia PDF Downloads 337
16738 Developing an ANN Model to Predict Anthropometric Dimensions Based on Real Anthropometric Database

Authors: Waleed A. Basuliman, Khalid S. AlSaleh, Mohamed Z. Ramadan

Abstract:

Applying the anthropometric dimensions is considered one of the important factors when designing any human-machine system. In this study, the estimation of anthropometric dimensions has been improved by developing artificial neural network that aims to predict the anthropometric measurements of the male in Saudi Arabia. A total of 1427 Saudi males from age 6 to 60 participated in measuring twenty anthropometric dimensions. These anthropometric measurements are important for designing the majority of work and life applications in Saudi Arabia. The data were collected during 8 months from different locations in Riyadh City. Five of these dimensions were used as predictors variables (inputs) of the model, and the remaining fifteen dimensions were set to be the measured variables (outcomes). The hidden layers have been varied during the structuring stage, and the best performance was achieved with the network structure 6-25-15. The results showed that the developed Neural Network model was significantly able to predict the body dimensions for the population of Saudi Arabia. The network mean absolute percentage error (MAPE) and the root mean squared error (RMSE) were found 0.0348 and 3.225 respectively. The accuracy of the developed neural network was evaluated by compare the predicted outcomes with a multiple regression model. The ANN model performed better and resulted excellent correlation coefficients between the predicted and actual dimensions.

Keywords: artificial neural network, anthropometric measurements, backpropagation, real anthropometric database

Procedia PDF Downloads 576
16737 Comparison of the Effectiveness of Tree Algorithms in Classification of Spongy Tissue Texture

Authors: Roza Dzierzak, Waldemar Wojcik, Piotr Kacejko

Abstract:

Analysis of the texture of medical images consists of determining the parameters and characteristics of the examined tissue. The main goal is to assign the analyzed area to one of two basic groups: as a healthy tissue or a tissue with pathological changes. The CT images of the thoracic lumbar spine from 15 healthy patients and 15 with confirmed osteoporosis were used for the analysis. As a result, 120 samples with dimensions of 50x50 pixels were obtained. The set of features has been obtained based on the histogram, gradient, run-length matrix, co-occurrence matrix, autoregressive model, and Haar wavelet. As a result of the image analysis, 290 descriptors of textural features were obtained. The dimension of the space of features was reduced by the use of three selection methods: Fisher coefficient (FC), mutual information (MI), minimization of the classification error probability and average correlation coefficients between the chosen features minimization of classification error probability (POE) and average correlation coefficients (ACC). Each of them returned ten features occupying the initial place in the ranking devised according to its own coefficient. As a result of the Fisher coefficient and mutual information selections, the same features arranged in a different order were obtained. In both rankings, the 50% percentile (Perc.50%) was found in the first place. The next selected features come from the co-occurrence matrix. The sets of features selected in the selection process were evaluated using six classification tree methods. These were: decision stump (DS), Hoeffding tree (HT), logistic model trees (LMT), random forest (RF), random tree (RT) and reduced error pruning tree (REPT). In order to assess the accuracy of classifiers, the following parameters were used: overall classification accuracy (ACC), true positive rate (TPR, classification sensitivity), true negative rate (TNR, classification specificity), positive predictive value (PPV) and negative predictive value (NPV). Taking into account the classification results, it should be stated that the best results were obtained for the Hoeffding tree and logistic model trees classifiers, using the set of features selected by the POE + ACC method. In the case of the Hoeffding tree classifier, the highest values of three parameters were obtained: ACC = 90%, TPR = 93.3% and PPV = 93.3%. Additionally, the values of the other two parameters, i.e., TNR = 86.7% and NPV = 86.6% were close to the maximum values obtained for the LMT classifier. In the case of logistic model trees classifier, the same ACC value was obtained ACC=90% and the highest values for TNR=88.3% and NPV= 88.3%. The values of the other two parameters remained at a level close to the highest TPR = 91.7% and PPV = 91.6%. The results obtained in the experiment show that the use of classification trees is an effective method of classification of texture features. This allows identifying the conditions of the spongy tissue for healthy cases and those with the porosis.

Keywords: classification, feature selection, texture analysis, tree algorithms

Procedia PDF Downloads 178
16736 Evaluation of Dual Polarization Rainfall Estimation Algorithm Applicability in Korea: A Case Study on Biseulsan Radar

Authors: Chulsang Yoo, Gildo Kim

Abstract:

Dual polarization radar provides comprehensive information about rainfall by measuring multiple parameters. In Korea, for the rainfall estimation, JPOLE and CSU-HIDRO algorithms are generally used. This study evaluated the local applicability of JPOLE and CSU-HIDRO algorithms in Korea by using the observed rainfall data collected on August, 2014 by the Biseulsan dual polarization radar data and KMA AWS. A total of 11,372 pairs of radar-ground rain rate data were classified according to thresholds of synthetic algorithms into suitable and unsuitable data. Then, evaluation criteria were derived by comparing radar rain rate and ground rain rate, respectively, for entire, suitable, unsuitable data. The results are as follows: (1) The radar rain rate equation including KDP, was found better in the rainfall estimation than the other equations for both JPOLE and CSU-HIDRO algorithms. The thresholds were found to be adequately applied for both algorithms including specific differential phase. (2) The radar rain rate equation including horizontal reflectivity and differential reflectivity were found poor compared to the others. The result was not improved even when only the suitable data were applied. Acknowledgments: This work was supported by the Basic Science Research Program through the National Research Foundation of Korea, funded by the Ministry of Education (NRF-2013R1A1A2011012).

Keywords: CSU-HIDRO algorithm, dual polarization radar, JPOLE algorithm, radar rainfall estimation algorithm

Procedia PDF Downloads 213
16735 An Adaptive Oversampling Technique for Imbalanced Datasets

Authors: Shaukat Ali Shahee, Usha Ananthakumar

Abstract:

A data set exhibits class imbalance problem when one class has very few examples compared to the other class, and this is also referred to as between class imbalance. The traditional classifiers fail to classify the minority class examples correctly due to its bias towards the majority class. Apart from between-class imbalance, imbalance within classes where classes are composed of a different number of sub-clusters with these sub-clusters containing different number of examples also deteriorates the performance of the classifier. Previously, many methods have been proposed for handling imbalanced dataset problem. These methods can be classified into four categories: data preprocessing, algorithmic based, cost-based methods and ensemble of classifier. Data preprocessing techniques have shown great potential as they attempt to improve data distribution rather than the classifier. Data preprocessing technique handles class imbalance either by increasing the minority class examples or by decreasing the majority class examples. Decreasing the majority class examples lead to loss of information and also when minority class has an absolute rarity, removing the majority class examples is generally not recommended. Existing methods available for handling class imbalance do not address both between-class imbalance and within-class imbalance simultaneously. In this paper, we propose a method that handles between class imbalance and within class imbalance simultaneously for binary classification problem. Removing between class imbalance and within class imbalance simultaneously eliminates the biases of the classifier towards bigger sub-clusters by minimizing the error domination of bigger sub-clusters in total error. The proposed method uses model-based clustering to find the presence of sub-clusters or sub-concepts in the dataset. The number of examples oversampled among the sub-clusters is determined based on the complexity of sub-clusters. The method also takes into consideration the scatter of the data in the feature space and also adaptively copes up with unseen test data using Lowner-John ellipsoid for increasing the accuracy of the classifier. In this study, neural network is being used as this is one such classifier where the total error is minimized and removing the between-class imbalance and within class imbalance simultaneously help the classifier in giving equal weight to all the sub-clusters irrespective of the classes. The proposed method is validated on 9 publicly available data sets and compared with three existing oversampling techniques that rely on the spatial location of minority class examples in the euclidean feature space. The experimental results show the proposed method to be statistically significantly superior to other methods in terms of various accuracy measures. Thus the proposed method can serve as a good alternative to handle various problem domains like credit scoring, customer churn prediction, financial distress, etc., that typically involve imbalanced data sets.

Keywords: classification, imbalanced dataset, Lowner-John ellipsoid, model based clustering, oversampling

Procedia PDF Downloads 418
16734 Computer Simulation Approach in the 3D Printing Operations of Surimi Paste

Authors: Timilehin Martins Oyinloye, Won Byong Yoon

Abstract:

Simulation technology is being adopted in many industries, with research focusing on the development of new ways in which technology becomes embedded within production, services, and society in general. 3D printing (3DP) technology is fast developing in the food industry. However, the limited processability of high-performance material restricts the robustness of the process in some cases. Significantly, the printability of materials becomes the foundation for extrusion-based 3DP, with residual stress being a major challenge in the printing of complex geometry. In many situations, the trial-a-error method is being used to determine the optimum printing condition, which results in time and resource wastage. In this report, the analysis of 3 moisture levels for surimi paste was investigated for an optimum 3DP material and printing conditions by probing its rheology, flow characteristics in the nozzle, and post-deposition process using the finite element method (FEM) model. Rheological tests revealed that surimi pastes with 82% moisture are suitable for 3DP. According to the FEM model, decreasing the nozzle diameter from 1.2 mm to 0.6 mm, increased the die swell from 9.8% to 14.1%. The die swell ratio increased due to an increase in the pressure gradient (1.15107 Pa to 7.80107 Pa) at the nozzle exit. The nozzle diameter influenced the fluid properties, i.e., the shear rate, velocity, and pressure in the flow field, as well as the residual stress and the deformation of the printed sample, according to FEM simulation. The post-printing stability of the model was investigated using the additive layer manufacturing (ALM) model. The ALM simulation revealed that the residual stress and total deformation of the sample were dependent on the nozzle diameter. A small nozzle diameter (0.6 mm) resulted in a greater total deformation (0.023), particularly at the top part of the model, which eventually resulted in the sample collapsing. As the nozzle diameter increased, the accuracy of the model improved until the optimum nozzle size (1.0 mm). Validation with 3D-printed surimi products confirmed that the nozzle diameter was a key parameter affecting the geometry accuracy of 3DP of surimi paste.

Keywords: 3D printing, deformation analysis, die swell, numerical simulation, surimi paste

Procedia PDF Downloads 68
16733 Rate of Profit as a Pricing Benchmark in Islamic Banking to Create Financial Stability

Authors: Trisiladi Supriyanto

Abstract:

Although much research has been done on the pricing benchmark both in terms of fiqh or Islamic economic perspective, but no substitution for the concept of interest (rate of interest) up to now in the application of Islamic Banking because some of the jurists from the middle east even allow the use of a benchmark rate such as LIBOR (London Interbank Offered Rate) as a measure of Islamic financial asset prices, so in other words, they equate the concept of rate of interest with the concept of rate of profit, which is the core reason (raison detre) for the replacement of usury as instructed in the Quran. This study aims to find the concept of rate of profit on Islamic banking that can create economic justice and stability in Islamic Banking and Capital market. Rate of profit that creates economic justice and stability can be achieved through its role in maintaining the stability of the financial system in which there is an equitable distribution of income and wealth. To determine the role of the rate of profit as the basis of the sharing system implemented in the Islamic financial system, we can see the connection of rate of profit in creating financial stability, especially in the asset-liability management of financial institutions that generate a stable net margin or the rate of profit that is not affected by the ups and downs of the market risk factors including indirect effect on interest rates. Furthermore, Islamic financial stability can be seen from the role of the rate of profit on the stability of the Islamic financial assets that are measured from the Islamic financial asset price volatility in Islamic Bond Market in Capital Market.

Keywords: Rate of profit, economic justice, stability, equitable distribution of income, equitable distribution of wealth

Procedia PDF Downloads 403
16732 0.13-µm Complementary Metal-Oxide Semiconductor Vector Modulator for Beamforming System

Authors: J. S. Kim

Abstract:

This paper presents a 0.13-µm Complementary Metal-Oxide Semiconductor (CMOS) vector modulator for beamforming system. The vector modulator features a 360° phase and gain range of -10 dB to 10 dB with a root mean square phase and amplitude error of only 2.2° and 0.45 dB, respectively. These features make it a suitable for wireless backhaul system in the 5 GHz industrial, scientific, and medical (ISM) bands. It draws a current of 20.4 mA from a 1.2 V supply. The total chip size is 1.87x1.34 mm².

Keywords: CMOS, vector modulator, beamforming, 802.11ac

Procedia PDF Downloads 210
16731 On the Performance Analysis of Coexistence between IEEE 802.11g and IEEE 802.15.4 Networks

Authors: Chompunut Jantarasorn, Chutima Prommak

Abstract:

This paper presents an intensive measurement studying of the network performance analysis when IEEE 802.11g Wireless Local Area Networks (WLAN) coexisting with IEEE 802.15.4 Wireless Personal Area Network (WPAN). The measurement results show that the coexistence between both networks could increase the Frame Error Rate (FER) of the IEEE 802.15.4 networks up to 60% and it could decrease the throughputs of the IEEE 802.11g networks up to 55%.

Keywords: wireless performance analysis, coexistence analysis, IEEE 802.11g, IEEE 802.15.4

Procedia PDF Downloads 552
16730 Estimating Anthropometric Dimensions for Saudi Males Using Artificial Neural Networks

Authors: Waleed Basuliman

Abstract:

Anthropometric dimensions are considered one of the important factors when designing human-machine systems. In this study, the estimation of anthropometric dimensions has been improved by using Artificial Neural Network (ANN) model that is able to predict the anthropometric measurements of Saudi males in Riyadh City. A total of 1427 Saudi males aged 6 to 60 years participated in measuring 20 anthropometric dimensions. These anthropometric measurements are considered important for designing the work and life applications in Saudi Arabia. The data were collected during eight months from different locations in Riyadh City. Five of these dimensions were used as predictors variables (inputs) of the model, and the remaining 15 dimensions were set to be the measured variables (Model’s outcomes). The hidden layers varied during the structuring stage, and the best performance was achieved with the network structure 6-25-15. The results showed that the developed Neural Network model was able to estimate the body dimensions of Saudi male population in Riyadh City. The network's mean absolute percentage error (MAPE) and the root mean squared error (RMSE) were found to be 0.0348 and 3.225, respectively. These results were found less, and then better, than the errors found in the literature. Finally, the accuracy of the developed neural network was evaluated by comparing the predicted outcomes with regression model. The ANN model showed higher coefficient of determination (R2) between the predicted and actual dimensions than the regression model.

Keywords: artificial neural network, anthropometric measurements, back-propagation

Procedia PDF Downloads 487
16729 Money and Inflation in Cambodia

Authors: Siphat Lim

Abstract:

The result of the study revealed that the interaction between money, exchange rate, and price level was mainly derived from the policy-induced by the central bank. Furthermore, the variation of inflation was explained weakly by exchange rate and money supply. In the period of twelfth-month, the variation of inflation which caused by exchange rate and money supply were not more than 1.78 percent and 9.77 percent, respectively.

Keywords: money supply, exchange rate, price level, VAR model

Procedia PDF Downloads 288
16728 Structural Damage Detection Using Modal Data Employing Teaching Learning Based Optimization

Authors: Subhajit Das, Nirjhar Dhang

Abstract:

Structural damage detection is a challenging work in the field of structural health monitoring (SHM). The damage detection methods mainly focused on the determination of the location and severity of the damage. Model updating is a well known method to locate and quantify the damage. In this method, an error function is defined in terms of difference between the signal measured from ‘experiment’ and signal obtained from undamaged finite element model. This error function is minimised with a proper algorithm, and the finite element model is updated accordingly to match the measured response. Thus, the damage location and severity can be identified from the updated model. In this paper, an error function is defined in terms of modal data viz. frequencies and modal assurance criteria (MAC). MAC is derived from Eigen vectors. This error function is minimized by teaching-learning-based optimization (TLBO) algorithm, and the finite element model is updated accordingly to locate and quantify the damage. Damage is introduced in the model by reduction of stiffness of the structural member. The ‘experimental’ data is simulated by the finite element modelling. The error due to experimental measurement is introduced in the synthetic ‘experimental’ data by adding random noise, which follows Gaussian distribution. The efficiency and robustness of this method are explained through three examples e.g., one truss, one beam and one frame problem. The result shows that TLBO algorithm is efficient to detect the damage location as well as the severity of damage using modal data.

Keywords: damage detection, finite element model updating, modal assurance criteria, structural health monitoring, teaching learning based optimization

Procedia PDF Downloads 215
16727 Development of an Automatic Control System for ex vivo Heart Perfusion

Authors: Pengzhou Lu, Liming Xin, Payam Tavakoli, Zhonghua Lin, Roberto V. P. Ribeiro, Mitesh V. Badiwala

Abstract:

Ex vivo Heart Perfusion (EVHP) has been developed as an alternative strategy to expand cardiac donation by enabling resuscitation and functional assessment of hearts donated from marginal donors, which were previously not accepted. EVHP parameters, such as perfusion flow (PF) and perfusion pressure (PP) are crucial for optimal organ preservation. However, with the heart’s constant physiological changes during EVHP, such as coronary vascular resistance, manual control of these parameters is rendered imprecise and cumbersome for the operator. Additionally, low control precision and the long adjusting time may lead to irreversible damage to the myocardial tissue. To solve this problem, an automatic heart perfusion system was developed by applying a Human-Machine Interface (HMI) and a Programmable-Logic-Controller (PLC)-based circuit to control PF and PP. The PLC-based control system collects the data of PF and PP through flow probes and pressure transducers. It has two control modes: the RPM-flow mode and the pressure mode. The RPM-flow control mode is an open-loop system. It influences PF through providing and maintaining the desired speed inputted through the HMI to the centrifugal pump with a maximum error of 20 rpm. The pressure control mode is a closed-loop system where the operator selects a target Mean Arterial Pressure (MAP) to control PP. The inputs of the pressure control mode are the target MAP, received through the HMI, and the real MAP, received from the pressure transducer. A PID algorithm is applied to maintain the real MAP at the target value with a maximum error of 1mmHg. The precision and control speed of the RPM-flow control mode were examined by comparing the PLC-based system to an experienced operator (EO) across seven RPM adjustment ranges (500, 1000, 2000 and random RPM changes; 8 trials per range) tested in a random order. System’s PID algorithm performance in pressure control was assessed during 10 EVHP experiments using porcine hearts. Precision was examined through monitoring the steady-state pressure error throughout perfusion period, and stabilizing speed was tested by performing two MAP adjustment changes (4 trials per change) of 15 and 20mmHg. A total of 56 trials were performed to validate the RPM-flow control mode. Overall, the PLC-based system demonstrated the significantly faster speed than the EO in all trials (PLC 1.21±0.03, EO 3.69±0.23 seconds; p < 0.001) and greater precision to reach the desired RPM (PLC 10±0.7, EO 33±2.7 mean RPM error; p < 0.001). Regarding pressure control, the PLC-based system has the median precision of ±1mmHg error and the median stabilizing times in changing 15 and 20mmHg of MAP are 15 and 19.5 seconds respectively. The novel PLC-based control system was 3 times faster with 60% less error than the EO for RPM-flow control. In pressure control mode, it demonstrates a high precision and fast stabilizing speed. In summary, this novel system successfully controlled perfusion flow and pressure with high precision, stability and a fast response time through a user-friendly interface. This design may provide a viable technique for future development of novel heart preservation and assessment strategies during EVHP.

Keywords: automatic control system, biomedical engineering, ex-vivo heart perfusion, human-machine interface, programmable logic controller

Procedia PDF Downloads 175
16726 Perception of TQM Implementation and Perceived Cost of Poor Quality: A Case Study of Local Automotive Company’s Supplier

Authors: Fakhruddin Esa, Yusri Yusof

Abstract:

The confirmatory of Total Quality Management (TQM) implementation is most vital in quality management. This paper focuses on employees' perceptions towards TQM implementation in a local automotive company supplier. The objectives of this study are first and foremost to determine the perception of TQM implementation among the staff, and secondly to ascertain the correlation between the variables, and lastly to identify the relative influence of the 10 TQM variables on the cost of poor quality (COPQ). The TQM implementation is perceived to be moderate. All correlation is found to be significant and five variables having positively moderate to high correlation. Out of 10 variables, quality system improvement, reward and recognition and customer focus influence the perceived COPQ. This study extended a discussion on these three variables contribution to TQM in general and the human resource development in the organization. A significant recommendation to lowering costs of internal error, such as trouble shooting and scraps are also discussed. Certain components of further research that would add value to this study have also been suggested and perhaps could be implemented at policy-level initiatives.

Keywords: cost of poor quality (COPQ), correlation, total quality management (TQM), variables

Procedia PDF Downloads 216
16725 Influence of Transportation Mode to the Deterioration Rate: Case Study of Food Transport by Ship

Authors: Danijela Tuljak-Suban, Valter Suban

Abstract:

Food as perishable goods represents a specific and sensitive part in the supply chain theory, since changing of its physical or chemical characteristics considerably influences the approach to stock management. The most delicate phase of this process is transportation, where it becomes difficult to ensure stability conditions that limit the deterioration, since the value of the deterioration rate could be easily influenced by the transportation mode. Fuzzy definition of variables allows taking into account these variations. Furthermore an appropriate choice of the defuzzification method permits to adapt results, as much as possible, to real conditions. In the article will be applied the those methods to the relationship between the deterioration rate of perishable goods and transportation by ship, with the aim: (a) to minimize the total costs function, defined as the sum of the ordering cost, holding cost, disposing cost and transportation costs, and (b) to improve supply chain sustainability by reducing the environmental impact and waste disposal costs.

Keywords: perishable goods, fuzzy reasoning, transport by ship, supply chain sustainability

Procedia PDF Downloads 543
16724 Allometric Models for Biomass Estimation in Savanna Woodland Area, Niger State, Nigeria

Authors: Abdullahi Jibrin, Aishetu Abdulkadir

Abstract:

The development of allometric models is crucial to accurate forest biomass/carbon stock assessment. The aim of this study was to develop a set of biomass prediction models that will enable the determination of total tree aboveground biomass for savannah woodland area in Niger State, Nigeria. Based on the data collected through biometric measurements of 1816 trees and destructive sampling of 36 trees, five species specific and one site specific models were developed. The sample size was distributed equally between the five most dominant species in the study site (Vitellaria paradoxa, Irvingia gabonensis, Parkia biglobosa, Anogeissus leiocarpus, Pterocarpus erinaceous). Firstly, the equations were developed for five individual species. Secondly these five species were mixed and were used to develop an allometric equation of mixed species. Overall, there was a strong positive relationship between total tree biomass and the stem diameter. The coefficient of determination (R2 values) ranging from 0.93 to 0.99 P < 0.001 were realised for the models; with considerable low standard error of the estimates (SEE) which confirms that the total tree above ground biomass has a significant relationship with the dbh. The F-test value for the biomass prediction models were also significant at p < 0.001 which indicates that the biomass prediction models are valid. This study recommends that for improved biomass estimates in the study site, the site specific biomass models should preferably be used instead of using generic models.

Keywords: allometriy, biomass, carbon stock , model, regression equation, woodland, inventory

Procedia PDF Downloads 448
16723 Driving Forces of Bank Liquidity: Evidence from Selected Ethiopian Private Commercial Banks

Authors: Tadele Tesfay Teame, Tsegaye Abrehame, Hágen István Zsombor

Abstract:

Liquidity is one of the main concerns for banks, and thus achieving the optimum level of liquidity is critical. The main objective of this study is to discover the driving force of selected private commercial banks’ liquidity. In order to achieve the objective explanatory research design and quantitative research approach were used. Data has been collected from a secondary source of the sampled Ethiopian private commercial banks’ financial statements, the National Bank of Ethiopia, and the Minister of Finance, the sample covering the period from 2011 to 2022. Bank-specific and macroeconomic variables were analyzed by using the balanced panel fixed effect regression model. Bank’s liquidity ratio is measured by the total liquid asset to total deposits. The findings of the study revealed that bank size, capital adequacy, loan growth rate, and non-performing loan had a statistically significant impact on private commercial banks’ liquidity, and annual inflation rate and interest rate margin had a statistically significant impact on the liquidity of Ethiopian private commercial banks measured by L1 (bank liquidity). Thus, banks in Ethiopia should not only be concerned about internal structures and policies/procedures, but they must consider both the internal environment and the macroeconomic environment together in developing their strategies to efficiently manage their liquidity position and private commercial banks to maintain their financial proficiency shall have bank liquidity management policy by assimilating both bank-specific and macro-economic variables.

Keywords: liquidity, Ethiopian private commercial banks, liquidity ratio, panel data regression analysis

Procedia PDF Downloads 99