Search results for: an optimal error bound
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5140

Search results for: an optimal error bound

4300 Surface Display of Lipase on Yarrowia lipolytica Cells

Authors: Evgeniya Y. Yuzbasheva, Tigran V. Yuzbashev, Natalia I. Perkovskaya, Elizaveta B. Mostova

Abstract:

Cell-surface display of lipase is of great interest as it has many applications in the field of biotechnology owing to its unique advantages: simplified product purification, and cost-effective downstream processing. One promising area of application for whole-cell biocatalysts with surface displayed lipase is biodiesel synthesis. Biodiesel is biodegradable, renewable, and nontoxic alternative fuel for diesel engines. Although the alkaline catalysis method has been widely used for biodiesel production, it has a number of limitations, such as rigorous feedstock specifications, complicated downstream processes, including removal of inorganic salts from the product, recovery of the salt-containing by-product glycerol, and treatment of alkaline wastewater. Enzymatic synthesis of biodiesel can overcome these drawbacks. In this study, Lip2p lipase was displayed on Yarrowia lipolytica cells via C- and N-terminal fusion variant. The active site of lipase is located near the C-terminus, therefore to prevent the activity loosing the insertion of glycine-serine linker between Lip2p and C-domains was performed. The hydrolytic activity of the displayed lipase reached 12,000–18,000 U/g of dry weight. However, leakage of enzyme from the cell wall was observed. In case of C-terminal fusion variant, the leakage was occurred due to the proteolytic cleavage within the linker peptide. In case of N-terminal fusion variant, the leaking enzyme was presented as three proteins, one of which corresponded to the whole hybrid protein. The calculated number of recombinant enzyme displayed on the cell surface is approximately 6–9 × 105 molecules per cell, which is close to the theoretical maximum (2 × 106 molecules/cell). Thus, we attribute the enzyme leakage to the limited space available on the cell surface. Nevertheless, cell-bound lipase exhibited greater stability to short-term and long-term temperature treatment than the native enzyme. It retained 74% of original activity at 60°C for 5 min of incubation, and 83% of original activity after incubation at 50°C during 5 h. Cell-bound lipase had also higher stability in organic solvents and detergents. The developed whole-cell biocatalyst was used for recycling biodiesel synthesis. Two repeated cycles of methanolysis yielded 84.1–% and 71.0–% methyl esters after 33–h and 45–h reactions, respectively.

Keywords: biodiesel, cell-surface display, lipase, whole-cell biocatalyst

Procedia PDF Downloads 470
4299 Mean-Field Type Modeling of Non-Local Congestion in Pedestrian Crowd Dynamics

Authors: Alexander Aurell

Abstract:

One of the latest trends in the modeling of human crowds is the mean-field game approach. In the mean-field game approach, the motion of a human crowd is described by a nonstandard stochastic optimal control problem. It is nonstandard since congestion is considered, introduced through a dependence in the performance functional on the distribution of the crowd. This study extends the class of mean-field pedestrian crowd models to allow for non-local congestion and arbitrary, but finitely, many interacting crowds. The new congestion feature grants pedestrians a 'personal space' where crowding is undesirable. The model is treated as a mean-field type game which is derived from a particle picture. This, in contrast to a mean-field game, better describes a situation where the crowd can be controlled by a central planner. The latter is suitable for decentralized situations. Solutions to the mean-field type game are characterized via a Pontryagin-type Maximum Principle.

Keywords: congestion, crowd dynamics, interacting populations, mean-field approximation, optimal control

Procedia PDF Downloads 429
4298 Comparative Study of Accuracy of Land Cover/Land Use Mapping Using Medium Resolution Satellite Imagery: A Case Study

Authors: M. C. Paliwal, A. K. Jain, S. K. Katiyar

Abstract:

Classification of satellite imagery is very important for the assessment of its accuracy. In order to determine the accuracy of the classified image, usually the assumed-true data are derived from ground truth data using Global Positioning System. The data collected from satellite imagery and ground truth data is then compared to find out the accuracy of data and error matrices are prepared. Overall and individual accuracies are calculated using different methods. The study illustrates advanced classification and accuracy assessment of land use/land cover mapping using satellite imagery. IRS-1C-LISS IV data were used for classification of satellite imagery. The satellite image was classified using the software in fourteen classes namely water bodies, agricultural fields, forest land, urban settlement, barren land and unclassified area etc. Classification of satellite imagery and calculation of accuracy was done by using ERDAS-Imagine software to find out the best method. This study is based on the data collected for Bhopal city boundaries of Madhya Pradesh State of India.

Keywords: resolution, accuracy assessment, land use mapping, satellite imagery, ground truth data, error matrices

Procedia PDF Downloads 490
4297 The Whale Optimization Algorithm and Its Implementation in MATLAB

Authors: S. Adhirai, R. P. Mahapatra, Paramjit Singh

Abstract:

Optimization is an important tool in making decisions and in analysing physical systems. In mathematical terms, an optimization problem is the problem of finding the best solution from among the set of all feasible solutions. The paper discusses the Whale Optimization Algorithm (WOA), and its applications in different fields. The algorithm is tested using MATLAB because of its unique and powerful features. The benchmark functions used in WOA algorithm are grouped as: unimodal (F1-F7), multimodal (F8-F13), and fixed-dimension multimodal (F14-F23). Out of these benchmark functions, we show the experimental results for F7, F11, and F19 for different number of iterations. The search space and objective space for the selected function are drawn, and finally, the best solution as well as the best optimal value of the objective function found by WOA is presented. The algorithmic results demonstrate that the WOA performs better than the state-of-the-art meta-heuristic and conventional algorithms.

Keywords: optimization, optimal value, objective function, optimization problems, meta-heuristic optimization algorithms, Whale Optimization Algorithm, implementation, MATLAB

Procedia PDF Downloads 353
4296 Ranking of Optimal Materials for Building Walls from the Perspective of Cost and Waste of Electricity and Gas Energy Using AHP-TOPSIS 1 Technique: Study Example: Sari City

Authors: Seyedomid Fatemi

Abstract:

The walls of the building, as the main intermediary between the outside and the inside of the building, play an important role in controlling the environmental conditions and ensuring the comfort of the residents, thus reducing the heating and cooling loads. Therefore, the use of suitable materials is considered one of the simplest and most effective ways to reduce the heating and cooling loads of the building, which will also save energy. Therefore, in order to achieve the goal of the research "Ranking of optimal materials for building walls," optimal materials for building walls in a temperate and humid climate (case example: Sari city) from the perspective of embodied energy, waste of electricity and gas energy, cost and reuse been investigated to achieve sustainable architecture. In this regard, using information obtained from Sari Municipality, design components have been presented by experts using the Delphi method. Considering the criteria of experts' opinions (cost and reuse), the amount of embodied energy of the materials, as well as the amount of waste of electricity and gas of different materials of the walls, with the help of the AHP weighting technique and finally with the TOPSIS technique, the best type of materials in the order of 1- 3-D Panel 2-ICF-, 3-Cement block with pumice, 4-Wallcrete block, 5-Clay block, 6-Autoclaved Aerated Concrete (AAC), 7-Foam cement block, 8-Aquapanel and 9-Reinforced concrete wall for use in The walls of the buildings were proposed in Sari city.

Keywords: optimum materials, building walls, moderate and humid climate, sustainable architecture, AHP-TOPSIS technique

Procedia PDF Downloads 54
4295 Identification of Bayesian Network with Convolutional Neural Network

Authors: Mohamed Raouf Benmakrelouf, Wafa Karouche, Joseph Rynkiewicz

Abstract:

In this paper, we propose an alternative method to construct a Bayesian Network (BN). This method relies on a convolutional neural network (CNN classifier), which determinates the edges of the network skeleton. We train a CNN on a normalized empirical probability density distribution (NEPDF) for predicting causal interactions and relationships. We have to find the optimal Bayesian network structure for causal inference. Indeed, we are undertaking a search for pair-wise causality, depending on considered causal assumptions. In order to avoid unreasonable causal structure, we consider a blacklist and a whitelist of causality senses. We tested the method on real data to assess the influence of education on the voting intention for the extreme right-wing party. We show that, with this method, we get a safer causal structure of variables (Bayesian Network) and make to identify a variable that satisfies the backdoor criterion.

Keywords: Bayesian network, structure learning, optimal search, convolutional neural network, causal inference

Procedia PDF Downloads 153
4294 The Combined Effect of Methane and Methanol on Growth and PHB Production in the Alphaproteobacterial Methanotroph Methylocystis Sp. Rockwell

Authors: Lazic Marina, Sugden Scott, Sharma Kanta Hem, Sauvageau Dominic, Stein Lisa

Abstract:

Methane is a highly potent greenhouse gas mostly released through anthropogenic activities. Methane represents a low-cost and sustainable feedstock used for the biological production of value-added compounds by bacteria known as methanotrophs. In addition to methane, these organisms can utilize methanol, another cheap carbon source that is a common industrial by-product. Alphaproteobacteria methanotrophs can utilize both methane and methanol to produce the biopolymer polyhydroxybutyrate. The goal of this study was to examine the effect of methanol on polyhydroxybutyrate production in Methylocystis sp. Rockwell and to identify the optimal methane: methanol ratio that will improve PHB without reducing biomass production. Three methane: methanol ratios (4, 2.5., and 0.5) and three nitrogen source (ammonium or nitrate) concentrations (10 mM, 1 mM, and 0.1 mM) were combined to generate 18 growing conditions (9 per carbon source). The production of polyhydroxybutyrate and biomass was analyzed at the end of growth. Overall, the methane: methanol ratios that promoted polyhydroxybutyrate synthesis without reducing biomass were 4 and 2.5 and the optimal nitrogen concentration was 1 mM for both ammonium and nitrate. The physiological mechanism behind the beneficial effect of combining methane and methanol as carbon sources remain to be discovered. One possibility is that methanol has a dual role as a carbon source at lower concentrations and as a stringent response trigger at higher concentrations. Nevertheless, the beneficial effect of methanol and optimal nitrogen concentration for PHB production was confirmed, providing a basis for future physiological analysis and conditions for process scale-up.

Keywords: methane, methanol, methanotrophs, polyhydroxybutyrate, methylocystis sp. rockwell, single carbon bioconversions

Procedia PDF Downloads 145
4293 Machine Learning Approach for Automating Electronic Component Error Classification and Detection

Authors: Monica Racha, Siva Chandrasekaran, Alex Stojcevski

Abstract:

The engineering programs focus on promoting students' personal and professional development by ensuring that students acquire technical and professional competencies during four-year studies. The traditional engineering laboratory provides an opportunity for students to "practice by doing," and laboratory facilities aid them in obtaining insight and understanding of their discipline. Due to rapid technological advancements and the current COVID-19 outbreak, the traditional labs were transforming into virtual learning environments. Aim: To better understand the limitations of the physical laboratory, this research study aims to use a Machine Learning (ML) algorithm that interfaces with the Augmented Reality HoloLens and predicts the image behavior to classify and detect the electronic components. The automated electronic components error classification and detection automatically detect and classify the position of all components on a breadboard by using the ML algorithm. This research will assist first-year undergraduate engineering students in conducting laboratory practices without any supervision. With the help of HoloLens, and ML algorithm, students will reduce component placement error on a breadboard and increase the efficiency of simple laboratory practices virtually. Method: The images of breadboards, resistors, capacitors, transistors, and other electrical components will be collected using HoloLens 2 and stored in a database. The collected image dataset will then be used for training a machine learning model. The raw images will be cleaned, processed, and labeled to facilitate further analysis of components error classification and detection. For instance, when students conduct laboratory experiments, the HoloLens captures images of students placing different components on a breadboard. The images are forwarded to the server for detection in the background. A hybrid Convolutional Neural Networks (CNNs) and Support Vector Machines (SVMs) algorithm will be used to train the dataset for object recognition and classification. The convolution layer extracts image features, which are then classified using Support Vector Machine (SVM). By adequately labeling the training data and classifying, the model will predict, categorize, and assess students in placing components correctly. As a result, the data acquired through HoloLens includes images of students assembling electronic components. It constantly checks to see if students appropriately position components in the breadboard and connect the components to function. When students misplace any components, the HoloLens predicts the error before the user places the components in the incorrect proportion and fosters students to correct their mistakes. This hybrid Convolutional Neural Networks (CNNs) and Support Vector Machines (SVMs) algorithm automating electronic component error classification and detection approach eliminates component connection problems and minimizes the risk of component damage. Conclusion: These augmented reality smart glasses powered by machine learning provide a wide range of benefits to supervisors, professionals, and students. It helps customize the learning experience, which is particularly beneficial in large classes with limited time. It determines the accuracy with which machine learning algorithms can forecast whether students are making the correct decisions and completing their laboratory tasks.

Keywords: augmented reality, machine learning, object recognition, virtual laboratories

Procedia PDF Downloads 119
4292 Optimization of Assay Parameters of L-Glutaminase from Bacillus cereus MTCC1305 Using Artificial Neural Network

Authors: P. Singh, R. M. Banik

Abstract:

Artificial neural network (ANN) was employed to optimize assay parameters viz., time, temperature, pH of reaction mixture, enzyme volume and substrate concentration of L-glutaminase from Bacillus cereus MTCC 1305. ANN model showed high value of coefficient of determination (0.9999), low value of root mean square error (0.6697) and low value of absolute average deviation. A multilayer perceptron neural network trained with an error back-propagation algorithm was incorporated for developing a predictive model and its topology was obtained as 5-3-1 after applying Levenberg Marquardt (LM) training algorithm. The predicted activity of L-glutaminase was obtained as 633.7349 U/l by considering optimum assay parameters, viz., pH of reaction mixture (7.5), reaction time (20 minutes), incubation temperature (35˚C), substrate concentration (40mM), and enzyme volume (0.5ml). The predicted data was verified by running experiment at simulated optimum assay condition and activity was obtained as 634.00 U/l. The application of ANN model for optimization of assay conditions improved the activity of L-glutaminase by 1.499 fold.

Keywords: Bacillus cereus, L-glutaminase, assay parameters, artificial neural network

Procedia PDF Downloads 416
4291 Effect of Punch Diameter on Optimal Loading Profiles in Hydromechanical Deep Drawing Process

Authors: Mehmet Halkaci, Ekrem Öztürk, Mevlüt Türköz, H. Selçuk Halkacı

Abstract:

Hydromechanical deep drawing (HMD) process is an advanced manufacturing process used to form deep parts with only one forming step. In this process, sheet metal blank can be drawn deeper by means of fluid pressure acting on sheet surface in the opposite direction of punch movement. High limiting drawing ratio, good surface quality, less springback characteristic and high dimensional accuracy are some of the advantages of this process. The performance of the HMD process is affected by various process parameters such as fluid pressure, blank holder force, punch-die radius, pre-bulging pressure and height, punch diameter, friction between sheet-die and sheet-punch. The fluid pressure and bank older force are the main loading parameters and affect the formability of HMD process significantly. The punch diameter also influences the limiting drawing ratio (the ratio of initial sheet diameter to punch diameter) of the sheet metal blank. In this research, optimal loading (fluid pressure and blank holder force) profiles were determined for AA 5754-O sheet material through fuzzy control algorithm developed in previous study using LS-DYNA finite element analysis (FEA) software. In the preceding study, the fuzzy control algorithm was developed utilizing geometrical criteria such as thinning and wrinkling. In order to obtain the final desired part with the developed algorithm in terms of the punch diameter requested, the effect of punch diameter, which is the one of the process parameters, on loading profiles was investigated separately using blank thickness of 1 mm. Thus, the practicality of the previously developed fuzzy control algorithm with different punch diameters was clarified. Also, thickness distributions of the sheet metal blank along a curvilinear distance were compared for the FEA in which different punch diameters were used. Consequently, it was found that the use of different punch diameters did not affect the optimal loading profiles too much.

Keywords: Finite Element Analysis (FEA), fuzzy control, hydromechanical deep drawing, optimal loading profiles, punch diameter

Procedia PDF Downloads 408
4290 The Optimization of Sun Collector Parameters

Authors: István Patkó, Hosam Bayoumi Hamuda, András Szeder

Abstract:

In order to efficiently solve the problems created by the deepening energy crisis affecting Europe and the world, governments cannot neglect the opportunities of using the energy produced by sun collectors. In many of the EU countries there are sun collectors producing heat energy, e.g. in 2011 in the area of EU27 (countries which belong to European Union) + Switzerland altogether 37519126 m2 were operated, which are capable of producing 26.3 GWh heat energy. The energy produced by these sun collectors is utilized at the place of production. In the near future governments will have to focus more on spreading and using sun collectors. Among the complex problems of operating sun collectors, this article deals with determining the optimal tilt angle, directions of sun collectors. We evaluate the contamination of glass surface of sun collector to the produced energy. Our theoretically results are confirmed by laboratory measurements. The purpose of our work is to help users and engineers in determination of optimal operation parameters of sun collectors.

Keywords: heat energy, tilt angle, direction of sun collector, contamination of surface

Procedia PDF Downloads 416
4289 Design of Parity-Preserving Reversible Logic Signed Array Multipliers

Authors: Mojtaba Valinataj

Abstract:

Reversible logic as a new favorable design domain can be used for various fields especially creating quantum computers because of its speed and intangible power consumption. However, its susceptibility to a variety of environmental effects may lead to yield the incorrect results. In this paper, because of the importance of multiplication operation in various computing systems, some novel reversible logic array multipliers are proposed with error detection capability by incorporating the parity-preserving gates. The new designs are presented for two main parts of array multipliers, partial product generation and multi-operand addition, by exploiting the new arrangements of existing gates, which results in two signed parity-preserving array multipliers. The experimental results reveal that the best proposed 4×4 multiplier in this paper reaches 12%, 24%, and 26% enhancements in the number of constant inputs, number of required gates, and quantum cost, respectively, compared to previous design. Moreover, the best proposed design is generalized for n×n multipliers with general formulations to estimate the main reversible logic criteria as the functions of the multiplier size.

Keywords: array multipliers, Baugh-Wooley method, error detection, parity-preserving gates, quantum computers, reversible logic

Procedia PDF Downloads 244
4288 The Moment of the Optimal Average Length of the Multivariate Exponentially Weighted Moving Average Control Chart for Equally Correlated Variables

Authors: Edokpa Idemudia Waziri, Salisu S. Umar

Abstract:

The Hotellng’s T^2 is a well-known statistic for detecting a shift in the mean vector of a multivariate normal distribution. Control charts based on T have been widely used in statistical process control for monitoring a multivariate process. Although it is a powerful tool, the T statistic is deficient when the shift to be detected in the mean vector of a multivariate process is small and consistent. The Multivariate Exponentially Weighted Moving Average (MEWMA) control chart is one of the control statistics used to overcome the drawback of the Hotellng’s T statistic. In this paper, the probability distribution of the Average Run Length (ARL) of the MEWMA control chart when the quality characteristics exhibit substantial cross correlation and when the process is in-control and out-of-control was derived using the Markov Chain algorithm. The derivation of the probability functions and the moments of the run length distribution were also obtained and they were consistent with some existing results for the in-control and out-of-control situation. By simulation process, the procedure identified a class of ARL for the MEWMA control when the process is in-control and out-of-control. From our study, it was observed that the MEWMA scheme is quite adequate for detecting a small shift and a good way to improve the quality of goods and services in a multivariate situation. It was also observed that as the in-control average run length ARL0¬ or the number of variables (p) increases, the optimum value of the ARL0pt increases asymptotically and as the magnitude of the shift σ increases, the optimal ARLopt decreases. Finally, we use the example from the literature to illustrate our method and demonstrate its efficiency.

Keywords: average run length, markov chain, multivariate exponentially weighted moving average, optimal smoothing parameter

Procedia PDF Downloads 402
4287 Flow Field Optimization for Proton Exchange Membrane Fuel Cells

Authors: Xiao-Dong Wang, Wei-Mon Yan

Abstract:

The flow field design in the bipolar plates affects the performance of the proton exchange membrane (PEM) fuel cell. This work adopted a combined optimization procedure, including a simplified conjugate-gradient method and a completely three-dimensional, two-phase, non-isothermal fuel cell model, to look for optimal flow field design for a single serpentine fuel cell of size 9×9 mm with five channels. For the direct solution, the two-fluid method was adopted to incorporate the heat effects using energy equations for entire cells. The model assumes that the system is steady; the inlet reactants are ideal gases; the flow is laminar; and the porous layers such as the diffusion layer, catalyst layer and PEM are isotropic. The model includes continuity, momentum and species equations for gaseous species, liquid water transport equations in the channels, gas diffusion layers, and catalyst layers, water transport equation in the membrane, electron and proton transport equations. The Bulter-Volumer equation was used to describe electrochemical reactions in the catalyst layers. The cell output power density Pcell is maximized subjected to an optimal set of channel heights, H1-H5, and channel widths, W2-W5. The basic case with all channel heights and widths set at 1 mm yields a Pcell=7260 Wm-2. The optimal design displays a tapered characteristic for channels 1, 3 and 4, and a diverging characteristic in height for channels 2 and 5, producing a Pcell=8894 Wm-2, about 22.5% increment. The reduced channel heights of channels 2-4 significantly increase the sub-rib convection and widths for effectively removing liquid water and oxygen transport in gas diffusion layer. The final diverging channel minimizes the leakage of fuel to outlet via sub-rib convection from channel 4 to channel 5. Near-optimal design without huge loss in cell performance but is easily manufactured is tested. The use of a straight, final channel of 0.1 mm height has led to 7.37% power loss, while the design with all channel widths to be 1 mm with optimal channel heights obtained above yields only 1.68% loss of current density. The presence of a final, diverging channel has greater impact on cell performance than the fine adjustment of channel width at the simulation conditions set herein studied.

Keywords: optimization, flow field design, simplified conjugate-gradient method, serpentine flow field, sub-rib convection

Procedia PDF Downloads 282
4286 Reconstruction of Signal in Plastic Scintillator of PET Using Tikhonov Regularization

Authors: L. Raczynski, P. Moskal, P. Kowalski, W. Wislicki, T. Bednarski, P. Bialas, E. Czerwinski, A. Gajos, L. Kaplon, A. Kochanowski, G. Korcyl, J. Kowal, T. Kozik, W. Krzemien, E. Kubicz, Sz. Niedzwiecki, M. Palka, Z. Rudy, O. Rundel, P. Salabura, N.G. Sharma, M. Silarski, A. Slomski, J. Smyrski, A. Strzelecki, A. Wieczorek, M. Zielinski, N. Zon

Abstract:

The J-PET scanner, which allows for single bed imaging of the whole human body, is currently under development at the Jagiellonian University. The J-PET detector improves the TOF resolution due to the use of fast plastic scintillators. Since registration of the waveform of signals with duration times of few nanoseconds is not feasible, a novel front-end electronics allowing for sampling in a voltage domain at four thresholds was developed. To take fully advantage of these fast signals a novel scheme of recovery of the waveform of the signal, based on ideas from the Tikhonov regularization (TR) and Compressive Sensing methods, is presented. The prior distribution of sparse representation is evaluated based on the linear transformation of the training set of waveform of the signals by using the Principal Component Analysis (PCA) decomposition. Beside the advantage of including the additional information from training signals, a further benefit of the TR approach is that the problem of signal recovery has an optimal solution which can be determined explicitly. Moreover, from the Bayes theory the properties of regularized solution, especially its covariance matrix, may be easily derived. This step is crucial to introduce and prove the formula for calculations of the signal recovery error. It has been proven that an average recovery error is approximately inversely proportional to the number of samples at voltage levels. The method is tested using signals registered by means of the single detection module of the J-PET detector built out from the 30 cm long BC-420 plastic scintillator strip. It is demonstrated that the experimental and theoretical functions describing the recovery errors in the J-PET scenario are largely consistent. The specificity and limitations of the signal recovery method in this application are discussed. It is shown that the PCA basis offers high level of information compression and an accurate recovery with just eight samples, from four voltage levels, for each signal waveform. Moreover, it is demonstrated that using the recovered waveform of the signals, instead of samples at four voltage levels alone, improves the spatial resolution of the hit position reconstruction. The experiment shows that spatial resolution evaluated based on information from four voltage levels, without a recovery of the waveform of the signal, is equal to 1.05 cm. After the application of an information from four voltage levels to the recovery of the signal waveform, the spatial resolution is improved to 0.94 cm. Moreover, the obtained result is only slightly worse than the one evaluated using the original raw-signal. The spatial resolution calculated under these conditions is equal to 0.93 cm. It is very important information since, limiting the number of threshold levels in the electronic devices to four, leads to significant reduction of the overall cost of the scanner. The developed recovery scheme is general and may be incorporated in any other investigation where a prior knowledge about the signals of interest may be utilized.

Keywords: plastic scintillators, positron emission tomography, statistical analysis, tikhonov regularization

Procedia PDF Downloads 425
4285 Multiple Linear Regression for Rapid Estimation of Subsurface Resistivity from Apparent Resistivity Measurements

Authors: Sabiu Bala Muhammad, Rosli Saad

Abstract:

Multiple linear regression (MLR) models for fast estimation of true subsurface resistivity from apparent resistivity field measurements are developed and assessed in this study. The parameters investigated were apparent resistivity (ρₐ), horizontal location (X) and depth (Z) of measurement as the independent variables; and true resistivity (ρₜ) as the dependent variable. To achieve linearity in both resistivity variables, datasets were first transformed into logarithmic domain following diagnostic checks of normality of the dependent variable and heteroscedasticity to ensure accurate models. Four MLR models were developed based on hierarchical combination of the independent variables. The generated MLR coefficients were applied to another data set to estimate ρₜ values for validation. Contours of the estimated ρₜ values were plotted and compared to the observed data plots at the colour scale and blanking for visual assessment. The accuracy of the models was assessed using coefficient of determination (R²), standard error (SE) and weighted mean absolute percentage error (wMAPE). It is concluded that the MLR models can estimate ρₜ for with high level of accuracy.

Keywords: apparent resistivity, depth, horizontal location, multiple linear regression, true resistivity

Procedia PDF Downloads 258
4284 Study on the Retaining Sleeve Structure for the Reduction of Eddy Current in SPMSM

Authors: Hyun-Woo Jun, In-Gun Kim, Hyun Seok Hong, Dong-Woo Kang, Ju Lee

Abstract:

In high-speed SPMSM design, the rotor-retaining sleeve is inserted into rotor to prevent permanent magnet’s damage. It is quite efficient way considering manufacturability, but the sleeve becomes major source of ohm loss in high-speed operation. In this paper, the high-speed motor for turbo-blower at the rating of 100kW was introduced. To improve its efficiency, the retaining sleeve’s optimal design was needed. Within the range of satisfies the mechanical safety, sleeve’s some design variables have been changed. The effect of changing design variables of the sleeve was studied. This paper presents the optimized sleeve’s advantages in electrical efficiency from the result of electromagnetic FEA (finite element analysis) software. Finally, it suggests the optimal sleeve design to reduce eddy current loss, which is related to motor shape.

Keywords: SPMSM, sleeve, eddy current, high efficiency

Procedia PDF Downloads 410
4283 Nurse-Reported Perceptions of Medication Safety in Private Hospitals in Gauteng Province.

Authors: Madre Paarlber, Alwiena Blignaut

Abstract:

Background: Medication administration errors remains a global patient safety problem targeted by the WHO (World Health Organization), yet research on this matter is sparce within the South African context. Objective: The aim was to explore and describe nurses’ (medication administrators) perceptions regarding medication administration safety-related culture, incidence, causes, and reporting in the Gauteng Province of South Africa, and to determine any relationships between perceived variables concerned with medication safety (safety culture, incidences, causes, reporting of incidences, and reasons for non-reporting). Method: A quantitative research design was used through which self-administered online surveys were sent to 768 nurses (medication administrators) (n=217). The response rate was 28.26%. The survey instrument was synthesised from the Agency of Healthcare Research and Quality (AHRQ) Hospital Survey on Patient Safety Culture, the Registered Nurse Forecasting (RN4CAST) survey, a survey list prepared from a systematic review aimed at generating a comprehensive list of medication administration error causes and the Medication Administration Error Reporting Survey from Wakefield. Exploratory and confirmatory factor analyses were used to determine the validity and reliability of the survey. Descriptive and inferential statistical data analysis were used to analyse quantitative data. Relationships and correlations were identified between items, subscales and biographic data by using Spearmans’ Rank correlations, T-Tests and ANOVAs (Analysis of Variance). Nurses reported on their perceptions of medication administration safety-related culture, incidence, causes, and reporting in the Gauteng Province. Results: Units’ teamwork deemed satisfactory, punitive responses to errors accentuated. “Crisis mode” working, concerns regarding mistake recording and long working hours disclosed as impacting patient safety. Overall medication safety graded mostly positively. Work overload, high patient-nurse ratios, and inadequate staffing implicated as error-inducing. Medication administration errors were reported regularly. Fear and administrative response to errors effected non-report. Non-report of errors’ reasons was affected by non-punitive safety culture. Conclusions: Medication administration safety improvement is contingent on fostering a non-punitive safety culture within units. Anonymous medication error reporting systems and auditing nurses’ workload are recommended in the quest of improved medication safety within Gauteng Province private hospitals.

Keywords: incidence, medication administration errors, medication safety, reporting, safety culture

Procedia PDF Downloads 30
4282 Optimal Trailing Edge Flap Positions of Helicopter Rotor for Various Thrust Coefficient to Solidity (Ct/σ) Ratios

Authors: K. K. Saijaand, K. Prabhakaran Nair

Abstract:

This study aims to determine change in optimal lo-cations of dual trailing-edge flaps for various thrust coefficient to solidity (Ct /σ) ratios of helicopter to achieve minimum hub vibration levels, with low penalty in terms of required trailing-edge flap control power. Polynomial response functions are used to approximate hub vibration and flap power objective functions. Single objective and multi-objective optimization is carried with the objective of minimizing hub vibration and flap power. The optimization results shows that the inboard flap location at low Ct/σ ratio move farther from the baseline value and at high Ct/σ ratio move towards the root of the blade for minimizing hub vibration.

Keywords: helicopter rotor, trailing-edge flap, thrust coefficient to solidity (Ct /σ) ratio, optimization

Procedia PDF Downloads 458
4281 Stochastic Optimization of a Vendor-Managed Inventory Problem in a Two-Echelon Supply Chain

Authors: Bita Payami-Shabestari, Dariush Eslami

Abstract:

The purpose of this paper is to develop a multi-product economic production quantity model under vendor management inventory policy and restrictions including limited warehouse space, budget, and number of orders, average shortage time and maximum permissible shortage. Since the “costs” cannot be predicted with certainty, it is assumed that data behave under uncertain environment. The problem is first formulated into the framework of a bi-objective of multi-product economic production quantity model. Then, the problem is solved with three multi-objective decision-making (MODM) methods. Then following this, three methods had been compared on information on the optimal value of the two objective functions and the central processing unit (CPU) time with the statistical analysis method and the multi-attribute decision-making (MADM). The results are compared with statistical analysis method and the MADM. The results of the study demonstrate that augmented-constraint in terms of optimal value of the two objective functions and the CPU time perform better than global criteria, and goal programming. Sensitivity analysis is done to illustrate the effect of parameter variations on the optimal solution. The contribution of this research is the use of random costs data in developing a multi-product economic production quantity model under vendor management inventory policy with several constraints.

Keywords: economic production quantity, random cost, supply chain management, vendor-managed inventory

Procedia PDF Downloads 112
4280 Preventative Maintenance, Impact on the Optimal Replacement Strategy of Secondhand Products

Authors: Pin-Wei Chiang, Wen-Liang Chang, Ruey-Huei Yeh

Abstract:

This paper investigates optimal replacement and preventative maintenance policies of secondhand products under a Finite Planning Horizon (FPH). Any consumer wishing to replace their product under FPH would have it undergo minimal repairs. The replacement provided would be required to undergo periodical preventive maintenance done to avoid product failure. Then, a mathematical formula for disbursement cost for products under FPH can be derived. Optimal policies are then obtained to minimize cost. In the first of two segments of the paper, a model for initial product purchase of either new or secondhand products is used. This model is built by analyzing product purchasing price, surplus value of product, as well as the minimal repair cost. The second segment uses a model for replacement products, which are also secondhand products with no limit on usage. This model analyzes the same components as the first as well as expected preventative maintenance cost. Using these two models, a formula for the expected final total cost can be developed. The formula requires four variables (optimal preventive maintenance level, preventive maintenance frequency, replacement timing, age of replacement product) to find minimal cost requirement. Based on analysis of the variables using the expected total final cost model, it was found that the purchasing price and length of ownership were directly related. Also, consumers should choose the secondhand product with the higher usage for replacement. Products with higher initial usage upon acquisition require an earlier replacement schedule. In this case, replacements should be made with a secondhand product with less usage. In addition, preventative maintenance also significantly reduces cost. Consumers that plan to use products for longer periods of time replace their products later. Hence these consumers should choose the secondhand product with lesser initial usage for replacement. Preventative maintenance also creates significant total cost savings in this case. This study provides consumers with a method of calculating both the ideal amount of usage of the products they should purchase as well as the frequency and level of preventative maintenance that should be conducted in order to minimize cost and maintain product function.

Keywords: finite planning horizon, second hand product, replacement, preventive maintenance, minimal repair

Procedia PDF Downloads 458
4279 Modal Analysis for Optimal Location of Doubly Fed Induction-Generator-Based Wind Farms for Reduction of Small Signal Oscillation

Authors: Meet Patel, Darshan Patel, Nilay Shah

Abstract:

Excess growth of wind-based renewable energy sources is required to identify the optimal location and damping capacity of doubly fed induction-generator-based (DFIG) wind farms while it penetrates into the transmission network. In this analysis, various ratings of DFIG wind farms are penetrated into the Single Machine Infinite Bus (SMIB ) at a different distance of the transmission line. On the basis of detailed examinations, a prime position is evaluated to maximize the stability of overall systems. A damping controller is designed at an optimum location to mitigate the small oscillations. The proposed model was validated using eigenvalue analysis, calculation of the participation factor, and time-domain simulation.

Keywords: DFIG, small signal stability, eigenvalues, time domain simulation

Procedia PDF Downloads 91
4278 Numerical Investigation of Geotextile Application in Clay Reinforcement in ABAQUS Software

Authors: Seyed Abolhasan Naeini, Eisa Aliagahei

Abstract:

Today, the use of geosynthetic materials in geotechnical activities is increasing significantly. One of the main uses of these materials is to increase the compressive strength of clay reinforced by geotextile layers. In the present study, the effect of clay reinforcement by geotextile layers in increasing the compressive strength of clay has been investigated using modeling in ABAQUS 6.11.3 software. For this purpose, the modified Drager Prager model has been chosen to simulate the stress-strain behavior of soil layers and the linear elastic model for the geotextile layer. Unreinforced samples and reinforced samples are modeled by geotextile layers (1, 2 and 3 geotextile layers) by software. In order to validate the results, an article in the same field was used and the numerical modeling results were calibrated with the laboratory results. Based on the obtained results, the software has a suitable capability for modeling and the results of the numerical model overlap with the laboratory results to a very acceptable extent, by increasing the number of geotextile layers, the error between the results of the laboratory sample and the software model increases. The highest amount of error is related to the sample reinforced with three layers of geotextile and is 7.3%.

Keywords: Abaqus, cap model, clay, geotextile layer, reinforced soil

Procedia PDF Downloads 73
4277 Selection of Soil Quality Indicators of Rice Cropping Systems Using Minimum Data Set Influenced by Imbalanced Fertilization

Authors: Theresa K., Shanmugasundaram R., Kennedy J. S.

Abstract:

Nutrient supplements are indispensable for raising crops and to reap determining productivity. The nutrient imbalance between replenishment and crop uptake is attempted through the input of inorganic fertilizers. Excessive dumping of inorganic nutrients in soil cause stagnant and decline in yield. Imbalanced N-P-K ratio in the soil exacerbates and agitates the soil ecosystems. The study evaluated the fertilization practices of conventional (CFs), organic and Integrated Nutrient Management system (INM) on soil quality using key indicators and soil quality indices. Twelve rice farming fields of which, ten fields were having conventional cultivation practices, one field each was organic farming based and INM based cultivated under monocropping sequence in the Thondamuthur block of Coimbatore district were fixed and properties viz., physical, chemical and biological were studied for four cropping seasons to determine soil quality index (SQI). SQI was computed for conventional, organic and INM fields. Comparing conventional farming (CF) with organic and INM, CF was recorded with a lower soil quality index. While in organic and INM fields, the higher SQI value of 0.99 and 0.88 respectively were registered. CF₄ received with a super-optimal dose of N (250%) showed a lesser SQI value (0.573) as well as the yield (3.20 t ha⁻¹) and the CF6 which received 125 % N recorded the highest SQI (0.715) and yield (6.20 t ha⁻¹). Likewise, most of the CFs received higher N beyond the level of 125 % except CF₃ and CF₉, which recorded lower yields. CFs which received super-optimal P in the order of CF₆&CF₇>CF₁&CF₁₀ recorded lesser yields except for CF₆. Super-optimal K application also recorded lesser yield in CF₄, CF₇ and CF₉.

Keywords: rice cropping system, soil quality indicators, imbalanced fertilization, yield

Procedia PDF Downloads 138
4276 Optimal Risk and Financial Stability

Authors: Rahmoune Abdelhaq

Abstract:

Systemic risk is a key concern for central banks charged with safeguarding overall financial stability. In this work, we investigate how systemic risk is affected by the structure of the financial system. We construct banking systems that are composed of a number of banks that are connected by interbank linkages. We then vary the key parameters that define the structure of the financial system — including its level of capitalization, the degree to which banks are connected, the size of interbank exposures and the degree of concentration of the system — and analyses the influence of these parameters on the likelihood of contagious (knock-on) defaults. First, we find that the better-capitalized banks are, the more resilient is the banking system against contagious defaults and this effect is non-linear. Second, the effect of the degree of connectivity is non-monotonic, that is, initially a small increase in connectivity increases the contagion effect; but after a certain threshold value, connectivity improves the ability of a banking system to absorb shocks. Third, the size of interbank liabilities tends to increase the risk of knock-on default, even if banks hold capital against such exposures. Fourth, more concentrated banking systems are shown to be prone to larger systemic risk, all else equal. In an extension to the main analysis, we study how liquidity effects interact with banking structure to produce a greater chance of systemic breakdown. We finally consider how the risk of contagion might depend on the degree of asymmetry (tier) inherent in the structure of the banking system. A number of our results have important implications for public policy, which this paper also draws out. This paper also discusses why bank risk management is needed to get the optimal one.

Keywords: financial stability, contagion, liquidity risk, optimal risk

Procedia PDF Downloads 382
4275 Optimizing Human Diet Problem Using Linear Programming Approach: A Case Study

Authors: P. Priyanka, S. Shruthi, N. Guruprasad

Abstract:

Health is a common theme in most cultures. In fact all communities have their concepts of health, as part of their culture. Health continues to be a neglected entity. Planning of Human diet should be done very careful by selecting the food items or groups of food items also the composition involved. Low price and good taste of foods are regarded as two major factors for optimal human nutrition. Linear programming techniques have been extensively used for human diet formulation for quiet good number of years. Through the process, we mainly apply “The Simplex Method” which is a very useful statistical tool based on the theorem of Elementary Row Operation from Linear Algebra and also incorporate some other necessary rules set by the Simplex Method to help solve the problem. The study done by us is an attempt to develop a programming model for optimal planning and best use of nutrient ingredients.

Keywords: diet formulation, linear programming, nutrient ingredients, optimization, simplex method

Procedia PDF Downloads 541
4274 A Longitudinal Case Study of Greek as a Second Language

Authors: M. Vassou, A. Karasimos

Abstract:

A primary concern in the field of Second Language Acquisition (SLA) research is to determine the innate mechanisms of second language learning and acquisition through the systematic study of a learner's interlanguage. Errors emerge while a learner attempts to communicate using the target-language and can be seen either as the observable linguistic product of the latent cognitive and language process of mental representations or as an indispensable learning mechanism. Therefore, the study of the learner’s erroneous forms may depict the various strategies and mechanisms that take place during the language acquisition process resulting in deviations from the target-language norms and difficulties in communication. Mapping the erroneous utterances of a late adult learner in the process of acquiring Greek as a second language constitutes one of the main aims of this study. For our research purposes, we created an error-tagged learner corpus composed of the participant’s written texts produced throughout a period of a 4- year instructed language acquisition. Error analysis and interlanguage theory constitute the methodological and theoretical framework, respectively. The research questions pertain to the learner's most frequent errors per linguistic category and per year as well as his choices concerning the Greek Article System. According to the quantitative analysis of the data, the most frequent errors are observed in the categories of the stress system and syntax, whereas a significant fluctuation and/or gradual reduction throughout the 4 years of instructed acquisition indicate the emergence of developmental stages. The findings with regard to the article usage bespeak fossilization of erroneous structures in certain contexts. In general, our results point towards the existence and further development of an established learner’s (inter-) language system governed not only by mother- tongue and target-language influences but also by the learner’s assumptions and set of rules as the result of a complex cognitive process. It is expected that this study will contribute not only to the knowledge in the field of Greek as a second language and SLA generally, but it will also provide an insight into the cognitive mechanisms and strategies developed by multilingual learners of late adulthood.

Keywords: Greek as a second language, error analysis, interlanguage, late adult learner

Procedia PDF Downloads 112
4273 Optimal Image Representation for Linear Canonical Transform Multiplexing

Authors: Navdeep Goel, Salvador Gabarda

Abstract:

Digital images are widely used in computer applications. To store or transmit the uncompressed images requires considerable storage capacity and transmission bandwidth. Image compression is a means to perform transmission or storage of visual data in the most economical way. This paper explains about how images can be encoded to be transmitted in a multiplexing time-frequency domain channel. Multiplexing involves packing signals together whose representations are compact in the working domain. In order to optimize transmission resources each 4x4 pixel block of the image is transformed by a suitable polynomial approximation, into a minimal number of coefficients. Less than 4*4 coefficients in one block spares a significant amount of transmitted information, but some information is lost. Different approximations for image transformation have been evaluated as polynomial representation (Vandermonde matrix), least squares + gradient descent, 1-D Chebyshev polynomials, 2-D Chebyshev polynomials or singular value decomposition (SVD). Results have been compared in terms of nominal compression rate (NCR), compression ratio (CR) and peak signal-to-noise ratio (PSNR) in order to minimize the error function defined as the difference between the original pixel gray levels and the approximated polynomial output. Polynomial coefficients have been later encoded and handled for generating chirps in a target rate of about two chirps per 4*4 pixel block and then submitted to a transmission multiplexing operation in the time-frequency domain.

Keywords: chirp signals, image multiplexing, image transformation, linear canonical transform, polynomial approximation

Procedia PDF Downloads 400
4272 An Integrated Approach for Optimal Selection of Machining Parameters in Laser Micro-Machining Process

Authors: A. Gopala Krishna, M. Lakshmi Chaitanya, V. Kalyana Manohar

Abstract:

In the existent analysis, laser micro machining (LMM) of Silicon carbide (SiCp) reinforced Aluminum 7075 Metal Matrix Composite (Al7075/SiCp MMC) was studied. While machining, Because of the intense heat generated, A layer gets formed on the work piece surface which is called recast layer and this layer is detrimental to the surface quality of the component. The recast layer needs to be as small as possible for precise applications. Therefore, The height of recast layer and the depth of groove which are conflicting in nature were considered as the significant manufacturing criteria, Which determines the pursuit of a machining process obtained in LMM of Al7075/10%SiCp composite. The present work formulates the depth of groove and height of recast layer in relation to the machining parameters using the Response Surface Methodology (RSM) and correspondingly, The formulated mathematical models were put to use for optimization. Since the effect of machining parameters on the depth of groove and height of recast layer was contradictory, The problem was explicated as a multi objective optimization problem. Moreover, An evolutionary Non-dominated sorting genetic algorithm (NSGA-II) was employed to optimize the model established by RSM. Subsequently this algorithm was also adapted to achieve the Pareto optimal set of solutions that provide a detailed illustration for making the optimal solutions. Eventually experiments were conducted to affirm the results obtained from RSM and NSGA-II.

Keywords: Laser Micro Machining (LMM), depth of groove, Height of recast layer, Response Surface Methodology (RSM), non-dominated sorting genetic algorithm

Procedia PDF Downloads 331
4271 Reducing Energy Consumption and GHG Emission by Integration of Flare Gas with Fuel Gas Network in Refinery

Authors: N. Tahouni, M. Gholami, M. H. Panjeshahi

Abstract:

Gas flaring is one of the most GHG emitting sources in the oil and gas industries. It is also a major way for wasting such an energy that could be better utilized and even generates revenue. Minimize flaring is an effective approach for reducing GHG emissions and also conserving energy in flaring systems. Integrating waste and flared gases into the fuel gas networks (FGN) of refineries is an efficient tool. A fuel gas network collects fuel gases from various source streams and mixes them in an optimal manner, and supplies them to different fuel sinks such as furnaces, boilers, turbines, etc. In this article we use fuel gas network model proposed by Hasan et al. as a base model and modify some of its features and add constraints on emission pollution by gas flaring to reduce GHG emissions as possible. Results for a refinery case study showed that integration of flare gas stream with waste and natural gas streams to construct an optimal FGN can significantly reduce total annualized cost and flaring emissions.

Keywords: flaring, fuel gas network, GHG emissions, stream

Procedia PDF Downloads 317