Search results for: sequential units distribution
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6787

Search results for: sequential units distribution

6097 Electronic Nose for Monitoring Fungal Deterioration of Stored Rapeseed

Authors: Robert Rusinek, Marek Gancarz, Jolanta Wawrzyniak, Marzena Gawrysiak-Witulska, Dariusz Wiącek, Agnieszka Nawrocka

Abstract:

Investigations were performed to examine the possibility of using an electronic nose to monitor the development of fungal microflora during the first eighteen days of rapeseed storage. The Cyranose 320 device with polymer-composite sensors was used. Each sample of infected material was divided into three parts, and the degree of spoilage was measured in three ways: analysis of colony forming units (CFU), determination of ergosterol content (ERG), and measurement with the eNose. Principal component analysis (PCA) was performed on the generated patterns of signals, and six groups of different spoilage levels were isolated. The electronic nose with polymer-composite sensors under laboratory conditions distinguished between species of spoiled and unspoiled seeds with 100% accuracy. Despite some minor differences in the CFU and ergosterol content, the electronic nose provided responses correctly corresponding to the level of spoilage with 85% accuracy. Therefore, the main conclusion from the study is that the electronic nose is a promising tool for quick and non-destructive detection of the level of oil seed spoilage. The research was supported by the National Centre for Research and Development (NCBR), Grant No. PBS2/A8/22/2013.

Keywords: colony forming units, electronic nose, ergosterol, rapeseed

Procedia PDF Downloads 321
6096 Estimation of Particle Size Distribution Using Magnetization Data

Authors: Navneet Kaur, S. D. Tiwari

Abstract:

Magnetic nanoparticles possess fascinating properties which make their behavior unique in comparison to corresponding bulk materials. Superparamagnetism is one such interesting phenomenon exhibited only by small particles of magnetic materials. In this state, the thermal energy of particles become more than their magnetic anisotropy energy, and so particle magnetic moment vectors fluctuate between states of minimum energy. This situation is similar to paramagnetism of non-interacting ions and termed as superparamagnetism. The magnetization of such systems has been described by Langevin function. But, the estimated fit parameters, in this case, are found to be unphysical. It is due to non-consideration of particle size distribution. In this work, analysis of magnetization data on NiO nanoparticles is presented considering the effect of particle size distribution. Nanoparticles of NiO of two different sizes are prepared by heating freshly synthesized Ni(OH)₂ at different temperatures. Room temperature X-ray diffraction patterns confirm the formation of single phase of NiO. The diffraction lines are seen to be quite broad indicating the nanocrystalline nature of the samples. The average crystallite size are estimated to be about 6 and 8 nm. The samples are also characterized by transmission electron microscope. Magnetization of both sample is measured as function of temperature and applied magnetic field. Zero field cooled and field cooled magnetization are measured as a function of temperature to determine the bifurcation temperature. The magnetization is also measured at several temperatures in superparamagnetic region. The data are fitted to an appropriate expression considering a distribution in particle size following a least square fit procedure. The computer codes are written in PYTHON. The presented analysis is found to be very useful for estimating the particle size distribution present in the samples. The estimated distributions are compared with those determined from transmission electron micrographs.

Keywords: anisotropy, magnetization, nanoparticles, superparamagnetism

Procedia PDF Downloads 141
6095 Tracking the Effect of Ibutilide on Amplitude and Frequency of Fibrillatory Intracardiac Electrograms Using the Regression Analysis

Authors: H. Hajimolahoseini, J. Hashemi, D. Redfearn

Abstract:

Background: Catheter ablation is an effective therapy for symptomatic atrial fibrillation (AF). The intracardiac electrocardiogram (IEGM) collected during this procedure contains precious information that has not been explored to its full capacity. Novel processing techniques allow looking at these recordings from different perspectives which can lead to improved therapeutic approaches. In our previous study, we showed that variation in amplitude measured through Shannon Entropy could be used as an AF recurrence risk stratification factor in patients who received Ibutilide before the electrograms were recorded. The aim of this study is to further investigate the effect of Ibutilide on characteristics of the recorded signals from the left atrium (LA) of a patient with persistent AF before and after administration of the drug. Methods: The IEGMs collected from different intra-atrial sites of 12 patients were studied and compared before and after Ibutilide administration. First, the before and after Ibutilide IEGMs that were recorded within a Euclidian distance of 3 mm in LA were selected as pairs for comparison. For every selected pair of IEGMs, the Probability Distribution Function (PDF) of the amplitude in time domain and magnitude in frequency domain was estimated using the regression analysis. The PDF represents the relative likelihood of a variable falling within a specific range of values. Results: Our observations showed that in time domain, the PDF of amplitudes was fitted to a Gaussian distribution while in frequency domain, it was fitted to a Rayleigh distribution. Our observations also revealed that after Ibutilide administration, the IEGMs would have significantly narrower short-tailed PDFs both in time and frequency domains. Conclusion: This study shows that the PDFs of the IEGMs before and after administration of Ibutilide represents significantly different properties, both in time and frequency domains. Hence, by fitting the PDF of IEGMs in time domain to a Gaussian distribution or in frequency domain to a Rayleigh distribution, the effect of Ibutilide can easily be tracked using the statistics of their PDF (e.g., standard deviation) while this is difficult through the waveform of IEGMs itself.

Keywords: atrial fibrillation, catheter ablation, probability distribution function, time-frequency characteristics

Procedia PDF Downloads 158
6094 Functioning of Public Distribution System and Calories Intake in the State of Maharashtra

Authors: Balasaheb Bansode, L. Ladusingh

Abstract:

The public distribution system is an important component of food security. It is a massive welfare program undertaken by Government of India and implemented by state government since India being a federal state; for achieving multiple objectives like eliminating hunger, reduction in malnutrition and making food consumption affordable. This program reaches at the community level through the various agencies of the government. The paper focuses on the accessibility of PDS at household level and how the present policy framework results in exclusion and inclusion errors. It tries to explore the sanctioned food grain quantity received by differentiated ration cards according to income criterion at household level, and also it has highlighted on the type of corruption in food distribution that is generated by the PDS system. The data used is of secondary nature from NSSO 68 round conducted in 2012. Bivariate and multivariate techniques have been used to understand the working and consumption of food for this paper.

Keywords: calories intake, entitle food quantity, poverty aliviation through PDS, target error

Procedia PDF Downloads 331
6093 An ANOVA-based Sequential Forward Channel Selection Framework for Brain-Computer Interface Application based on EEG Signals Driven by Motor Imagery

Authors: Forouzan Salehi Fergeni

Abstract:

Converting the movement intents of a person into commands for action employing brain signals like electroencephalogram signals is a brain-computer interface (BCI) system. When left or right-hand motions are imagined, different patterns of brain activity appear, which can be employed as BCI signals for control. To make better the brain-computer interface (BCI) structures, effective and accurate techniques for increasing the classifying precision of motor imagery (MI) based on electroencephalography (EEG) are greatly needed. Subject dependency and non-stationary are two features of EEG signals. So, EEG signals must be effectively processed before being used in BCI applications. In the present study, after applying an 8 to 30 band-pass filter, a car spatial filter is rendered for the purpose of denoising, and then, a method of analysis of variance is used to select more appropriate and informative channels from a category of a large number of different channels. After ordering channels based on their efficiencies, a sequential forward channel selection is employed to choose just a few reliable ones. Features from two domains of time and wavelet are extracted and shortlisted with the help of a statistical technique, namely the t-test. Finally, the selected features are classified with different machine learning and neural network classifiers being k-nearest neighbor, Probabilistic neural network, support-vector-machine, Extreme learning machine, decision tree, Multi-layer perceptron, and linear discriminant analysis with the purpose of comparing their performance in this application. Utilizing a ten-fold cross-validation approach, tests are performed on a motor imagery dataset found in the BCI competition III. Outcomes demonstrated that the SVM classifier got the greatest classification precision of 97% when compared to the other available approaches. The entire investigative findings confirm that the suggested framework is reliable and computationally effective for the construction of BCI systems and surpasses the existing methods.

Keywords: brain-computer interface, channel selection, motor imagery, support-vector-machine

Procedia PDF Downloads 47
6092 Nonparametric Copula Approximations

Authors: Serge Provost, Yishan Zang

Abstract:

Copulas are currently utilized in finance, reliability theory, machine learning, signal processing, geodesy, hydrology and biostatistics, among several other fields of scientific investigation. It follows from Sklar's theorem that the joint distribution function of a multidimensional random vector can be expressed in terms of its associated copula and marginals. Since marginal distributions can easily be determined by making use of a variety of techniques, we address the problem of securing the distribution of the copula. This will be done by using several approaches. For example, we will obtain bivariate least-squares approximations of the empirical copulas, modify the kernel density estimation technique and propose a criterion for selecting appropriate bandwidths, differentiate linearized empirical copulas, secure Bernstein polynomial approximations of suitable degrees, and apply a corollary to Sklar's result. Illustrative examples involving actual observations will be presented. The proposed methodologies will as well be applied to a sample generated from a known copula distribution in order to validate their effectiveness.

Keywords: copulas, Bernstein polynomial approximation, least-squares polynomial approximation, kernel density estimation, density approximation

Procedia PDF Downloads 72
6091 Influence of the Non-Uniform Distribution of Filler Porosity on the Thermal Performance of Sensible Heat Thermocline Storage Tanks

Authors: Yuchao Hua, Lingai Luo

Abstract:

Thermal energy storage is of critical importance for the highly-efficient utilization of renewable energy sources. Over the past decades, single-tank thermocline technology has attracted much attention owing to its high cost-effectiveness. In the present work, we investigate the influence of the filler porosity’s non-uniform distribution on the thermal performance of the packed-bed sensible heat thermocline storage tanks on the basis of the analytical model obtained by the Laplace transform. It is found that when the total amount of filler materials (i.e., the integration of porosity) is fixed, the different porosity distributions can result in the significantly-different behaviors of outlet temperature and thus the varied charging and discharging efficiencies. Our results indicate that a non-uniform distribution of the fillers with the proper design can improve the heat storage performance without changing the total amount of the filling materials.

Keywords: energy storage, heat thermocline storage tank, packed bed, transient thermal analysis

Procedia PDF Downloads 93
6090 Numerical Investigation of Fluid Flow and Temperature Distribution on Power Transformer Windings Using Open Foam

Authors: Saeed Khandan Siar, Stefan Tenbohlen, Christian Breuer, Raphael Lebreton

Abstract:

The goal of this article is to investigate the detailed temperature distribution and the fluid flow of an oil cooled winding of a power transformer by means of computational fluid dynamics (CFD). The experimental setup consists of three passes of a zig-zag cooled disc type winding, in which losses are modeled by heating cartridges in each winding segment. A precise temperature sensor measures the temperature of each turn. The laboratory setup allows the exact control of the boundary conditions, e.g. the oil flow rate and the inlet temperature. Furthermore, a simulation model is solved using the open source computational fluid dynamics solver OpenFOAM and validated with the experimental results. The model utilizes the laminar and turbulent flow for the different mass flow rate of the oil. The good agreement of the simulation results with experimental measurements validates the model.

Keywords: CFD, conjugated heat transfer, power transformers, temperature distribution

Procedia PDF Downloads 419
6089 Top-Down Approach for Fabricating Hematite Nanowire Arrays

Authors: Seungmin Shin, Jin-Baek Kim

Abstract:

Hematite (α-Fe2O3) has very good semiconducting properties with a band gap of 2.1 eV and is antiferromagnetic. Due to its electrochemical stability, low toxicity, wide abundance, and low-cost, hematite, it is a particularly attractive material for photoelectrochemical cells. Additionally, hematite has also found applications in gas sensing, field emission, heterogeneous catalysis, and lithium-ion battery electrodes. Here, we discovered a new universal top-down method for the synthesis of one-dimensional hematite nanowire arrays. Various shapes and lengths of hematite nanowire have been easily fabricated over large areas by sequential processes. The obtained hematite nanowire arrays are promising candidates as photoanodes in photoelectrochemical solar cells.

Keywords: hematite, lithography, nanowire, top-down process

Procedia PDF Downloads 247
6088 Enhancing the Implementation Strategy of Simultaneous Operations (SIMOPS) for the Major Turnaround at Pertamina Plaju Refinery

Authors: Fahrur Rozi, Daniswara Krisna Prabatha, Latief Zulfikar Chusaini

Abstract:

Amidst the backdrop of Pertamina Plaju Refinery, which stands as the oldest and historically less technologically advanced among Pertamina's refineries, lies a unique challenge. Originally integrating facilities established by Shell in 1904 and Stanvac (originally Standard Oil) in 1926, the primary challenge at Plaju Refinery does not solely revolve around complexity; instead, it lies in ensuring reliability, considering its operational history of over a century. After centuries of existence, Plaju Refinery has never undergone a comprehensive major turnaround encompassing all its units. The usual practice involves partial turnarounds that are sequentially conducted across its primary, secondary, and tertiary units (utilities and offsite). However, a significant shift is on the horizon. In the Q-IV of 2023, the refinery embarks on its first-ever major turnaround since its establishment. This decision was driven by the alignment of maintenance timelines across various units. Plaju Refinery's major turnaround was scheduled for October-November 2023, spanning 45 calendar days, with the objective of enhancing the operational reliability of all refinery units. The extensive job list for this turnaround encompasses 1583 tasks across 18 units/areas, involving approximately 9000 contracted workers. In this context, the Strategy of Simultaneous Operations (SIMOPS) execution emerges as a pivotal tool to optimize time efficiency and ensure safety. A Hazard Effect Management Process (HEMP) has been employed to assess the risk ratings of each task within the turnaround. Out of the tasks assessed, 22 are deemed high-risk and necessitate mitigation. The SIMOPS approach serves as a preventive measure against potential incidents. It is noteworthy that every turnaround period at Pertamina Plaju Refinery involves SIMOPS-related tasks. In this context, enhancing the implementation strategy of "Simultaneous Operations (SIMOPS)" becomes imperative to minimize the occurrence of incidents. At least four improvements have been introduced in the enhancement process for the major turnaround at Refinery Plaju. The first improvement involves conducting systematic risk assessment and potential hazard mitigation studies for SIMOPS tasks before task execution, as opposed to the previous on-site approach. The second improvement includes the completion of SIMOPS Job Mitigation and Work Matrices Sheets, which was often neglected in the past. The third improvement emphasizes comprehensive awareness to workers/contractors regarding potential hazards and mitigation strategies for SIMOPS tasks before and during the major turnaround. The final improvement is the introduction of a daily program for inspecting and observing work in progress for SIMOPS tasks. Prior to these improvements, there was no established program for monitoring ongoing activities related to SIMOPS tasks during the turnaround. This study elucidates the steps taken to enhance SIMOPS within Pertamina, drawing from the experiences of Plaju Refinery as a guide. A real actual case study will be provided from our experience in the operational unit. In conclusion, these efforts are essential for the success of the first-ever major turnaround at Plaju Refinery, with the SIMOPS strategy serving as a central component. Based on these experiences, enhancements have been made to Pertamina's official Internal Guidelines for Executing SIMOPS Risk Mitigation, benefiting all Pertamina units.

Keywords: process safety management, turn around, oil refinery, risk assessment

Procedia PDF Downloads 72
6087 The Effect of Different Parameters on a Single Invariant Lateral Displacement Distribution to Consider the Higher Modes Effect in a Displacement-Based Pushover Procedure

Authors: Mohamad Amin Amini, Mehdi Poursha

Abstract:

Nonlinear response history analysis (NL-RHA) is a robust analytical tool for estimating the seismic demands of structures responding in the inelastic range. However, because of its conceptual and numerical complications, the nonlinear static procedure (NSP) is being increasingly used as a suitable tool for seismic performance evaluation of structures. The conventional pushover analysis methods presented in various codes (FEMA 356; Eurocode-8; ATC-40), are limited to the first-mode-dominated structures, and cannot take higher modes effect into consideration. Therefore, since more than a decade ago, researchers developed enhanced pushover analysis procedures to take higher modes effect into account. The main objective of this study is to propose an enhanced invariant lateral displacement distribution to take higher modes effect into consideration in performing a displacement-based pushover analysis, whereby a set of laterally applied displacements, rather than forces, is monotonically applied to the structure. For this purpose, the effect of different parameters such as the spectral displacement of ground motion, the modal participation factor, and the effective modal participating mass ratio on the lateral displacement distribution is investigated to find the best distribution. The major simplification of this procedure is that the effect of higher modes is concentrated into a single invariant lateral load distribution. Therefore, only one pushover analysis is sufficient without any need to utilize a modal combination rule for combining the responses. The invariant lateral displacement distribution for pushover analysis is then calculated by combining the modal story displacements using the modal combination rules. The seismic demands resulting from the different procedures are compared to those from the more accurate nonlinear response history analysis (NL-RHA) as a benchmark solution. Two structures of different heights including 10 and 20-story special steel moment resisting frames (MRFs) were selected and evaluated. Twenty ground motion records were used to conduct the NL-RHA. The results show that more accurate responses can be obtained in comparison with the conventional lateral loads when the enhanced modal lateral displacement distributions are used.

Keywords: displacement-based pushover, enhanced lateral load distribution, higher modes effect, nonlinear response history analysis (NL-RHA)

Procedia PDF Downloads 275
6086 Placer Gold Deposits in Madari Gold Mine, Southern Eastern Desert, Egypt: Orientation, Source and Distribution

Authors: Tarek Sedki

Abstract:

Madari gold mine is delineated by latitudes 22° 30' 29" and 22° 32' 33" N and longitudes 36° 24' 03" and 35°11' 44" E. Geologically, Madari rock units are classified into dismembered ophiolites, arc volcanic assemblage, syntectonic metagabbro-diorites and Mineralized quartz diorite and granodiorite. Deposition of gold in area occurred as a direct result of weathering of nearby gold-bearing veins. Main concentrations of gold are supposed to ensue close to the bed rock. Nevertheless, the several shallow channel-fill features covering lag deposits, arising throughout the alluvial fan sequence would definitely contain a percentage of the finer gold due to the limited washing and sorting capacity of the uncommon flood events. Gold deposits arise as disseminated and separate gold with limited pyrite, arsenopyrite and chalcopyrite everywhere veins in the wall rocks and lode gold deposits in quartz veins. In places, the wall rocks, in near district of the quartz vein, are grieved strong silicification, chloritization and pyritization as a result of a metasomatic alteration due to purification of external hydrothermal fluids. Quartz veins are mostly steeply dipping and display banding features and frequently sheared and brecciated.

Keywords: Madari gold mine, placer deposits, southern eastern desert, gold mineralization, quartz veins

Procedia PDF Downloads 139
6085 Numerical Approach of RC Structural MembersExposed to Fire and After-Cooling Analysis

Authors: Ju-young Hwang, Hyo-Gyoung Kwak, Hong Jae Yim

Abstract:

This paper introduces a numerical analysis method for reinforced-concrete (RC) structures exposed to fire and compares the result with experimental results. The proposed analysis method for RC structure under the high temperature consists of two procedures. First step is to decide the temperature distribution across the section through the heat transfer analysis by using the time-temperature curve. After determination of the temperature distribution, the nonlinear analysis is followed. By considering material and geometrical non-linearity with the temperature distribution, nonlinear analysis predicts the behavior of RC structure under the fire by the exposed time. The proposed method is validated by the comparison with the experimental results. Finally, Prediction model to describe the status of after-cooling concrete can also be introduced based on the results of additional experiment. The product of this study is expected to be embedded for smart structure monitoring system against fire in u-City.

Keywords: RC structures, heat transfer analysis, nonlinear analysis, after-cooling concrete model

Procedia PDF Downloads 365
6084 Loss Allocation in Radial Distribution Networks for Loads of Composite Types

Authors: Sumit Banerjee, Chandan Kumar Chanda

Abstract:

The paper presents allocation of active power losses and energy losses to consumers connected to radial distribution networks in a deregulated environment for loads of composite types. A detailed comparison among four algorithms, namely quadratic loss allocation, proportional loss allocation, pro rata loss allocation and exact loss allocation methods are presented. Quadratic and proportional loss allocations are based on identifying the active and reactive components of current in each branch and the losses are allocated to each consumer, pro rata loss allocation method is based on the load demand of each consumer and exact loss allocation method is based on the actual contribution of active power loss by each consumer. The effectiveness of the proposed comparison among four algorithms for composite load is demonstrated through an example.

Keywords: composite type, deregulation, loss allocation, radial distribution networks

Procedia PDF Downloads 286
6083 Passenger Flow Characteristics of Seoul Metropolitan Subway Network

Authors: Kang Won Lee, Jung Won Lee

Abstract:

Characterizing the network flow is of fundamental importance to understand the complex dynamics of networks. And passenger flow characteristics of the subway network are very relevant for an effective transportation management in urban cities. In this study, passenger flow of Seoul metropolitan subway network is investigated and characterized through statistical analysis. Traditional betweenness centrality measure considers only topological structure of the network and ignores the transportation factors. This paper proposes a weighted betweenness centrality measure that incorporates monthly passenger flow volume. We apply the proposed measure on the Seoul metropolitan subway network involving 493 stations and 16 lines. Several interesting insights about the network are derived from the new measures. Using Kolmogorov-Smirnov test, we also find out that monthly passenger flow between any two stations follows a power-law distribution and other traffic characteristics such as congestion level and throughflow traffic follow exponential distribution.

Keywords: betweenness centrality, correlation coefficient, power-law distribution, Korea traffic DB

Procedia PDF Downloads 289
6082 Residual Stress Around Embedded Particles in Bulk YBa2Cu3Oy Samples

Authors: Anjela Koblischka-Veneva, Michael R. Koblischka

Abstract:

To increase the flux pinning performance of bulk YBa2Cu3O7-δ (YBCO or Y-123) superconductors, it is common to employ secondary phase particles, either Y2BaCuO5 (Y-211) particles created during the growth of the samples or additionally added (nano)particles of various types, embedded in the superconducting Y-123 matrix. As the crystallographic parameters of all the particles indicate a misfit to Y-123, there will be residual strain within the Y-123 matrix around such particles. With a dedicated analysis of electron backscatter diffraction (EBSD) data obtained on various bulk, Y-123 superconductor samples, the strain distribution around such embedded secondary phase particles can be revealed. The results obtained are presented in form of Kernel Average Misorientation (KAM) mappings. Around large Y-211 particles, the strain can be so large that YBCO subgrains are formed. Therefore, it is essential to properly control the particle size as well as their distribution within the bulk sample to obtain the best performance. The impact of the strain distribution on the flux pinning properties is discussed.

Keywords: Bulk superconductors, EBSD, Strain, YBa2Cu3Oy

Procedia PDF Downloads 148
6081 Bayesian Parameter Inference for Continuous Time Markov Chains with Intractable Likelihood

Authors: Randa Alharbi, Vladislav Vyshemirsky

Abstract:

Systems biology is an important field in science which focuses on studying behaviour of biological systems. Modelling is required to produce detailed description of the elements of a biological system, their function, and their interactions. A well-designed model requires selecting a suitable mechanism which can capture the main features of the system, define the essential components of the system and represent an appropriate law that can define the interactions between its components. Complex biological systems exhibit stochastic behaviour. Thus, using probabilistic models are suitable to describe and analyse biological systems. Continuous-Time Markov Chain (CTMC) is one of the probabilistic models that describe the system as a set of discrete states with continuous time transitions between them. The system is then characterised by a set of probability distributions that describe the transition from one state to another at a given time. The evolution of these probabilities through time can be obtained by chemical master equation which is analytically intractable but it can be simulated. Uncertain parameters of such a model can be inferred using methods of Bayesian inference. Yet, inference in such a complex system is challenging as it requires the evaluation of the likelihood which is intractable in most cases. There are different statistical methods that allow simulating from the model despite intractability of the likelihood. Approximate Bayesian computation is a common approach for tackling inference which relies on simulation of the model to approximate the intractable likelihood. Particle Markov chain Monte Carlo (PMCMC) is another approach which is based on using sequential Monte Carlo to estimate intractable likelihood. However, both methods are computationally expensive. In this paper we discuss the efficiency and possible practical issues for each method, taking into account the computational time for these methods. We demonstrate likelihood-free inference by performing analysing a model of the Repressilator using both methods. Detailed investigation is performed to quantify the difference between these methods in terms of efficiency and computational cost.

Keywords: Approximate Bayesian computation(ABC), Continuous-Time Markov Chains, Sequential Monte Carlo, Particle Markov chain Monte Carlo (PMCMC)

Procedia PDF Downloads 201
6080 Multi-Subpopulation Genetic Algorithm with Estimation of Distribution Algorithm for Textile Batch Dyeing Scheduling Problem

Authors: Nhat-To Huynh, Chen-Fu Chien

Abstract:

Textile batch dyeing scheduling problem is complicated which includes batch formation, batch assignment on machines, batch sequencing with sequence-dependent setup time. Most manufacturers schedule their orders manually that are time consuming and inefficient. More power methods are needed to improve the solution. Motivated by the real needs, this study aims to propose approaches in which genetic algorithm is developed with multi-subpopulation and hybridised with estimation of distribution algorithm to solve the constructed problem for minimising the makespan. A heuristic algorithm is designed and embedded into the proposed algorithms to improve the ability to get out of the local optima. In addition, an empirical study is conducted in a textile company in Taiwan to validate the proposed approaches. The results have showed that proposed approaches are more efficient than simulated annealing algorithm.

Keywords: estimation of distribution algorithm, genetic algorithm, multi-subpopulation, scheduling, textile dyeing

Procedia PDF Downloads 298
6079 Comparison of Heuristic Methods for Solving Traveling Salesman Problem

Authors: Regita P. Permata, Ulfa S. Nuraini

Abstract:

Traveling Salesman Problem (TSP) is the most studied problem in combinatorial optimization. In simple language, TSP can be described as a problem of finding a minimum distance tour to a city, starting and ending in the same city, and exactly visiting another city. In product distribution, companies often get problems in determining the minimum distance that affects the time allocation. In this research, we aim to apply TSP heuristic methods to simulate nodes as city coordinates in product distribution. The heuristics used are sub tour reversal, nearest neighbor, farthest insertion, cheapest insertion, nearest insertion, and arbitrary insertion. We have done simulation nodes using Euclidean distances to compare the number of cities and processing time, thus we get optimum heuristic method. The results show that the optimum heuristic methods are farthest insertion and nearest insertion. These two methods can be recommended to solve product distribution problems in certain companies.

Keywords: Euclidean, heuristics, simulation, TSP

Procedia PDF Downloads 125
6078 Extreme Temperature Forecast in Mbonge, Cameroon Through Return Level Analysis of the Generalized Extreme Value (GEV) Distribution

Authors: Nkongho Ayuketang Arreyndip, Ebobenow Joseph

Abstract:

In this paper, temperature extremes are forecast by employing the block maxima method of the generalized extreme value (GEV) distribution to analyse temperature data from the Cameroon Development Corporation (CDC). By considering two sets of data (raw data and simulated data) and two (stationary and non-stationary) models of the GEV distribution, return levels analysis is carried out and it was found that in the stationary model, the return values are constant over time with the raw data, while in the simulated data the return values show an increasing trend with an upper bound. In the non-stationary model, the return levels of both the raw data and simulated data show an increasing trend with an upper bound. This clearly shows that although temperatures in the tropics show a sign of increase in the future, there is a maximum temperature at which there is no exceedance. The results of this paper are very vital in agricultural and environmental research.

Keywords: forecasting, generalized extreme value (GEV), meteorology, return level

Procedia PDF Downloads 477
6077 Estimation of Carbon Dioxide Absorption in DKI Jakarta Green Space

Authors: Mario Belseran

Abstract:

The issue of climate change become world attention where one of them increase in air temperature due to greenhouse gas emissions. This climate change is caused by gases in the atmosphere, one of which is CO2. DKI Jakarta as the capital has a dense population with a variety of existing land use. Land use that is dominated by settlements resulting in fewer green space, which functions to absorb atmospheric CO2. Image interpretation SPOT-7 is used to determine the greenness level of vegetation on a green space using the vegetation index NDVI, EVI, GNDVI and OSAVI. Measuring the diameter and height of trees were also performed to obtain the value of biomass that will be used as the CO2 absorption value. The CO2 absorption value that spread in Jakarta are classified into three classes: high, medium, and low. The distribution pattern of CO2 absorption value at green space in Jakarta dominance in the medium class with the distribution pattern is located in South Jakarta, East Jakarta, North Jakarta and West Jakarta. The distribution pattern of green space in Jakarta scattered randomly and more dominate in East Jakarta and South Jakarta

Keywords: carbon dioxide, DKI Jakarta, green space, SPOT-7, vegetation index

Procedia PDF Downloads 278
6076 Superordinated Control for Increasing Feed-in Capacity and Improving Power Quality in Low Voltage Distribution Grids

Authors: Markus Meyer, Bastian Maucher, Rolf Witzmann

Abstract:

The ever increasing amount of distributed generation in low voltage distribution grids (mainly PV and micro-CHP) can lead to reverse load flows from low to medium/high voltage levels at times of high feed-in. Reverse load flow leads to rising voltages that may even exceed the limits specified in the grid codes. Furthermore, the share of electrical loads connected to low voltage distribution grids via switched power supplies continuously increases. In combination with inverter-based feed-in, this results in high harmonic levels reducing overall power quality. Especially high levels of third-order harmonic currents can lead to neutral conductor overload, which is even more critical if lines with reduced neutral conductor section areas are used. This paper illustrates a possible concept for smart grids in order to increase the feed-in capacity, improve power quality and to ensure safe operation of low voltage distribution grids at all times. The key feature of the concept is a hierarchically structured control strategy that is run on a superordinated controller, which is connected to several distributed grid analyzers and inverters via broad band powerline (BPL). The strategy is devised to ensure both quick response time as well as the technically and economically reasonable use of the available inverters in the grid (PV-inverters, batteries, stepless line voltage regulators). These inverters are provided with standard features for voltage control, e.g. voltage dependent reactive power control. In addition they can receive reactive power set points transmitted by the superordinated controller. To further improve power quality, the inverters are capable of active harmonic filtering, as well as voltage balancing, whereas the latter is primarily done by the stepless line voltage regulators. By additionally connecting the superordinated controller to the control center of the grid operator, supervisory control and data acquisition capabilities for the low voltage distribution grid are enabled, which allows easy monitoring and manual input. Such a low voltage distribution grid can also be used as a virtual power plant.

Keywords: distributed generation, distribution grid, power quality, smart grid, virtual power plant, voltage control

Procedia PDF Downloads 266
6075 Exploring the Correlation between Population Distribution and Urban Heat Island under Urban Data: Taking Shenzhen Urban Heat Island as an Example

Authors: Wang Yang

Abstract:

Shenzhen is a modern city of China's reform and opening-up policy, the development of urban morphology has been established on the administration of the Chinese government. This city`s planning paradigm is primarily affected by the spatial structure and human behavior. The subjective urban agglomeration center is divided into several groups and centers. In comparisons of this effect, the city development law has better to be neglected. With the continuous development of the internet, extensive data technology has been introduced in China. Data mining and data analysis has become important tools in municipal research. Data mining has been utilized to improve data cleaning such as receiving business data, traffic data and population data. Prior to data mining, government data were collected by traditional means, then were analyzed using city-relationship research, delaying the timeliness of urban development, especially for the contemporary city. Data update speed is very fast and based on the Internet. The city's point of interest (POI) in the excavation serves as data source affecting the city design, while satellite remote sensing is used as a reference object, city analysis is conducted in both directions, the administrative paradigm of government is broken and urban research is restored. Therefore, the use of data mining in urban analysis is very important. The satellite remote sensing data of the Shenzhen city in July 2018 were measured by the satellite Modis sensor and can be utilized to perform land surface temperature inversion, and analyze city heat island distribution of Shenzhen. This article acquired and classified the data from Shenzhen by using Data crawler technology. Data of Shenzhen heat island and interest points were simulated and analyzed in the GIS platform to discover the main features of functional equivalent distribution influence. Shenzhen is located in the east-west area of China. The city’s main streets are also determined according to the direction of city development. Therefore, it is determined that the functional area of the city is also distributed in the east-west direction. The urban heat island can express the heat map according to the functional urban area. Regional POI has correspondence. The research result clearly explains that the distribution of the urban heat island and the distribution of urban POIs are one-to-one correspondence. Urban heat island is primarily influenced by the properties of the underlying surface, avoiding the impact of urban climate. Using urban POIs as analysis object, the distribution of municipal POIs and population aggregation are closely connected, so that the distribution of the population corresponded with the distribution of the urban heat island.

Keywords: POI, satellite remote sensing, the population distribution, urban heat island thermal map

Procedia PDF Downloads 103
6074 Finite Elemental Simulation of the Combined Process of Asymmetric Rolling and Plastic Bending

Authors: A. Pesin, D. Pustovoytov, M. Sverdlik

Abstract:

Traditionally, the need in items represents a large body of rotation (e.g. shrouds of various process units: a converter, a mixer, a scrubber, a steel ladle and etc.) is satisfied by using them at engineering enterprises. At these enterprises large parts of bodies of rotation are made on stamping units or bending and forming machines. In Nosov Magnitogorsk State Technical University in alliance with JSC "Magnitogorsk Metal and Steel Works" there was suggested and implemented the technology for producing such items based on a combination of asymmetric rolling processes and plastic bending under conditions of the plate mill. In this paper, based on finite elemental mathematical simulation in technology of a combined process of asymmetric rolling and bending plastic has been improved. It is shown that for the same curvature along the entire length of the metal sheet it is necessary to introduce additional asymmetry speed when rolling front end and tape trailer. Production of large bodies of rotation at mill 4500 JSC "Magnitogorsk Metal and Steel Works" showed good convergence of theoretical and experimental values of the curvature of the metal. Economic effect obtained more than 1.0 million dollars.

Keywords: asymmetric rolling, plastic bending, combined process, FEM

Procedia PDF Downloads 319
6073 Mixing Behaviors of Shear-Thinning Fluids in Serpentine-Channel Micromixers

Authors: Rei-Tang Tsai, Chih-Yang Wu, Chia-Yuan Chang, Ming-Ying Kuo

Abstract:

This study aims to investigate the mixing behaviors of deionized (DI) water and carboxymethyl cellulose (CMC) solutions in C-shaped serpentine micromixers over a wide range of flow conditions. The flow of CMC solutions exhibits shear-thinning behaviors. Numerical simulations are performed to investigate the effects of the mean flow speed, fluid properties and geometry parameters on flow and mixing in the micromixers with serpentine channel of the same overall channel length. From the results, we can find the following trends. When fluid mixing is dominated by convection, the curvature-induced vortices enhance fluid mixing effectively. The mixing efficiency of a micromixer consisting of semicircular C-shaped repeating units with a smaller center-line radius is better than that of a micromixer consisting of major-segment repeating units with a larger center-line radius. The viscosity of DI water is less than the overall average apparent viscosity of CMC solutions, and so the effect of curvature-induced vortices on fluid mixing in DI water is larger than that in CMC solutions for the cases with the same mean flow speed.

Keywords: curved channel, microfluidics, mixing, non-newtonian fluids, vortex

Procedia PDF Downloads 440
6072 Modelling Hydrological Time Series Using Wakeby Distribution

Authors: Ilaria Lucrezia Amerise

Abstract:

The statistical modelling of precipitation data for a given portion of territory is fundamental for the monitoring of climatic conditions and for Hydrogeological Management Plans (HMP). This modelling is rendered particularly complex by the changes taking place in the frequency and intensity of precipitation, presumably to be attributed to the global climate change. This paper applies the Wakeby distribution (with 5 parameters) as a theoretical reference model. The number and the quality of the parameters indicate that this distribution may be the appropriate choice for the interpolations of the hydrological variables and, moreover, the Wakeby is particularly suitable for describing phenomena producing heavy tails. The proposed estimation methods for determining the value of the Wakeby parameters are the same as those used for density functions with heavy tails. The commonly used procedure is the classic method of moments weighed with probabilities (probability weighted moments, PWM) although this has often shown difficulty of convergence, or rather, convergence to a configuration of inappropriate parameters. In this paper, we analyze the problem of the likelihood estimation of a random variable expressed through its quantile function. The method of maximum likelihood, in this case, is more demanding than in the situations of more usual estimation. The reasons for this lie, in the sampling and asymptotic properties of the estimators of maximum likelihood which improve the estimates obtained with indications of their variability and, therefore, their accuracy and reliability. These features are highly appreciated in contexts where poor decisions, attributable to an inefficient or incomplete information base, can cause serious damages.

Keywords: generalized extreme values, likelihood estimation, precipitation data, Wakeby distribution

Procedia PDF Downloads 136
6071 Patient Safety Culture in Brazilian Hospitals from Nurse's Team Perspective

Authors: Carmen Silvia Gabriel, Dsniele Bernardi da Costa, Andrea Bernardes, Sabrina Elias Mikael, Daniele da Silva Ramos

Abstract:

The goal of this quantitative study is to investigate patient safety culture from the perspective of professional from the hospital nursing team.It was conducted in two Brazilian hospitals,.The sample included 282 nurses Data collection occurred in 2013, through the questionnaire Hospital Survey on Patient Safety Culture.Based on the assessment of the dimensions is stressed that, in the dimension teamwork across hospital units, 69.4% of professionals agree that when a lot of work needs to be done quickly, they work together as a team; about the dimension supervisor/ manager expectations and actions promoting safety, 70.2% agree that their supervisor overlooks patient safety problems.Related to organizational learning and continuous improvement, 56.5% agree that there is evaluation of the effectiveness of the changes after its implementation.On hospital management support for patient safety, 52.8% refer that the actions of hospital management show that patient safety is a top priority.On the overall perception of patient safety, 57.2% disagree that patient safety is never compromised due to higher amount of work to be completed.In what refers to feedback and communication about error, 57.7% refer that always and usually receive such information. Relative to communication openness, 42.9% said they never or rarely feel free to question the decisions / actions of their superiors.On frequency of event reporting, 64.7% said often and always notify events with no damages to patients..About teamwork across hospital units is noted similarity between the percentages of agreement and disagreement, as on the item there is a good cooperation among hospital units that need to work together, that indicates 41.4% and 40.5% respectively.Related to adequacy of professionals, 77.8 % disagree on the existence of sufficient amount of employees to do the job, 52.4% agree that shift changes are problematic for patients. On nonpunitive response to errors, 71.7% indicate that when an event is reported it seems that the focus is on the person.On the patient safety grade of the institution, 41.6 % classified it as very good. it is concluded that there are positive points in the safety culture, and some weaknesses as a punitive culture and impaired patient safety due to work overload .

Keywords: quality of health care, health services evaluation, safety culture, patient safety, nursing team

Procedia PDF Downloads 298
6070 Optimal Opportunistic Maintenance Policy for a Two-Unit System

Authors: Nooshin Salari, Viliam Makis, Jane Doe

Abstract:

This paper presents a maintenance policy for a system consisting of two units. Unit 1 is gradually deteriorating and is subject to soft failure. Unit 2 has a general lifetime distribution and is subject to hard failure. Condition of unit 1 of the system is monitored periodically and it is considered as failed when its deterioration level reaches or exceeds a critical level N. At the failure time of unit 2 system is considered as failed, and unit 2 will be correctively replaced by the next inspection epoch. Unit 1 or 2 are preventively replaced when deterioration level of unit 1 or age of unit 2 exceeds the related preventive maintenance (PM) levels. At the time of corrective or preventive replacement of unit 2, there is an opportunity to replace unit 1 if its deterioration level reaches the opportunistic maintenance (OM) level. If unit 2 fails in an inspection interval, system stops operating although unit 1 has not failed. A mathematical model is derived to find the preventive and opportunistic replacement levels for unit 1 and preventive replacement age for unit 2, that minimize the long run expected average cost per unit time. The problem is formulated and solved in the semi-Markov decision process (SMDP) framework. Numerical example is provided to illustrate the performance of the proposed model and the comparison of the proposed model with an optimal policy without opportunistic maintenance level for unit 1 is carried out.

Keywords: condition-based maintenance, opportunistic maintenance, preventive maintenance, two-unit system

Procedia PDF Downloads 198
6069 Genetic Polymorphisms of the Human Organic Cation Transporter 2 gene, SLC22A2, in the Zulu population

Authors: N. Hoosain, S. Nene, B. Pearce, C. Jacobs, M. Du Plessis, M. Benjeddou

Abstract:

Organic Cation Transporters play a vital role in the absorption, tissue distribution and elimination of various substrates. Numerous studies have suggested that variations in non-synonymous single nucleotide polymorphisms (SNPs) of SLC22A2 could influence an individual’s response to various treatments, including clinically important drugs. This study is the first to determine the baseline frequency distribution for twenty SNPs of SLC22A2in the Zulu population. DNA was collected from 101 unrelated “healthy” Zulu participants. Genotypes of all samples were determined using a multiplex PCR and SNaPshot assay followed by the generation of the haplotype structure. This is the first time that the baseline frequency distribution of SNPs is reported for the Zulu population. Data from this study could be used in in vitro and in vivo pharmacogenetic and pharmacokinetic studies to evaluate the potential role the studied SNPs play in the therapeutic efficacy of clinically important drugs.

Keywords: SLC22A2 gene, SNaPshot assay, PCR, Zulu population

Procedia PDF Downloads 287
6068 Estimating Housing Prices Using Automatic Linear Modeling in the Metropolis of Mashhad, Iran

Authors: Mohammad Rahim Rahnama

Abstract:

Market-transaction price for housing is the main criteria for determining municipality taxes and is determined and announced on an annual basis. Of course, there is a discrepancy between the actual value of transactions in the Bureau of Finance (P for short) or municipality (P´ for short) and the real price on the market (P˝). The present research aims to determine the real price of housing in the metropolis of Mashhad and to pinpoint the price gap with those of the aforementioned apparatuses and identify the factors affecting it. In order to reach this practical objective, Automatic Linear Modeling, which calls for an explanatory research, was utilized. The population of the research consisted of all the residential units in Mashhad, from which 317 residential units were randomly selected. Through cluster sampling, out of the 170 income blocks defined by the municipality, three blocks form high-income (Kosar), middle-income (Elahieh), and low-income (Seyyedi) strata were surveyed using questionnaires during February and March of 2015 and the information regarding the price and specifications of residential units were gathered. In order to estimate the effect of various factors on the price, the relationship between independent variables (8 variables) and the dependent variable of the housing price was calculated using Automatic Linear Modeling in SPSS. The results revealed that the average for housing price index is 788$ per square meter, compared to the Bureau of Finance’s prices which is 10$ and that of municipality’s which is 378$. Correlation coefficient among dependent and independent variables was calculated to be R²=0.81. Out of the eight initial variables, three were omitted. The most influential factor affecting the housing prices is the quality of Quality of construction (Ordinary, Full, Luxury). The least important factor influencing the housing prices is the variable of number of sides. The price gap between low-income (Seyyedi) and middle-income (Elahieh) districts was not confirmed via One-Way ANOVA but their gap with the high-income district (Kosar) was confirmed. It is suggested that city be divided into two low-income and high-income sections, as opposed three, in terms of housing prices.

Keywords: automatic linear modeling, housing prices, Mashhad, Iran

Procedia PDF Downloads 254