Search results for: Rayleigh Distribution
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1992

Search results for: Rayleigh Distribution

1302 Copper Contamination in the Sediments of Northern Kaohsiung Harbor, Taiwan

Authors: Chiu-Wen Chen, Chih-Feng Chen, Cheng-Di Dong

Abstract:

The distribution, enrichment, accumulation, and potential ecological risk of copper (Cu) in the surface sediments of northern Kaohsiung Harbor, Taiwan were investigated. Sediment samples from 12 locations of northern Kaohsiung Harbor were collected and characterized for Cu, aluminum, water content, organic matter, total nitrogen, total phosphorous, total grease and grain size. Results showed that the Cu concentrations varied from 6.9–244 mg/kg with an average of 109±66 mg/kg. The spatial distribution of Cu reveals that the Cu concentration is relatively high in the river mouth region, and gradually diminishes toward the harbor entrance region. This indicates that upstream industrial and municipal wastewater discharges along the river bank are major sources of Cu pollution. Results from the enrichment factor and geo-accumulation index analyses imply that the sediments collected from the river mouth can be characterized between moderate and moderately severe degree enrichment and between none to medium and moderate accumulation of Cu, respectively. However, results of potential ecological risk index indicate that the sediment has low ecological potential risk.

Keywords: Accumulation, ecological risk, enrichment, copper, sediment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1335
1301 An Hybrid Approach for Loss Reduction in Distribution Systems using Harmony Search Algorithm

Authors: R. Srinivasa Rao

Abstract:

Individually Network reconfiguration or Capacitor control perform well in minimizing power loss and improving voltage profile of the distribution system. But for heavy reactive power loads network reconfiguration and for heavy active power loads capacitor placement can not effectively reduce power loss and enhance voltage profiles in the system. In this paper, an hybrid approach that combine network reconfiguration and capacitor placement using Harmony Search Algorithm (HSA) is proposed to minimize power loss reduction and improve voltage profile. The proposed approach is tested on standard IEEE 33 and 16 bus systems. Computational results show that the proposed hybrid approach can minimize losses more efficiently than Network reconfiguration or Capacitor control. The results of proposed method are also compared with results obtained by Simulated Annealing (SA). The proposed method has outperformed in terms of the quality of solution compared to SA.

Keywords: Capacitor Control, Network Reconfiguration, HarmonySearch Algorithm, Loss Reduction, Voltage Profile.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2168
1300 The Effects of Misspecification of Stochastic Processes on Investment Appraisal

Authors: George Yungchih Wang

Abstract:

For decades financial economists have been attempted to determine the optimal investment policy by recognizing the option value embedded in irreversible investment whose project value evolves as a geometric Brownian motion (GBM). This paper aims to examine the effects of the optimal investment trigger and of the misspecification of stochastic processes on investment in real options applications. Specifically, the former explores the consequence of adopting optimal investment rules on the distributions of corporate value under the correct assumption of stochastic process while the latter analyzes the influence on the distributions of corporate value as a result of the misspecification of stochastic processes, i.e., mistaking an alternative process as a GBM. It is found that adopting the correct optimal investment policy may increase corporate value by shifting the value distribution rightward, and the misspecification effect may decrease corporate value by shifting the value distribution leftward. The adoption of the optimal investment trigger has a major impact on investment to such an extent that the downside risk of investment is truncated at the project value of zero, thereby moving the value distributions rightward. The analytical framework is also extended to situations where collection lags are in place, and the result indicates that collection lags reduce the effects of investment trigger and misspecification on investment in an opposite way.

Keywords: GBM, real options, investment trigger, misspecification, collection lags

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513
1299 Retarding Potential Analyzer Design and Result Analysis for Ion Energy Distribution Measurement of the Thruster Plume in the Laboratory

Authors: Ma Ya-li, Tang Fu-jun, Xue Yu-xiong, Chen Yi-feng, Gao Xin, Wang Yi, Tian Kai, Yan Ze-dong

Abstract:

Plasma plume will be produced and arrive at spacecraft when the electric thruster operates on orbit. It-s important to characterize the thruster plasma parameters because the plume has significant effects or hazards on spacecraft sub-systems and parts. Through the ground test data of the desired parameters, the major characteristics of the thruster plume will be achieved. Also it is very important for optimizing design of Ion thruster. Retarding Potential Analyzer (RPA) is an effective instrument for plasma ion energy per unit charge distribution measurement. Special RPA should be designed according to certain plume plasma parameters range and feature. In this paper, major principles usable for good RPA design are discussed carefully. Conform to these principles, a four-grid planar electrostatic energy analyzer RPA was designed to avoid false data, and details were discussed including construction, materials, aperture diameter and so on. At the same time, it was designed more suitable for credible and long-duration measurements in the laboratory. In the end, RPA measurement results in the laboratory were given and discussed.

Keywords: Thruster plume ion energy distributions, retarding potential analyzer, ground test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4008
1298 A Coupled Model for Two-Phase Simulation of a Heavy Water Pressure Vessel Reactor

Authors: Damian Ramajo, Santiago Corzo, Norberto Nigro

Abstract:

A Multi-dimensional computational fluid dynamics (CFD) two-phase model was developed with the aim to simulate the in-core coolant circuit of a pressurized heavy water reactor (PHWR) of a commercial nuclear power plant (NPP). Due to the fact that this PHWR is a Reactor Pressure Vessel type (RPV), three-dimensional (3D) detailed modelling of the large reservoirs of the RPV (the upper and lower plenums and the downcomer) were coupled with an in-house finite volume one-dimensional (1D) code in order to model the 451 coolant channels housing the nuclear fuel. Regarding the 1D code, suitable empirical correlations for taking into account the in-channel distributed (friction losses) and concentrated (spacer grids, inlet and outlet throttles) pressure losses were used. A local power distribution at each one of the coolant channels was also taken into account. The heat transfer between the coolant and the surrounding moderator was accurately calculated using a two-dimensional theoretical model. The implementation of subcooled boiling and condensation models in the 1D code along with the use of functions for representing the thermal and dynamic properties of the coolant and moderator (heavy water) allow to have estimations of the in-core steam generation under nominal flow conditions for a generic fission power distribution. The in-core mass flow distribution results for steady state nominal conditions are in agreement with the expected from design, thus getting a first assessment of the coupled 1/3D model. Results for nominal condition were compared with those obtained with a previous 1/3D single-phase model getting more realistic temperature patterns, also allowing visualize low values of void fraction inside the upper plenum. It must be mentioned that the current results were obtained by imposing prescribed fission power functions from literature. Therefore, results are showed with the aim of point out the potentiality of the developed model.

Keywords: CFD, PHWR, Thermo-hydraulic, Two-phase flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2710
1297 The Influence of Fiber Volume Fraction on Thermal Conductivity of Pultruded Profile

Authors: V. Lukášová, P. Peukert, V. Votrubec

Abstract:

Thermal conductivity in the x, y and z-directions was measured on a pultruded profile that was manufactured by the technology of pulling from glass fibers and a polyester matrix. The results of measurements of thermal conductivity showed considerable variability in different directions. The caused variability in thermal conductivity was expected due fraction variations. The cross-section of the pultruded profile was scanned. An image analysis illustrated an uneven distribution of the fibers and the matrix in the cross-section. The distribution of these inequalities was processed into a Voronoi diagram in the observed area of the pultruded profile cross-section. In order to verify whether the variation of the fiber volume fraction in the pultruded profile can affect its thermal conductivity, the numerical simulations in the ANSYS Fluent were performed. The simulation was based on the geometry reconstructed from image analysis. The aim is to quantify thermal conductivity numerically. Above all, images with different volume fractions were chosen. The results of the measured thermal conductivity were compared with the calculated thermal conductivity. The evaluated data proved a strong correlation between volume fraction and thermal conductivity of the pultruded profile. Based on presented results, a modification of production technology may be proposed.

Keywords: Numerical simulation, pultruded profile, volume fraction, thermal conductivity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1183
1296 Regionalization of IDF Curves with L-Moments for Storm Events

Authors: Noratiqah Mohd Ariff, Abdul Aziz Jemain, Mohd Aftar Abu Bakar

Abstract:

The construction of Intensity-Duration-Frequency (IDF) curves is one of the most common and useful tools in order to design hydraulic structures and to provide a mathematical relationship between rainfall characteristics. IDF curves, especially those in Peninsular Malaysia, are often built using moving windows of rainfalls. However, these windows do not represent the actual rainfall events since the duration of rainfalls is usually prefixed. Hence, instead of using moving windows, this study aims to find regionalized distributions for IDF curves of extreme rainfalls based on storm events. Homogeneity test is performed on annual maximum of storm intensities to identify homogeneous regions of storms in Peninsular Malaysia. The L-moment method is then used to regionalized Generalized Extreme Value (GEV) distribution of these annual maximums and subsequently. IDF curves are constructed using the regional distributions. The differences between the IDF curves obtained and IDF curves found using at-site GEV distributions are observed through the computation of the coefficient of variation of root mean square error, mean percentage difference and the coefficient of determination. The small differences implied that the construction of IDF curves could be simplified by finding a general probability distribution of each region. This will also help in constructing IDF curves for sites with no rainfall station.

Keywords: IDF curves, L-moments, regionalization, storm events.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1714
1295 The Influence of Surface Roughness of Drawbead on Non-Symmetry Deep Drawing Cold Rolled Steel Sheet

Authors: A. Watanapa, S. Torsakul

Abstract:

This study was aimed to explain the influence of surface roughness of the drawbead on non-symmetry deep drawing cold rolled steel sheet to improve the drawability of cold rolled steel sheet. The variables used in this study included semi-circle drawbead with 3 levels of surface roughness which are 6.127 mm Ra, 0.963 mm Ra and 0.152 mm Ra and cold rolled steel sheet according to 3 grades of the JIS standards which are SPCC, SPCE and SPCD with the thickness of 1.0 mm and the blankholder force which is 50% of the drawing force and the depth of 50 mm. According to the test results, when there was the increase in the surface roughness of drawbead, there would be the increase in deep drawing force, especially the SPCC cold rolled steel sheet. This is similar to the increase in the equivalent strain and the wall thickness distribution when the surface roughness of the drawbead increased. It could be concluded that the surface roughness of drawbead has an influence on deep drawing cold rolled steel sheet, especially the drawing force, the equivalent strain and the wall thickness distribution.

Keywords: Drawbead, Deep Drawing, Drawing Force, Equivalent Strain, Surface roughness

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2425
1294 Numerical Investigation on Performance of Expanded Polystyrene Geofoam Block in Protecting Buried Lifeline Structures

Authors: M. Abdollahi, S. N. Moghaddas Tafreshi

Abstract:

Expanded polystyrene (EPS) geofoam is often used in below ground applications in geotechnical engineering. A most recent configuration system implemented in roadways to protect lifelines such as buried pipes, electrical cables and culvert systems could be consisted of two EPS geofoam blocks, “posts” placed on each side of the structure, an EPS block capping, “beam” put atop two posts, and soil cover on the beam. In this configuration, a rectangular void space will be built atop the lifeline. EPS blocks will stand all the imposed vertical forces due to their strength and deformability, thus the lifeline will experience no vertical stress. The present paper describes the results of a numerical study on the post and beam configuration subjected to the static loading. Three-dimensional finite element analysis using ABAQUS software is carried out to investigate the effect of different parameters such as beam thickness, soil thickness over the beam, post height to width ratio, EPS density, and free span between two posts, on the stress distribution and the deflection of the beam. The results show favorable performance of EPS geofoam for protecting sensitive infrastructures.

Keywords: Beam, EPS block, numerical analysis, post, stress distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1152
1293 Determination and Comparison of Fabric Pills Distribution Using Image Processing and Spatial Data Analysis Tools

Authors: Lenka Techniková, Maroš Tunák, Jiří Janáček

Abstract:

This work deals with the determination and comparison of pill patterns in 2 sets of fabric samples which differ in way of pill creation. The first set contains fabric samples with the pills created by simulation on a Martindale abrasion machine, while pills in the second set originated during normal wearing and maintenance. The goal of the study is to determine whether the pattern of the fabric pills created by simulation is the same as the pattern of naturally occurring pills. The system of determination and comparison of the pills is based on image processing and spatial data analysis tools. Firstly, 3D reconstruction of the fabric surfaces with the pills is realized with using a gradient fields method. The gradient fields method creates a 3D fabric surface from a set of 4 images. Thereafter, the pills are detected in 3D fabric surfaces using image-processing tools in the MATLAB software. Determination and comparison of the pills patterns of two sets of fabric samples is based on spatial data analysis using tools in R software.

Keywords: 3D reconstruction of the surface, image analysis tools, distribution of the pills, spatial data analysis tools.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173
1292 Effect of Geographical Co-Ordinates on the Parameters in the Rain Rate Model for Radio Propagation Applications

Authors: Olatinwo M. O., Oyeleke Olaosebikan, David Henry O.

Abstract:

Rain attenuation plays a lot of roles in the design of satellite and terrestrial microwave radio links, hence a good knowledge of its effect is of great interest to Engineers and scientists in that it is often required to give a high level of accuracy of the rainrate distribution that expresses rainrate from the lowest value to the highest. This study proposes a model to express rainrate parameters alpha (α) and beta (β) as a function of geographical location at 0.01% of the time. The tropical locations used in the development of the effect were Ilorin, Ile-Ife, Douala, Dar-es-Selam, Nairobi, Lusaka, and Brazilia.

This expression clearly confirms the variability of rainfall from place to place. When consistency test was carried out using the expression to generate rainrate for each location examined, the result obtained was reliable for rain intensities between 5mm/h and 200mm/h. The variability of α and β with latitude also shows that different latitudes have different cumulative rain distribution. The model proposed in this study would be one of the useful tools to Radio Engineers since the precipitation effect in the design of satellite and terrestrial microwave radio links is among the factors to consider when designing communication systems.

Keywords: Rain rate, attenuation, geographical location.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1714
1291 A Speeded up Robust Scale-Invariant Feature Transform Currency Recognition Algorithm

Authors: Daliyah S. Aljutaili, Redna A. Almutlaq, Suha A. Alharbi, Dina M. Ibrahim

Abstract:

All currencies around the world look very different from each other. For instance, the size, color, and pattern of the paper are different. With the development of modern banking services, automatic methods for paper currency recognition become important in many applications like vending machines. One of the currency recognition architecture’s phases is Feature detection and description. There are many algorithms that are used for this phase, but they still have some disadvantages. This paper proposes a feature detection algorithm, which merges the advantages given in the current SIFT and SURF algorithms, which we call, Speeded up Robust Scale-Invariant Feature Transform (SR-SIFT) algorithm. Our proposed SR-SIFT algorithm overcomes the problems of both the SIFT and SURF algorithms. The proposed algorithm aims to speed up the SIFT feature detection algorithm and keep it robust. Simulation results demonstrate that the proposed SR-SIFT algorithm decreases the average response time, especially in small and minimum number of best key points, increases the distribution of the number of best key points on the surface of the currency. Furthermore, the proposed algorithm increases the accuracy of the true best point distribution inside the currency edge than the other two algorithms.

Keywords: Currency recognition, feature detection and description, SIFT algorithm, SURF algorithm, speeded up and robust features.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 864
1290 Impact of Zn/Cr Ratio on ZnCrOx-SAPO-34 Bifunctional Catalyst for Direct Conversion of Syngas to Light Olefins

Authors: Yuxuan Huang, Weixin Qian, Hongfang Ma, Haitao Zhang, Weiyong Ying

Abstract:

Light olefins are important building blocks for chemical industry. Direct conversion of syngas to light olefins has been investigated for decades. Meanwhile, the limit for light olefins selectivity described by Anderson-Schulz-Flory (ASF) distribution model is still a great challenge to conventional Fischer-Tropsch synthesis. The emerging strategy called oxide-zeolite concept (OX-ZEO) is a promising way to get rid of this limit. ZnCrOx was prepared by co-precipitation method and (NH4)2CO3 was used as precipitant. SAPO-34 was prepared by hydrothermal synthesis, and Tetraethylammonium hydroxide (TEAOH) was used as template, while silica sol, pseudo-boehmite, and phosphoric acid were Al, Si and P source, respectively. The bifunctional catalyst was prepared by mechanical mixing of ZnCrOx and SAPO-34. Catalytic reactions were carried out under H2/CO=2, 380 ℃, 1 MPa and 6000 mL·gcat-1·h-1 in a fixed-bed reactor with a quartz lining. Catalysts were characterized by XRD, N2 adsorption-desorption, NH3-TPD, H2-TPR, and CO-TPD. The addition of Al as structure promoter enhances CO conversion and selectivity to light olefins. Zn/Cr ratio, which decides the active component content and chemisorption property of the catalyst, influences CO conversion and selectivity to light olefins at the same time. C2-4= distribution of 86% among hydrocarbons at CO conversion of 14% was reached when Zn/Cr=1.5.

Keywords: Light olefins, OX-ZEO, syngas, ZnCrOx.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1021
1289 Stochastic Modeling for Parameters of Modified Car-Following Model in Area-Based Traffic Flow

Authors: N. C. Sarkar, A. Bhaskar, Z. Zheng

Abstract:

The driving behavior in area-based (i.e., non-lane based) traffic is induced by the presence of other individuals in the choice space from the driver’s visual perception area. The driving behavior of a subject vehicle is constrained by the potential leaders and leaders are frequently changed over time. This paper is to determine a stochastic model for a parameter of modified intelligent driver model (MIDM) in area-based traffic (as in developing countries). The parametric and non-parametric distributions are presented to fit the parameters of MIDM. The goodness of fit for each parameter is measured in two different ways such as graphically and statistically. The quantile-quantile (Q-Q) plot is used for a graphical representation of a theoretical distribution to model a parameter and the Kolmogorov-Smirnov (K-S) test is used for a statistical measure of fitness for a parameter with a theoretical distribution. The distributions are performed on a set of estimated parameters of MIDM. The parameters are estimated on the real vehicle trajectory data from India. The fitness of each parameter with a stochastic model is well represented. The results support the applicability of the proposed modeling for parameters of MIDM in area-based traffic flow simulation.

Keywords: Area-based traffic, car-following model, micro-simulation, stochastic modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 774
1288 CFD Prediction of the Round Elbow Fitting Loss Coefficient

Authors: Ana Paula P. dos Santos, Claudia R. Andrade, Edson L. Zaparoli

Abstract:

Pressure loss in ductworks is an important factor to be considered in design of engineering systems such as power-plants, refineries, HVAC systems to reduce energy costs. Ductwork can be composed by straight ducts and different types of fittings (elbows, transitions, converging and diverging tees and wyes). Duct fittings are significant sources of pressure loss in fluid distribution systems. Fitting losses can be even more significant than equipment components such as coils, filters, and dampers. At the present work, a conventional 90o round elbow under turbulent incompressible airflow is studied. Mass, momentum, and k-e turbulence model equations are solved employing the finite volume method. The SIMPLE algorithm is used for the pressure-velocity coupling. In order to validate the numerical tool, the elbow pressure loss coefficient is determined using the same conditions to compare with ASHRAE database. Furthermore, the effect of Reynolds number variation on the elbow pressure loss coefficient is investigated. These results can be useful to perform better preliminary design of air distribution ductworks in air conditioning systems.

Keywords: Duct fitting, Pressure loss, Elbow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4851
1287 Investigation of Fire Damaged Reinforced Concrete Walls with Axial Force

Authors: Hyun Ah Yoon, Ji Yeon Kang, Hee Sun Kim, Yeong Soo Shin

Abstract:

Reinforced concrete (RC) shear wall system of residential buildings is popular in South Korea. RC walls are subjected to axial forces in common and the effect of axial forces on the strength loss of the fire damaged walls has not been investigated. This paper aims at investigating temperature distribution on fire damaged concrete walls having different axial loads. In the experiments, a variable of specimens is axial force ratio. RC walls are fabricated with 150mm of wall thicknesses, 750mm of lengths and 1,300mm of heights having concrete strength of 24MPa. After curing, specimens are heated on one surface with ISO-834 standard time-temperature curve for 2 hours and temperature distributions during the test are measured using thermocouples inside the walls. The experimental results show that the temperature of the RC walls exposed to fire increases as axial force ratio increases. To verify the experiments, finite element (FE) models are generated for coupled temperature-structure analyses. The analytical results of thermal behaviors are in good agreement with the experimental results. The predicted displacement of the walls decreases when the axial force increases. 

Keywords: Axial force ratio, coupled analysis, fire, reinforced concrete wall, temperature distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738
1286 Analysis of Dust Particles in Snow Cover in the Surroundings of the City of Ostrava: Particle Size Distribution, Zeta Potential and Heavy Metal Content

Authors: Roman Marsalek

Abstract:

In this paper, snow samples containing dust particles from several sampling points around the city of Ostrava were analyzed. The pH values of sampled snow were measured and solid particles analyzed. Particle size, zeta potential and content of selected heavy metals were determined in solid particles. The pH values of most samples lay in the slightly acid region. Mean values of particle size ranged from 290.5 to 620.5 nm. Zeta potential values varied between -5 and -26.5 mV. The following heavy metal concentration ranges were found: copper 0.08-0.75 mg/g, lead 0.05-0.9 mg/g, manganese 0.45-5.9 mg/g and iron 25.7-280.46 mg/g. The highest values of copper and lead were found in the vicinity of busy crossroads, and on the contrary, the highest levels of manganese and iron were detected close to a large steelworks. The proportion between pH values, zeta potentials, particle sizes and heavy metal contents was established. Zeta potential decreased with rising pH values and, simultaneously, heavy metal content in solid particles increased. At the same time, higher metal content corresponded to lower particle size.

Keywords: Dust, snow, zeta potential, particles size distribution, heavy metals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1983
1285 MONPAR - A Page Replacement Algorithm for a Spatiotemporal Database

Authors: U. Kalay, O. Kalıpsız

Abstract:

For a spatiotemporal database management system, I/O cost of queries and other operations is an important performance criterion. In order to optimize this cost, an intense research on designing robust index structures has been done in the past decade. With these major considerations, there are still other design issues that deserve addressing due to their direct impact on the I/O cost. Having said this, an efficient buffer management strategy plays a key role on reducing redundant disk access. In this paper, we proposed an efficient buffer strategy for a spatiotemporal database index structure, specifically indexing objects moving over a network of roads. The proposed strategy, namely MONPAR, is based on the data type (i.e. spatiotemporal data) and the structure of the index structure. For the purpose of an experimental evaluation, we set up a simulation environment that counts the number of disk accesses while executing a number of spatiotemporal range-queries over the index. We reiterated simulations with query sets with different distributions, such as uniform query distribution and skewed query distribution. Based on the comparison of our strategy with wellknown page-replacement techniques, like LRU-based and Prioritybased buffers, we conclude that MONPAR behaves better than its competitors for small and medium size buffers under all used query-distributions.

Keywords: Buffer Management, Spatiotemporal databases.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1476
1284 User-Driven Product Line Engineering for Assembling Large Families of Software

Authors: Zhaopeng Xuan, Yuan Bian, C. Cailleaux, Jing Qin, S. Traore

Abstract:

Traditional software engineering allows engineers to propose to their clients multiple specialized software distributions assembled from a shared set of software assets. The management of these assets however requires a trade-off between client satisfaction and software engineering process. Clients have more and more difficult to find a distribution or components based on their needs from all of distributed repositories.

This paper proposes a software engineering for a user-driven software product line in which engineers define a Feature Model but users drive the actual software distribution on demand. This approach makes the user become final actor as a release manager in software engineering process, increasing user product satisfaction and simplifying user operations to find required components. In addition, it provides a way for engineers to manage and assembly large software families.

As a proof of concept, a user-driven software product line is implemented for Eclipse, an integrated development environment. An Eclipse feature model is defined, which is exposed to users on a cloud-based built platform from which clients can download individualized Eclipse distributions.

Keywords: Software Product Line, Model-driven Development, Reverse Engineering and Refactoring, Agile Method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1831
1283 The Effect of Loperamide and Fentanyl on the Distribution Kinetics of Verapamil in the Lung and Brain in Sprague Dawley Rats

Authors: Iman A. Elkiweri, Ph.D, Martha C. Tissot van Patot, Ph.D., Yan Ling Zhang, Ph.D., Uwe Christians, Ph.D., Thomas K. Henthorn, M.D.,

Abstract:

Verapamil has been shown to inhibit fentanyl uptake in vitro and is a potent P-glycoprotein inhibitor. Tissue partitioning of loperamide, a commercially available opioid, is closely controlled by the P-gp efflux transporter. The following studies were designed to evaluate the effect of opioids on verapamil partitioning in the lung and brain, in vivo. Opioid (fentanyl or loperamide) was administered by intravenous infusion to Sprague Dawley rats alone or in combination with verapamil and plasma, with lung and brain tissues were collected at 1, 5, 6, 8, 10 and 60 minutes. Drug dispositions were modeled by recirculatory pharmacokinetic models. Fentanyl slightly increased the verapamil lung (PL) partition coefficient yet decreased the brain (PB) partition coefficient. Furthermore, loperamide significantly increased PLand PB. Fentanyl reduced the verapamil volume of distribution (V1) and verapamil elimination clearance (ClE). Fentanyl decreased verapamil brain partitioning, yet increased verapamil lung partitioning. Also, loperamide increased lung and brain partitioning in vivo. These results suggest that verapamil and fentanyl may be substrates of an unidentified inward transporter in brain tissue and confirm that verapamil and loperamide are substrates of the efflux transporter P-gp.

Keywords: Efflux transporter, elimination clearance, partition coefficient, verapamil

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793
1282 Analytical Solution of the Boundary Value Problem of Delaminated Doubly-Curved Composite Shells

Authors: András Szekrényes

Abstract:

Delamination is one of the major failure modes in laminated composite structures. Delamination tips are mostly captured by spatial numerical models in order to predict crack growth. This paper presents some mechanical models of delaminated composite shells based on shallow shell theories. The mechanical fields are based on a third-order displacement field in terms of the through-thickness coordinate of the laminated shell. The undelaminated and delaminated parts are captured by separate models and the continuity and boundary conditions are also formulated in a general way providing a large size boundary value problem. The system of differential equations is solved by the state space method for an elliptic delaminated shell having simply supported edges. The comparison of the proposed and a numerical model indicates that the primary indicator of the model is the deflection, the secondary is the widthwise distribution of the energy release rate. The model is promising and suitable to determine accurately the J-integral distribution along the delamination front. Based on the proposed model it is also possible to develop finite elements which are able to replace the computationally expensive spatial models of delaminated structures.

Keywords: J-integral, Lévy method, third-order shell theory, state space solution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 599
1281 Analyzing of Temperature-Dependent Thermal Conductivity Effect in the Numerical Modeling of Fin-Tube Radiators: Introduction of a New Method

Authors: Farzad Bazdidi-Tehrani, Mohammad Hadi Kamrava

Abstract:

In all industries which are related to heat, suitable thermal ranges are defined for each device to operate well. Consideration of these limits requires a thermal control unit beside the main system. The Satellite Thermal Control Unit exploits from different methods and facilities individually or mixed. For enhancing heat transfer between primary surface and the environment, utilization of radiating extended surfaces are common. Especially for large temperature differences; variable thermal conductivity has a strong effect on performance of such a surface .In most literatures, thermo-physical properties, such as thermal conductivity, are assumed as constant. However, in some recent researches the variation of these parameters is considered. This may be helpful for the evaluation of fin-s temperature distribution in relatively large temperature differences. A new method is introduced to evaluate temperature-dependent thermal conductivity values. The finite volume method is employed to simulate numerically the temperature distribution in a space radiating fin. The present modeling is carried out for Aluminum as fin material and compared with previous method. The present results are also compared with those of two other analytical methods and good agreement is shown.

Keywords: Variable thermal conductivity, New method, Finitevolume method, Combined heat transfer, Extended Surface

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2329
1280 A Data Driven Approach for the Degradation of a Lithium-Ion Battery Based on Accelerated Life Test

Authors: Alyaa M. Younes, Nermine Harraz, Mohammad H. Elwany

Abstract:

Lithium ion batteries are currently used for many applications including satellites, electric vehicles and mobile electronics. Their ability to store relatively large amount of energy in a limited space make them most appropriate for critical applications. Evaluation of the life of these batteries and their reliability becomes crucial to the systems they support. Reliability of Li-Ion batteries has been mainly considered based on its lifetime. However, another important factor that can be considered critical in many applications such as in electric vehicles is the cycle duration. The present work presents the results of an experimental investigation on the degradation behavior of a Laptop Li-ion battery (type TKV2V) and the effect of applied load on the battery cycle time. The reliability was evaluated using an accelerated life test. Least squares linear regression with median rank estimation was used to estimate the Weibull distribution parameters needed for the reliability functions estimation. The probability density function, failure rate and reliability function under each of the applied loads were evaluated and compared. An inverse power model is introduced that can predict cycle time at any stress level given.

Keywords: Accelerated life test, inverse power law, lithium ion battery, reliability evaluation, Weibull distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 831
1279 Distributed Generator Placement for Loss Reduction and Improvement in Reliability

Authors: Priyanka Paliwal, N.P. Patidar

Abstract:

Distributed Power generation has gained a lot of attention in recent times due to constraints associated with conventional power generation and new advancements in DG technologies .The need to operate the power system economically and with optimum levels of reliability has further led to an increase in interest in Distributed Generation. However it is important to place Distributed Generator on an optimum location so that the purpose of loss minimization and voltage regulation is dully served on the feeder. This paper investigates the impact of DG units installation on electric losses, reliability and voltage profile of distribution networks. In this paper, our aim would be to find optimal distributed generation allocation for loss reduction subjected to constraint of voltage regulation in distribution network. The system is further analyzed for increased levels of Reliability. Distributed Generator offers the additional advantage of increase in reliability levels as suggested by the improvements in various reliability indices such as SAIDI, CAIDI and AENS. Comparative studies are performed and related results are addressed. An analytical technique is used in order to find the optimal location of Distributed Generator. The suggested technique is programmed under MATLAB software. The results clearly indicate that DG can reduce the electrical line loss while simultaneously improving the reliability of the system.

Keywords: AENS, CAIDI, Distributed Generation, lossreduction, Reliability, SAIDI

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3101
1278 An Improved Algorithm for Channel Estimations of OFDM System based Pilot Signal

Authors: Ahmed N. H. Alnuaimy, Mahamod Ismail, Mohd. A. M. Ali, Kasmiran Jumari, Ayman A. El-Saleh

Abstract:

This paper presents a new algorithm for the channel estimation of the OFDM system based on a pilot signal for the new generation of high data rate communication systems. In orthogonal frequency division multiplexing (OFDM) systems over fast-varying fading channels, channel estimation and tracking is generally carried out by transmitting known pilot symbols in given positions of the frequency-time grid. In this paper, we propose to derive an improved algorithm based on the calculation of the mean and the variance of the adjacent pilot signals for a specific distribution of the pilot signals in the OFDM frequency-time grid then calculating of the entire unknown channel coefficients from the equation of the mean and the variance. Simulation results shows that the performance of the OFDM system increase as the length of the channel increase where the accuracy of the estimated channel will be increased using this low complexity algorithm, also the number of the pilot signal needed to be inserted in the OFDM signal will be reduced which lead to increase in the throughput of the signal over the OFDM system in compared with other type of the distribution such as Comb type and Block type channel estimation.

Keywords: Channel estimation, orthogonal frequency divisionmultiplexing (OFDM), comb type channel estimation, block typechannel estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1416
1277 An Implementation of Fuzzy Logic Technique for Prediction of the Power Transformer Faults

Authors: Omar M. Elmabrouk., Roaa Y. Taha., Najat M. Ebrahim, Sabbreen A. Mohammed

Abstract:

Power transformers are the most crucial part of power electrical system, distribution and transmission grid. This part is maintained using predictive or condition-based maintenance approach. The diagnosis of power transformer condition is performed based on Dissolved Gas Analysis (DGA). There are five main methods utilized for analyzing these gases. These methods are International Electrotechnical Commission (IEC) gas ratio, Key Gas, Roger gas ratio, Doernenburg, and Duval Triangle. Moreover, due to the importance of the transformers, there is a need for an accurate technique to diagnose and hence predict the transformer condition. The main objective of this technique is to avoid the transformer faults and hence to maintain the power electrical system, distribution and transmission grid. In this paper, the DGA was utilized based on the data collected from the transformer records available in the General Electricity Company of Libya (GECOL) which is located in Benghazi-Libya. The Fuzzy Logic (FL) technique was implemented as a diagnostic approach based on IEC gas ratio method. The FL technique gave better results and approved to be used as an accurate prediction technique for power transformer faults. Also, this technique is approved to be a quite interesting for the readers and the concern researchers in the area of FL mathematics and power transformer.

Keywords: Fuzzy logic, dissolved gas-in-oil analysis, DGA, prediction, power transformer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1357
1276 Improvement of Parallel Compressor Model in Dealing Outlet Unequal Pressure Distribution

Authors: Kewei Xu, Jens Friedrich, Kevin Dwinger, Wei Fan, Xijin Zhang

Abstract:

Parallel Compressor Model (PCM) is a simplified approach to predict compressor performance with inlet distortions. In PCM calculation, it is assumed that the sub-compressors’ outlet static pressure is uniform and therefore simplifies PCM calculation procedure. However, if the compressor’s outlet duct is not long and straight, such assumption frequently induces error ranging from 10% to 15%. This paper provides a revised calculation method of PCM that can correct the error. The revised method employs energy equation, momentum equation and continuity equation to acquire needed parameters and replace the equal static pressure assumption. Based on the revised method, PCM is applied on two compression system with different blades types. The predictions of their performance in non-uniform inlet conditions are yielded through the revised calculation method and are employed to evaluate the method’s efficiency. Validating the results by experimental data, it is found that although little deviation occurs, calculated result agrees well with experiment data whose error ranges from 0.1% to 3%. Therefore, this proves the revised calculation method of PCM possesses great advantages in predicting the performance of the distorted compressor with limited exhaust duct.

Keywords: Parallel Compressor Model (PCM), Revised Calculation Method, Inlet Distortion, Outlet Unequal Pressure Distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1688
1275 The Impacts of Local Decision Making on Customisation Process Speed across Distributed Boundaries: A Case Study

Authors: A. M. Qahtani, G. B. Wills, A. M. Gravell

Abstract:

Communicating and managing customers’ requirements in software development projects play a vital role in the software development process. While it is difficult to do so locally, it is even more difficult to communicate these requirements over distributed boundaries and to convey them to multiple distribution customers. This paper discusses the communication of multiple distribution customers’ requirements in the context of customised software products. The main purpose is to understand the challenges of communicating and managing customisation requirements across distributed boundaries. We propose a model for Communicating Customisation Requirements of Multi-Clients in a Distributed Domain (CCRD). Thereafter, we evaluate that model by presenting the findings of a case study conducted with a company with customisation projects for 18 distributed customers. Then, we compare the outputs of the real case process and the outputs of the CCRD model using simulation methods. Our conjecture is that the CCRD model can reduce the challenge of communication requirements over distributed organisational boundaries, and the delay in decision making and in the entire customisation process time.

Keywords: Customisation Software Products, Global Software Engineering, Local Decision Making, Requirement Engineering, Simulation Model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897
1274 Application of Gamma Frailty Model in Survival of Liver Cirrhosis Patients

Authors: Elnaz Saeedi, Jamileh Abolaghasemi, Mohsen Nasiri Tousi, Saeedeh Khosravi

Abstract:

Goals and Objectives: A typical analysis of survival data involves the modeling of time-to-event data, such as the time till death. A frailty model is a random effect model for time-to-event data, where the random effect has a multiplicative influence on the baseline hazard function. This article aims to investigate the use of gamma frailty model with concomitant variable in order to individualize the prognostic factors that influence the liver cirrhosis patients’ survival times. Methods: During the one-year study period (May 2008-May 2009), data have been used from the recorded information of patients with liver cirrhosis who were scheduled for liver transplantation and were followed up for at least seven years in Imam Khomeini Hospital in Iran. In order to determine the effective factors for cirrhotic patients’ survival in the presence of latent variables, the gamma frailty distribution has been applied. In this article, it was considering the parametric model, such as Exponential and Weibull distributions for survival time. Data analysis is performed using R software, and the error level of 0.05 was considered for all tests. Results: 305 patients with liver cirrhosis including 180 (59%) men and 125 (41%) women were studied. The age average of patients was 39.8 years. At the end of the study, 82 (26%) patients died, among them 48 (58%) were men and 34 (42%) women. The main cause of liver cirrhosis was found hepatitis 'B' with 23%, followed by cryptogenic with 22.6% were identified as the second factor. Generally, 7-year’s survival was 28.44 months, for dead patients and for censoring was 19.33 and 31.79 months, respectively. Using multi-parametric survival models of progressive and regressive, Exponential and Weibull models with regard to the gamma frailty distribution were fitted to the cirrhosis data. In both models, factors including, age, bilirubin serum, albumin serum, and encephalopathy had a significant effect on survival time of cirrhotic patients. Conclusion: To investigate the effective factors for the time of patients’ death with liver cirrhosis in the presence of latent variables, gamma frailty model with parametric distributions seems desirable.

Keywords: Frailty model, latent variables, liver cirrhosis, parametric distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1058
1273 A New Approach for Image Segmentation using Pillar-Kmeans Algorithm

Authors: Ali Ridho Barakbah, Yasushi Kiyoki

Abstract:

This paper presents a new approach for image segmentation by applying Pillar-Kmeans algorithm. This segmentation process includes a new mechanism for clustering the elements of high-resolution images in order to improve precision and reduce computation time. The system applies K-means clustering to the image segmentation after optimized by Pillar Algorithm. The Pillar algorithm considers the pillars- placement which should be located as far as possible from each other to withstand against the pressure distribution of a roof, as identical to the number of centroids amongst the data distribution. This algorithm is able to optimize the K-means clustering for image segmentation in aspects of precision and computation time. It designates the initial centroids- positions by calculating the accumulated distance metric between each data point and all previous centroids, and then selects data points which have the maximum distance as new initial centroids. This algorithm distributes all initial centroids according to the maximum accumulated distance metric. This paper evaluates the proposed approach for image segmentation by comparing with K-means and Gaussian Mixture Model algorithm and involving RGB, HSV, HSL and CIELAB color spaces. The experimental results clarify the effectiveness of our approach to improve the segmentation quality in aspects of precision and computational time.

Keywords: Image segmentation, K-means clustering, Pillaralgorithm, color spaces.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3372