Search results for: agent based simulation
11946 A New Approach to Design Low Power Continues-Time Sigma-Delta Modulators
Authors: E. Farshidi
Abstract:
This paper presents the design of a low power second-order continuous-time sigma-delta modulator for low power applications. The loop filter of this modulator has been implemented based on the nonlinear transconductance-capacitor (Gm-C) by employing current-mode technique. The nonlinear transconductance uses floating gate MOS (FG-MOS) transistors that operate in weak inversion region. The proposed modulator features low power consumption (<80uW), low supply voltage (1V) and 62dB dynamic range. Simulation results by HSPICE confirm that it is very suitable for low power biomedical instrumentation designs.
Keywords: Sigma-delta, modulator, Current-mode, Nonlinear Transconductance, FG-MOS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 152511945 On the Modeling and State Estimation for Dynamic Power System
Authors: A. Thabet, M. Boutayeb, M. N. Abdelkrim
Abstract:
This paper investigates a method for the state estimation of nonlinear systems described by a class of differential-algebraic equation (DAE) models using the extended Kalman filter. The method involves the use of a transformation from a DAE to ordinary differential equation (ODE). A relevant dynamic power system model using decoupled techniques will be proposed. The estimation technique consists of a state estimator based on the EKF technique as well as the local stability analysis. High performances are illustrated through a simulation study applied on IEEE 13 buses test system.
Keywords: Power system, Dynamic decoupled model, Extended Kalman Filter, Convergence analysis, Time computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 274311944 Method and Experiment of Fabricating and Cutting the Burr for Y Shape Nanochannel
Authors: Zone-Ching Lin, Hao-Yuan Jheng, Shih-Hung Ma
Abstract:
The present paper proposes using atomic force microscopy (AFM) and the concept of specific down force energy (SDFE) to establish a method for fabricating and cutting the burr for Y shape nanochannel on silicon (Si) substrate. For fabricating Y shape nanochannel, it first makes the experimental cutting path planning for fabricating Y shape nanochannel until the fifth cutting layer. Using the constant down force by AFM and SDFE theory and following the experimental cutting path planning, the cutting depth and width of each pass of Y shape nanochannel can be predicted by simulation. The paper plans the path for cutting the burr at the edge of Y shape nanochannel. Then, it carries out cutting the burr along the Y nanochannel edge by using a smaller down force. The height of standing burr at the edge is required to be below the set value of 0.54 nm. The results of simulation and experiment of fabricating and cutting the burr for Y shape nanochannel is further compared.Keywords: Atomic force microscopy, nanochannel, specific down force energy, Y shape, burr, silicon.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 108811943 Generator Capability Curve Constraint for PSO Based Optimal Power Flow
Authors: Mat Syai'in, Adi Soeprijanto, Takashi Hiyama
Abstract:
An optimal power flow (OPF) based on particle swarm optimization (PSO) was developed with more realistic generator security constraint using the capability curve instead of only Pmin/Pmax and Qmin/Qmax. Neural network (NN) was used in designing digital capability curve and the security check algorithm. The algorithm is very simple and flexible especially for representing non linear generation operation limit near steady state stability limit and under excitation operation area. In effort to avoid local optimal power flow solution, the particle swarm optimization was implemented with enough widespread initial population. The objective function used in the optimization process is electric production cost which is dominated by fuel cost. The proposed method was implemented at Java Bali 500 kV power systems contain of 7 generators and 20 buses. The simulation result shows that the combination of generator power output resulted from the proposed method was more economic compared with the result using conventional constraint but operated at more marginal operating point.Keywords: Optimal Power Flow, Generator Capability Curve, Particle Swarm Optimization, Neural Network
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 258611942 Transport of Analytes under Mixed Electroosmotic and Pressure Driven Flow of Power Law Fluid
Authors: Naren Bag, S. Bhattacharyya, Partha P. Gopmandal
Abstract:
In this study, we have analyzed the transport of analytes under a two dimensional steady incompressible flow of power-law fluids through rectangular nanochannel. A mathematical model based on the Cauchy momentum-Nernst-Planck-Poisson equations is considered to study the combined effect of mixed electroosmotic (EO) and pressure driven (PD) flow. The coupled governing equations are solved numerically by finite volume method. We have studied extensively the effect of key parameters, e.g., flow behavior index, concentration of the electrolyte, surface potential, imposed pressure gradient and imposed electric field strength on the net average flow across the channel. In addition to study the effect of mixed EOF and PD on the analyte distribution across the channel, we consider a nonlinear model based on general convective-diffusion-electromigration equation. We have also presented the retention factor for various values of electrolyte concentration and flow behavior index.Keywords: Electric double layer, finite volume method, flow behavior index, mixed electroosmotic/pressure driven flow, Non-Newtonian power-law fluids, numerical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 121111941 Low Overhead Dynamic Channel Selection with Cluster-Based Spatial-Temporal Station Reporting in Wireless Networks
Authors: Zeyad Abdelmageid, Xianbin Wang
Abstract:
Choosing the operational channel for a WLAN access point (AP) in WLAN networks has been a static channel assignment process initiated by the user during the deployment process of the AP, which fails to cope with the dynamic conditions of the assigned channel at the station side afterwards. However, the dramatically growing number of Wi-Fi APs and stations operating in the unlicensed band has led to dynamic, distributed and often severe interference. This highlights the urgent need for the AP to dynamically select the best overall channel of operation for the basic service set (BSS) by considering the distributed and changing channel conditions at all stations. Consequently, dynamic channel selection algorithms which consider feedback from the station side have been developed. Despite the significant performance improvement, existing channel selection algorithms suffer from very high feedback overhead. Feedback latency from the STAs, due the high overhead, can cause the eventually selected channel to no longer be optimal for operation due to the dynamic sharing nature of the unlicensed band. This has inspired us to develop our own dynamic channel selection algorithm with reduced overhead through the proposed low-overhead, cluster-based station reporting mechanism. The main idea behind the cluster-based station reporting is the observation that STAs which are very close to each other tend to have very similar channel conditions. Instead of requesting each STA to report on every candidate channel while causing high overhead, the AP divides STAs into clusters then assigns each STA in each cluster one channel to report feedback on. With proper design of the cluster based reporting, the AP does not lose any information about the channel conditions at the station side while reducing feedback overhead. The simulation results show equal performance and at times better performance with a fraction of the overhead. We believe that this algorithm has great potential in designing future dynamic channel selection algorithms with low overhead.
Keywords: Channel assignment, Wi-Fi networks, clustering, DBSCAN, overhead.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39911940 Decision Support System for Hospital Selection in Emergency Medical Services: A Discrete Event Simulation Approach
Authors: D. Tedesco, G. Feletti, P. Trucco
Abstract:
The present study aims to develop a Decision Support System (DSS) to support operational decisions in Emergency Medical Service (EMS) systems regarding the assignment of medical emergency requests to Emergency Departments (ED). This problem is called “hospital selection” and concerns the definition of policies for the selection of the ED to which patients who require further treatment are transported by ambulance. The employed research methodology consists of a first phase of review of the technical-scientific literature concerning DSSs to support the EMS management and, in particular, the hospital selection decision. From the literature analysis, it emerged that current studies mainly focused on the EMS phases related to the ambulance service and consider a process that ends when the ambulance is available after completing a mission. Therefore, all the ED-related issues are excluded and considered as part of a separate process. Indeed, the most studied hospital selection policy turned out to be proximity, thus allowing to minimize the travelling time and to free-up the ambulance in the shortest possible time. The purpose of the present study consists in developing an optimization model for assigning medical emergency requests to the EDs also considering the expected time performance in the subsequent phases of the process, such as the case mix, the expected service throughput times, and the operational capacity of different EDs in hospitals. To this end, a Discrete Event Simulation (DES) model was created to compare different hospital selection policies. The model was implemented with the AnyLogic software and finally validated on a realistic case. The hospital selection policy that returned the best results was the minimization of the Time To Provider (TTP), considered as the time from the beginning of the ambulance journey to the ED at the beginning of the clinical evaluation by the doctor. Finally, two approaches were further compared: a static approach, based on a retrospective estimation of the TTP, and a dynamic approach, focused on a predictive estimation of the TTP which is determined with a constantly updated Winters forecasting model. Findings reveal that considering the minimization of TTP is the best hospital selection policy. It allows to significantly reducing service throughput times in the ED with a negligible increase in travel time. Furthermore, an immediate view of the saturation state of the ED is produced and the case mix present in the ED structures (i.e., the different triage codes) is considered, as different severity codes correspond to different service throughput times. Besides, the use of a predictive approach is certainly more reliable in terms on TTP estimation, than a retrospective approach. These considerations can support decision-makers in introducing different hospital selection policies to enhance EMSs performance.
Keywords: Emergency medical services, hospital selection, discrete event simulation, forecast model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24811939 Thermal Evaluation of Printed Circuit Board Design Options and Voids in Solder Interface by a Simulation Tool
Authors: B. Arzhanov, A. Correia, P. Delgado, J. Meireles
Abstract:
Quad Flat No-Lead (QFN) packages have become very popular for turners, converters and audio amplifiers, among others applications, needing efficient power dissipation in small footprints. Since semiconductor junction temperature (TJ) is a critical parameter in the product quality. And to ensure that die temperature does not exceed the maximum allowable TJ, a thermal analysis conducted in an earlier development phase is essential to avoid repeated re-designs process with huge losses in cost and time. A simulation tool capable to estimate die temperature of components with QFN package was developed. Allow establish a non-empirical way to define an acceptance criterion for amount of voids in solder interface between its exposed pad and Printed Circuit Board (PCB) to be applied during industrialization process, and evaluate the impact of PCB designs parameters. Targeting PCB layout designer as an end user for the application, a user-friendly interface (GUI) was implemented allowing user to introduce design parameters in a convenient and secure way and hiding all the complexity of finite element simulation process. This cost effective tool turns transparent a simulating process and provides useful outputs after acceptable time, which can be adopted by PCB designers, preventing potential risks during the design stage and make product economically efficient by not oversizing it. This article gathers relevant information related to the design and implementation of the developed tool, presenting a parametric study conducted with it. The simulation tool was experimentally validated using a Thermal-Test-Chip (TTC) in a QFN open-cavity, in order to measure junction temperature (TJ) directly on the die under controlled and knowing conditions. Providing a short overview about standard thermal solutions and impacts in exposed pad packages (i.e. QFN), accurately describe the methods and techniques that the system designer should use to achieve optimum thermal performance, and demonstrate the effect of system-level constraints on the thermal performance of the design.Keywords: Quad Flat No-Lead packages, exposed pads, junction temperature, thermal management and measurements.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 193311938 Simulation of Enhanced Biomass Gasification for Hydrogen Production using iCON
Authors: Mohd K. Yunus, Murni M. Ahmad, Abrar Inayat, Suzana Yusup
Abstract:
Due to the environmental and price issues of current energy crisis, scientists and technologists around the globe are intensively searching for new environmentally less-impact form of clean energy that will reduce the high dependency on fossil fuel. Particularly hydrogen can be produced from biomass via thermochemical processes including pyrolysis and gasification due to the economic advantage and can be further enhanced through in-situ carbon dioxide removal using calcium oxide. This work focuses on the synthesis and development of the flowsheet for the enhanced biomass gasification process in PETRONAS-s iCON process simulation software. This hydrogen prediction model is conducted at operating temperature between 600 to 1000oC at atmospheric pressure. Effects of temperature, steam-to-biomass ratio and adsorbent-to-biomass ratio were studied and 0.85 mol fraction of hydrogen is predicted in the product gas. Comparisons of the results are also made with experimental data from literature. The preliminary economic potential of developed system is RM 12.57 x 106 which equivalent to USD 3.77 x 106 annually shows economic viability of this process.Keywords: Biomass, Gasification, Hydrogen, iCON.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 261211937 Optimization of Multi-Zone Unconventional (Shale) Gas Reservoir Using Hydraulic Fracturing Technique
Authors: F.C. Amadi, G. C. Enyi, G. G. Nasr
Abstract:
Hydraulic fracturing is one of the most important stimulation techniques available to the petroleum engineer to extract hydrocarbons in tight gas sandstones. It allows more oil and gas production in tight reservoirs as compared to conventional means. The main aim of the study is to optimize the hydraulic fracturing as technique and for this purpose three multi-zones layer formation is considered and fractured contemporaneously. The three zones are named as Zone1 (upper zone), Zone2 (middle zone) and Zone3 (lower zone) respectively and they all occur in shale rock. Simulation was performed with Mfrac integrated software which gives a variety of 3D fracture options. This simulation process yielded an average fracture efficiency of 93.8%for the three respective zones and an increase of the average permeability of the rock system. An average fracture length of 909 ft with net height (propped height) of 210 ft (average) was achieved. Optimum fracturing results was also achieved with maximum fracture width of 0.379 inches at an injection rate of 13.01 bpm with 17995 Mscf of gas production.Keywords: Hydraulic fracturing, Mfrac, Optimisation, Tight reservoir.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 196711936 Coding based Synchronization Algorithm for Secondary Synchronization Channel in WCDMA
Authors: Deng Liao, Dongyu Qiu, Ahmed K. Elhakeem
Abstract:
A new code synchronization algorithm is proposed in this paper for the secondary cell-search stage in wideband CDMA systems. Rather than using the Cyclically Permutable (CP) code in the Secondary Synchronization Channel (S-SCH) to simultaneously determine the frame boundary and scrambling code group, the new synchronization algorithm implements the same function with less system complexity and less Mean Acquisition Time (MAT). The Secondary Synchronization Code (SSC) is redesigned by splitting into two sub-sequences. We treat the information of scrambling code group as data bits and use simple time diversity BCH coding for further reliability. It avoids involved and time-costly Reed-Solomon (RS) code computations and comparisons. Analysis and simulation results show that the Synchronization Error Rate (SER) yielded by the new algorithm in Rayleigh fading channels is close to that of the conventional algorithm in the standard. This new synchronization algorithm reduces system complexities, shortens the average cell-search time and can be implemented in the slot-based cell-search pipeline. By taking antenna diversity and pipelining correlation processes, the new algorithm also shows its flexible application in multiple antenna systems.Keywords: WCDMA cell-search, synchronization algorithm, secondary synchronization channel, antenna diversity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 239811935 A Framework of Monte Carlo Simulation for Examining the Uncertainty-Investment Relationship
Authors: George Yungchih Wang
Abstract:
This paper argues that increased uncertainty, in certain situations, may actually encourage investment. Since earlier studies mostly base their arguments on the assumption of geometric Brownian motion, the study extends the assumption to alternative stochastic processes, such as mixed diffusion-jump, mean-reverting process, and jump amplitude process. A general approach of Monte Carlo simulation is developed to derive optimal investment trigger for the situation that the closed-form solution could not be readily obtained under the assumption of alternative process. The main finding is that the overall effect of uncertainty on investment is interpreted by the probability of investing, and the relationship appears to be an invested U-shaped curve between uncertainty and investment. The implication is that uncertainty does not always discourage investment even under several sources of uncertainty. Furthermore, high-risk projects are not always dominated by low-risk projects because the high-risk projects may have a positive realization effect on encouraging investment.Keywords: real options, geometric Brownian motion, mixeddiffusion-jump process, mean- reverting process, jump amplitudeprocess
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155711934 An Improved Adaptive Dot-Shape Beamforming Algorithm Research on Frequency Diverse Array
Authors: Yanping Liao, Zenan Wu, Ruigang Zhao
Abstract:
Frequency diverse array (FDA) beamforming is a technology developed in recent years, and its antenna pattern has a unique angle-distance-dependent characteristic. However, the beam is always required to have strong concentration, high resolution and low sidelobe level to form the point-to-point interference in the concentrated set. In order to eliminate the angle-distance coupling of the traditional FDA and to make the beam energy more concentrated, this paper adopts a multi-carrier FDA structure based on proposed power exponential frequency offset to improve the array structure and frequency offset of the traditional FDA. The simulation results show that the beam pattern of the array can form a dot-shape beam with more concentrated energy, and its resolution and sidelobe level performance are improved. However, the covariance matrix of the signal in the traditional adaptive beamforming algorithm is estimated by the finite-time snapshot data. When the number of snapshots is limited, the algorithm has an underestimation problem, which leads to the estimation error of the covariance matrix to cause beam distortion, so that the output pattern cannot form a dot-shape beam. And it also has main lobe deviation and high sidelobe level problems in the case of limited snapshot. Aiming at these problems, an adaptive beamforming technique based on exponential correction for multi-carrier FDA is proposed to improve beamforming robustness. The steps are as follows: first, the beamforming of the multi-carrier FDA is formed under linear constrained minimum variance (LCMV) criteria. Then the eigenvalue decomposition of the covariance matrix is performed to obtain the diagonal matrix composed of the interference subspace, the noise subspace and the corresponding eigenvalues. Finally, the correction index is introduced to exponentially correct the small eigenvalues of the noise subspace, improve the divergence of small eigenvalues in the noise subspace, and improve the performance of beamforming. The theoretical analysis and simulation results show that the proposed algorithm can make the multi-carrier FDA form a dot-shape beam at limited snapshots, reduce the sidelobe level, improve the robustness of beamforming, and have better performance.
Keywords: Multi-carrier frequency diverse array, adaptive beamforming, correction index, limited snapshot, robust.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 68811933 Fuzzy Logic Based Improved Range Free Localization for Wireless Sensor Networks
Authors: Ashok Kumar, Vinod Kumar
Abstract:
Wireless Sensor Networks (WSNs) are used to monitor/observe vast inaccessible regions through deployment of large number of sensor nodes in the sensing area. For majority of WSN applications, the collected data needs to be combined with geographic information of its origin to make it useful for the user; information received from remote Sensor Nodes (SNs) that are several hops away from base station/sink is meaningless without knowledge of its source. In addition to this, location information of SNs can also be used to propose/develop new network protocols for WSNs to improve their energy efficiency and lifetime. In this paper, range free localization protocols for WSNs have been proposed. The proposed protocols are based on weighted centroid localization technique, where the edge weights of SNs are decided by utilizing fuzzy logic inference for received signal strength and link quality between the nodes. The fuzzification is carried out using (i) Mamdani, (ii) Sugeno, and (iii) Combined Mamdani Sugeno fuzzy logic inference. Simulation results demonstrate that proposed protocols provide better accuracy in node localization compared to conventional centroid based localization protocols despite presence of unintentional radio frequency interference from radio frequency (RF) sources operating in same frequency band.
Keywords: localization, range free, received signal strength, link quality indicator, Mamdani fuzzy logic inference, Sugeno fuzzy logic inference.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 263611932 Computer - based Systems for High Speed Vessels Navigators – Engineers Training
Authors: D. E. Gourgoulis, C. G. Yakinthos, M. G. Vassiliadou
Abstract:
With high speed vessels getting ever more sophisti-cated, travelling at higher and higher speeds and operating in With high speed vessels getting ever more sophisticated, travelling at higher and higher speeds and operating in areas of high maritime traffic density, training becomes of the highest priority to ensure that safety levels are maintained, and risks are adequately mitigated. Training onboard the actual craft on the actual route still remains the most effective way for crews to gain experience. However, operational experience and incidents during the last 10 years demonstrate the need for supplementary training whether in the area of simulation or man to man, man/ machine interaction. Training and familiarisation of the crew is the most important aspect in preventing incidents. The use of simulator, computer and web based training systems in conjunction with onboard training focusing on critical situations will improve the man machine interaction and thereby reduce the risk of accidents. Today, both ship simulator and bridge teamwork courses are now becoming the norm in order to improve further emergency response and crisis management skills. One of the main causes of accidents is the human factor. An efficient way to reduce human errors is to provide high-quality training to the personnel and to select the navigators carefully.areas of high maritime traffic density, training becomes of the highest priority to ensure that safety levels are maintained, and risks are adequately mitigated. Training onboard the actual craft on the actual route still remains the most effective way for crews to gain experience. How-ever, operational experience and incidents during the last 10 years demonstrate the need for supplementary training whether in the area of simulation or man to man, man/ machine interaction. Training and familiarisation of the crew is the most important aspect in preventing incidents. The use of simulator, computer and web based training systems in conjunction with onboard training focusing on critical situations will improve the man machine interaction and thereby reduce the risk of accidents. Today, both ship simulator and bridge teamwork courses are now becoming the norm in order to improve further emergency response and crisis management skills. One of the main causes of accidents is the human factor. An efficient way to reduce human errors is to provide high-quality training to the person-nel and to select the navigators carefully. KeywordsCBT - WBT systems, Human factors.Keywords: CBT - WBT systems, Human factors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 153311931 CFD Analysis of Natural Ventilation Behaviour in Four Sided Wind Catcher
Authors: M. Hossein Ghadiri, Mohd Farid Mohamed, N. Lukman N. Ibrahim
Abstract:
Wind catchers are traditional natural ventilation systems attached to buildings in order to ventilate the indoor air. The most common type of wind catcher is four sided one which is capable to catch wind in all directions. CFD simulation is the perfect way to evaluate the wind catcher performance. The accuracy of CFD results is the issue of concern, so sensitivity analyses is crucial to find out the effect of different settings of CFD on results. This paper presents a series of 3D steady RANS simulations for a generic isolated four-sided wind catcher attached to a room subjected to wind direction ranging from 0º to 180º with an interval of 45º. The CFD simulations are validated with detailed wind tunnel experiments. The influence of an extensive range of computational parameters is explored in this paper, including the resolution of the computational grid, the size of the computational domain and the turbulence model. This study found that CFD simulation is a reliable method for wind catcher study, but it is less accurate in prediction of models with non perpendicular wind directions.Keywords: Wind catcher, CFD, natural ventilation, sensitivity study.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 270011930 Sliding Mode Control Based on Backstepping Approach for an UAV Type-Quadrotor
Authors: H. Bouadi, M. Bouchoucha, M. Tadjine
Abstract:
In this paper; we are interested principally in dynamic modelling of quadrotor while taking into account the high-order nonholonomic constraints in order to develop a new control scheme as well as the various physical phenomena, which can influence the dynamics of a flying structure. These permit us to introduce a new state-space representation. After, the use of Backstepping approach for the synthesis of tracking errors and Lyapunov functions, a sliding mode controller is developed in order to ensure Lyapunov stability, the handling of all system nonlinearities and desired tracking trajectories. Finally simulation results are also provided in order to illustrate the performances of the proposed controller.
Keywords: Dynamic modeling, nonholonomic constraints, Backstepping, sliding mode.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 591211929 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics
Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi
Abstract:
Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.Keywords: Reactive rotational molding, free surface flows, simulation, smoothed particle hydrodynamics, surface tension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 107911928 Detection of Breast Cancer in the JPEG2000 Domain
Authors: Fayez M. Idris, Nehal I. AlZubaidi
Abstract:
Breast cancer detection techniques have been reported to aid radiologists in analyzing mammograms. We note that most techniques are performed on uncompressed digital mammograms. Mammogram images are huge in size necessitating the use of compression to reduce storage/transmission requirements. In this paper, we present an algorithm for the detection of microcalcifications in the JPEG2000 domain. The algorithm is based on the statistical properties of the wavelet transform that the JPEG2000 coder employs. Simulation results were carried out at different compression ratios. The sensitivity of this algorithm ranges from 92% with a false positive rate of 4.7 down to 66% with a false positive rate of 2.1 using lossless compression and lossy compression at a compression ratio of 100:1, respectively.Keywords: Breast cancer, JPEG2000, mammography, microcalcifications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 158211927 Finite Element Modeling of the Mechanical Behavior of Municipal Solid Waste Incineration Bottom Ash with the Mohr-Coulomb Model
Authors: Le Ngoc Hung, Abriak Nor Edine, Binetruy Christophe, Benzerzour Mahfoud, Shahrour Isam, Patrice Rivard
Abstract:
Bottom ash from Municipal Solid Waste Incineration (MSWI) can be viewed as a typical granular material because these industrial by-products result from the incineration of various domestic wastes. MSWI bottom ash is mainly used in road engineering in substitution of the traditional natural aggregates. As the characterization of their mechanical behavior is essential in order to use them, specific studies have been led over the past few years. In the first part of this paper, the mechanical behavior of MSWI bottom ash is studied with triaxial tests. After, analysis of the experiment results, the simulation of triaxial tests is carried out by using the software package CESAR-LCPC. As the first approach in modeling of this new class material, the Mohr-Coulomb model was chosen to describe the evolution of material under the influence of external mechanical actions.Keywords: Bottom ash, granular material, triaxial test, mechanical behavior, simulation, Mohr-Coulomb model, CESARLCPC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 190211926 Design Optimization for Efficient Erbium-Doped Fiber Amplifiers
Authors: Parekhan M. Aljaff, Banaz O. Rasheed
Abstract:
The exact gain shape profile of erbium doped fiber amplifiers (EDFA`s) are depends on fiber length and Er3 ion densities. This paper optimized several of erbium doped fiber parameters to obtain high performance characteristic at pump wavelengths of λp= 980 nm and λs= 1550 nm for three different pump powers. The maximum gain obtained for pump powers (10, 30 and 50mw) is nearly (19, 30 and 33 dB) at optimizations. The required numerical aperture NA to obtain maximum gain becomes less when pump power increased. The amplifier gain is increase when Er+3doped near the center of the fiber core. The simulation has been done by using optisystem 5.0 software (CAD for Photonics, a license product of a Canadian based company) at 2.5 Gbps.Keywords: EDFA, Erbium Doped Fiber, optimization OpticalAmplifiers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 349611925 Non-Sensitive Solutions in Multi-Objective Optimization of a Solar Photovoltaic/Thermal(PV/T) Air Collector
Authors: F. Sarhaddi, S. Farahat, M .A. Alavi, F. Sobhnamayan
Abstract:
In this paper, an attempt has been made to obtain nonsensitive solutions in the multi-objective optimization of a photovoltaic/thermal (PV/T) air collector. The selected objective functions are overall energy efficiency and exergy efficiency. Improved thermal, electrical and exergy models are used to calculate the thermal and electrical parameters, overall energy efficiency, exergy components and exergy efficiency of a typical PV/T air collector. A computer simulation program is also developed. The results of numerical simulation are in good agreement with the experimental measurements noted in the previous literature. Finally, multi-objective optimization has been carried out under given climatic, operating and design parameters. The optimized ranges of inlet air velocity, duct depth and the objective functions in optimal Pareto front have been obtained. Furthermore, non-sensitive solutions from energy or exergy point of view in the results of multi-objective optimization have been shown.Keywords: Solar photovoltaic thermal (PV/T) air collector, Overall energy efficiency, Exergy efficiency, Multi-objectiveoptimization, Sensitivity analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 216411924 SVM-Based Detection of SAR Images in Partially Developed Speckle Noise
Authors: J. P. Dubois, O. M. Abdul-Latif
Abstract:
Support Vector Machine (SVM) is a statistical learning tool that was initially developed by Vapnik in 1979 and later developed to a more complex concept of structural risk minimization (SRM). SVM is playing an increasing role in applications to detection problems in various engineering problems, notably in statistical signal processing, pattern recognition, image analysis, and communication systems. In this paper, SVM was applied to the detection of SAR (synthetic aperture radar) images in the presence of partially developed speckle noise. The simulation was done for single look and multi-look speckle models to give a complete overlook and insight to the new proposed model of the SVM-based detector. The structure of the SVM was derived and applied to real SAR images and its performance in terms of the mean square error (MSE) metric was calculated. We showed that the SVM-detected SAR images have a very low MSE and are of good quality. The quality of the processed speckled images improved for the multi-look model. Furthermore, the contrast of the SVM detected images was higher than that of the original non-noisy images, indicating that the SVM approach increased the distance between the pixel reflectivity levels (the detection hypotheses) in the original images.Keywords: Least Square-Support Vector Machine, SyntheticAperture Radar. Partially Developed Speckle, Multi-Look Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 154211923 Determination of Safety Distance Around Gas Pipelines Using Numerical Methods
Authors: Omid Adibi, Nategheh Najafpour, Bijan Farhanieh, Hossein Afshin
Abstract:
Energy transmission pipelines are one of the most vital parts of each country which several strict laws have been conducted to enhance the safety of these lines and their vicinity. One of these laws is the safety distance around high pressure gas pipelines. Safety distance refers to the minimum distance from the pipeline where people and equipment do not confront with serious damages. In the present study, safety distance around high pressure gas transmission pipelines were determined by using numerical methods. For this purpose, gas leakages from cracked pipeline and created jet fires were simulated as continuous ignition, three dimensional, unsteady and turbulent cases. Numerical simulations were based on finite volume method and turbulence of flow was considered using k-ω SST model. Also, the combustion of natural gas and air mixture was applied using the eddy dissipation method. The results show that, due to the high pressure difference between pipeline and environment, flow chocks in the cracked area and velocity of the exhausted gas reaches to sound speed. Also, analysis of the incident radiation results shows that safety distances around 42 inches high pressure natural gas pipeline based on 5 and 15 kW/m2 criteria are 205 and 272 meters, respectively.
Keywords: Gas pipelines, incident radiation, numerical simulation, safety distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 113911922 Design and Manufacturing of a Propeller for Axial-Flow Fan
Authors: D. Almazo, M. Toledo, C. Rodríguez
Abstract:
This work presents a methodology for the design and manufacture of propellers oriented to the experimental verification of theoretical results based on the combined model. The design process begins by using algorithms in Matlab which output data contain the coordinates of the points that define the blade airfoils, in this case the NACA 6512 airfoil was used. The modeling for the propeller blade was made in NX7, through the imported files in Matlab and with the help of surfaces. Later, the hub and the clamps were also modeled. Finally, NX 7 also made possible to create post-processed files to the required machine. It is possible to find the block of numbers with G & M codes about the type of driver on the machine. The file extension is .ptp. These files made possible to manufacture the blade, and the hub of the propeller.Keywords: Airfoil, CAM, manufacturing, mathematical algorithm, numeric control, propeller design, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 387611921 A Strategy of Direct Power Control for PWM Rectifier Reducing Ripple in Instantaneous Power
Authors: T. Mohammed Chikouche, K. Hartani
Abstract:
In order to solve the instantaneous power ripple and achieve better performance of direct power control (DPC) for a three-phase PWM rectifier, a control method is proposed in this paper. This control method is applied to overcome the instantaneous power ripple, to eliminate line current harmonics and therefore reduce the total harmonic distortion and to improve the power factor. A switching table is based on the analysis on the change of instantaneous active and reactive power, to select the optimum switching state of the three-phase PWM rectifier. The simulation result shows feasibility of this control method.
Keywords: Power quality, direct power control, power ripple, switching table, unity power factor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 117211920 Enhanced Frame-based Video Coding to Support Content-based Functionalities
Authors: Prabhudev Hosur, Rolando Carrasco
Abstract:
This paper presents the enhanced frame-based video coding scheme. The input source video to the enhanced frame-based video encoder consists of a rectangular-size video and shapes of arbitrarily-shaped objects on video frames. The rectangular frame texture is encoded by the conventional frame-based coding technique and the video object-s shape is encoded using the contour-based vertex coding. It is possible to achieve several useful content-based functionalities by utilizing the shape information in the bitstream at the cost of a very small overhead to the bitrate.
Keywords: Video coding, content-based, hyper video, interactivity, shape coding, polygon.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166811919 Combined Fuzzy and Predictive Controller for Unity Power Factor Converter
Authors: Abdelhalim Kessal
Abstract:
This paper treats a design of combined control of a single phase power factor correction (PFC). The strategy of the proposed control is based on two parts, the first, for the outer loop (DC output regulated voltage), and the second govern the input current of the converter in order to achieve a sinusoidal form in phase with the grid voltage. Two kinds of regulators are used, Fuzzy controller for the outer loop and predictive controller for the inner loop. The controllers are verified and discussed through simulation under MATLAB/Simulink platform. Also an experimental confirmation is applied. Results present a high dynamic performance under various parameters changes.Keywords: Boost converter, harmonic distortion, Fuzzy, prediction, unity power factor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 116311918 Delay-Distribution-Dependent Stability Criteria for BAM Neural Networks with Time-Varying Delays
Authors: J.H. Park, S. Lakshmanan, H.Y. Jung, S.M. Lee
Abstract:
This paper is concerned with the delay-distributiondependent stability criteria for bidirectional associative memory (BAM) neural networks with time-varying delays. Based on the Lyapunov-Krasovskii functional and stochastic analysis approach, a delay-probability-distribution-dependent sufficient condition is derived to achieve the globally asymptotically mean square stable of the considered BAM neural networks. The criteria are formulated in terms of a set of linear matrix inequalities (LMIs), which can be checked efficiently by use of some standard numerical packages. Finally, a numerical example and its simulation is given to demonstrate the usefulness and effectiveness of the proposed results.Keywords: BAM neural networks, Probabilistic time-varying delays, Stability criteria.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 142511917 Hydrodynamic Performance of a Moored Barge in Irregular Wave
Authors: Srinivasan Chandrasekaran, Shihas A. Khader
Abstract:
Motion response of floating structures is of great concern in marine engineering. Nonlinearity is an inherent property of any floating bodies subjected to irregular waves. These floating structures are continuously subjected to environmental loadings from wave, current, wind etc. This can result in undesirable motions of the vessel which may challenge the operability. For a floating body to remain in its position, it should be able to induce a restoring force when displaced. Mooring is provided to enable this restoring force. This paper discusses the hydrodynamic performance and motion characteristics of an 8 point spread mooring system applied to a pipe laying barge operating in the West African sea. The modelling of the barge is done using a computer aided-design (CAD) software RHINOCEROS. Irregular waves are generated using a suitable wave spectrum. Both frequency domain and time domain analysis is done. Numerical simulations based on potential theory are carried out to find the responses and hydrodynamic performance of the barge in both free floating as well as moored conditions. Initially, potential flow frequency domain analysis is done to obtain the Response Amplitude Operator (RAO) which gives an idea about the structural motion in free floating state. RAOs for different wave headings are analyzed. In the following step, a time domain analysis is carried out to obtain the responses of the structure in the moored condition. In this study, wave induced motions are only taken into consideration. Wind and current loads are ruled out and shall be included in further studies. For the current study, 2000 seconds simulation is taken. The results represent wave induced motion responses, mooring line tensions and identify critical mooring lines.
Keywords: Irregular wave, moored barge, time domain analysis, numerical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2766