Search results for: Moderate random particle swarm optimization method
9238 An Efficient Ant Colony Optimization Algorithm for Multiobjective Flow Shop Scheduling Problem
Authors: Ahmad Rabanimotlagh
Abstract:
In this paper an ant colony optimization algorithm is developed to solve the permutation flow shop scheduling problem. In the permutation flow shop scheduling problem which has been vastly studied in the literature, there are a set of m machines and a set of n jobs. All the jobs are processed on all the machines and the sequence of jobs being processed is the same on all the machines. Here this problem is optimized considering two criteria, makespan and total flow time. Then the results are compared with the ones obtained by previously developed algorithms. Finally it is visible that our proposed approach performs best among all other algorithms in the literature.Keywords: Scheduling, Flow shop, Ant colony optimization, Makespan, Flow time
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24079237 A Novel Reversible Watermarking Method based on Adaptive Thresholding and Companding Technique
Authors: Nisar Ahmed Memon
Abstract:
Embedding and extraction of a secret information as well as the restoration of the original un-watermarked image is highly desirable in sensitive applications like military, medical, and law enforcement imaging. This paper presents a novel reversible data-hiding method for digital images using integer to integer wavelet transform and companding technique which can embed and recover the secret information as well as can restore the image to its pristine state. The novel method takes advantage of block based watermarking and iterative optimization of threshold for companding which avoids histogram pre and post-processing. Consequently, it reduces the associated overhead usually required in most of the reversible watermarking techniques. As a result, it keeps the distortion small between the marked and the original images. Experimental results show that the proposed method outperforms the existing reversible data hiding schemes reported in the literature.Keywords: Adaptive Thresholding, Companding Technique, Integer Wavelet Transform, Reversible Watermarking
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18549236 Surrogate based Evolutionary Algorithm for Design Optimization
Authors: Maumita Bhattacharya
Abstract:
Optimization is often a critical issue for most system design problems. Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, finding optimal solution to complex high dimensional, multimodal problems often require highly computationally expensive function evaluations and hence are practically prohibitive. The Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model presented in our earlier work [14] reduced computation time by controlled use of meta-models to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the meta-model are generated from a single uniform model. Situations like model formation involving variable input dimensions and noisy data certainly can not be covered by this assumption. In this paper we present an enhanced version of DAFHEA that incorporates a multiple-model based learning approach for the SVM approximator. DAFHEA-II (the enhanced version of the DAFHEA framework) also overcomes the high computational expense involved with additional clustering requirements of the original DAFHEA framework. The proposed framework has been tested on several benchmark functions and the empirical results illustrate the advantages of the proposed technique.Keywords: Evolutionary algorithm, Fitness function, Optimization, Meta-model, Stochastic method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15619235 A New Stabilizing GPC for Nonminimum Phase LTI Systems Using Time Varying Weighting
Authors: Mahdi Yaghobi, Mohammad Haeri
Abstract:
In this paper, we show that the stability can not be achieved with current stabilizing MPC methods for some unstable processes. Hence we present a new method for stabilizing these processes. The main idea is to use a new time varying weighted cost function for traditional GPC. This stabilizes the closed loop system without adding soft or hard constraint in optimization problem. By studying different examples it is shown that using the proposed method, the closed-loop stability of unstable nonminimum phase process is achieved.Keywords: GPC, Stability, Varying Weighting Coefficients.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14489234 Enhanced Interference Management Technique for Multi-Cell Multi-Antenna System
Authors: Simon E. Uguru, Victor E. Idigo, Obinna S. Oguejiofor, Naveed Nawaz
Abstract:
As the deployment of the Fifth Generation (5G) mobile communication networks take shape all over the world, achieving spectral efficiency, energy efficiency, and dealing with interference are among the greatest challenges encountered so far. The aim of this study is to mitigate inter-cell interference (ICI) in a multi-cell multi-antenna system while maximizing the spectral efficiency of the system. In this study, a system model was devised that showed a miniature representation of a multi-cell multi-antenna system. Based on this system model, a convex optimization problem was formulated to maximize the spectral efficiency of the system while mitigating the ICI. This optimization problem was solved using CVX, which is a modeling system for constructing and solving discipline convex programs. The solutions to the optimization problem are sub-optimal coordinated beamformers. These coordinated beamformers direct each data to the served user equipments (UEs) in each cell without interference during downlink transmission, thereby maximizing the system-wide spectral efficiency.
Keywords: coordinated beamforming, convex optimization, inter-cell interference, multi-antenna, multi-cell, spectral efficiency
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4189233 Effect of Changing Iron Content and Excitation Frequency on Magnetic Particle Imaging Signal: A Comparative Study of Synomag® Nanoparticles
Authors: Kalthoum Riahi, Max T. Rietberg, Javier Perez y Perez, Corné Dijkstra, Bennie ten Haken, Lejla Alic
Abstract:
Magnetic nanoparticles (MNPs) are widely used to facilitate magnetic particle imaging (MPI) which has the potential to become the leading diagnostic instrument for biomedical imaging. This comparative study assesses the effects of changing iron content and excitation frequency on point-spread function (PSF) representing the effect of magnetization reversal. PSF is quantified by features of interest for MPI: i.e., drive field amplitude and full-width-at-half-maximum (FWHM). A superparamagnetic quantifier (SPaQ) is used to assess differential magnetic susceptibility of two commercially available MNPs: Synomag®-D50 and Synomag®-D70. For both MNPs, the signal output depends on increase in drive field frequency and amount of iron-oxide, which might be hampering the sensitivity of MPI systems that perform on higher frequencies. Nevertheless, there is a clear potential of Synomag®-D for a stable MPI resolution, especially in case of 70 nm version, that is independent of either drive field frequency or amount of iron-oxide.
Keywords: Magnetic nanoparticles, MNPs, Differential magnetic susceptibility, DMS, Magnetic particle imaging, MPI, magnetic relaxation, Synomag®-D.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6619232 An EWMA p Chart Based On Improved Square Root Transformation
Authors: S. Sukparungsee
Abstract:
Generally, the traditional Shewhart p chart has been developed by for charting the binomial data. This chart has been developed using the normal approximation with condition as low defect level and the small to moderate sample size. In real applications, however, are away from these assumptions due to skewness in the exact distribution. In this paper, a modified Exponentially Weighted Moving Average (EWMA) control chat for detecting a change in binomial data by improving square root transformations, namely ISRT p EWMA control chart. The numerical results show that ISRT p EWMA chart is superior to ISRT p chart for small to moderate shifts, otherwise, the latter is better for large shifts.
Keywords: Number of defects, Exponentially Weighted Moving Average, Average Run Length, Square root transformations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24639231 Preparation of ATO Conductive Particles with Narrow Size Distribution
Authors: Yueying Wu, Fengzhu Lv, Yihe Zhang, Zixian Xu
Abstract:
Antimosy-doped tin oxide (ATO) particles were prepared via chemical coprecipitation and reverse emulsion. The size and size distribution of ATO particles were obviously decreased via reverse microemulsion method. At the relatively high yield the ATO particles were nearly spherical in shape, meanwhile the crystalline structure and excellent conductivity were reserved, which could satisfy the requirement as composite fillers, such as dielectric filler of polyimide film.Keywords: ATO particle, Conductivity, Distribution, Reverse emulsion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18659230 Lane Detection Using Labeling Based RANSAC Algorithm
Authors: Yeongyu Choi, Ju H. Park, Ho-Youl Jung
Abstract:
In this paper, we propose labeling based RANSAC algorithm for lane detection. Advanced driver assistance systems (ADAS) have been widely researched to avoid unexpected accidents. Lane detection is a necessary system to assist keeping lane and lane departure prevention. The proposed vision based lane detection method applies Canny edge detection, inverse perspective mapping (IPM), K-means algorithm, mathematical morphology operations and 8 connected-component labeling. Next, random samples are selected from each labeling region for RANSAC. The sampling method selects the points of lane with a high probability. Finally, lane parameters of straight line or curve equations are estimated. Through the simulations tested on video recorded at daytime and nighttime, we show that the proposed method has better performance than the existing RANSAC algorithm in various environments.
Keywords: Canny edge detection, k-means algorithm, RANSAC, inverse perspective mapping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11759229 On Discretization of Second-order Derivatives in Smoothed Particle Hydrodynamics
Authors: R. Fatehi, M.A. Fayazbakhsh, M.T. Manzari
Abstract:
Discretization of spatial derivatives is an important issue in meshfree methods especially when the derivative terms contain non-linear coefficients. In this paper, various methods used for discretization of second-order spatial derivatives are investigated in the context of Smoothed Particle Hydrodynamics. Three popular forms (i.e. "double summation", "second-order kernel derivation", and "difference scheme") are studied using one-dimensional unsteady heat conduction equation. To assess these schemes, transient response to a step function initial condition is considered. Due to parabolic nature of the heat equation, one can expect smooth and monotone solutions. It is shown, however in this paper, that regardless of the type of kernel function used and the size of smoothing radius, the double summation discretization form leads to non-physical oscillations which persist in the solution. Also, results show that when a second-order kernel derivative is used, a high-order kernel function shall be employed in such a way that the distance of inflection point from origin in the kernel function be less than the nearest particle distance. Otherwise, solutions may exhibit oscillations near discontinuities unlike the "difference scheme" which unconditionally produces monotone results.Keywords: Heat conduction, Meshfree methods, Smoothed ParticleHydrodynamics (SPH), Second-order derivatives.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30689228 Optimization of Surface Roughness in Additive Manufacturing Processes via Taguchi Methodology
Authors: Anjian Chen, Joseph C. Chen
Abstract:
This paper studies a case where the targeted surface roughness of fused deposition modeling (FDM) additive manufacturing process is improved. The process is designing to reduce or eliminate the defects and improve the process capability index Cp and Cpk for an FDM additive manufacturing process. The baseline Cp is 0.274 and Cpk is 0.654. This research utilizes the Taguchi methodology, to eliminate defects and improve the process. The Taguchi method is used to optimize the additive manufacturing process and printing parameters that affect the targeted surface roughness of FDM additive manufacturing. The Taguchi L9 orthogonal array is used to organize the parameters' (four controllable parameters and one non-controllable parameter) effectiveness on the FDM additive manufacturing process. The four controllable parameters are nozzle temperature [°C], layer thickness [mm], nozzle speed [mm/s], and extruder speed [%]. The non-controllable parameter is the environmental temperature [°C]. After the optimization of the parameters, a confirmation print was printed to prove that the results can reduce the amount of defects and improve the process capability index Cp from 0.274 to 1.605 and the Cpk from 0.654 to 1.233 for the FDM additive manufacturing process. The final results confirmed that the Taguchi methodology is sufficient to improve the surface roughness of FDM additive manufacturing process.
Keywords: Additive manufacturing, fused deposition modeling, surface roughness, Six-Sigma, Taguchi method, 3D printing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13649227 Numerical Optimization within Vector of Parameters Estimation in Volatility Models
Authors: J. Arneric, A. Rozga
Abstract:
In this paper usefulness of quasi-Newton iteration procedure in parameters estimation of the conditional variance equation within BHHH algorithm is presented. Analytical solution of maximization of the likelihood function using first and second derivatives is too complex when the variance is time-varying. The advantage of BHHH algorithm in comparison to the other optimization algorithms is that requires no third derivatives with assured convergence. To simplify optimization procedure BHHH algorithm uses the approximation of the matrix of second derivatives according to information identity. However, parameters estimation in a/symmetric GARCH(1,1) model assuming normal distribution of returns is not that simple, i.e. it is difficult to solve it analytically. Maximum of the likelihood function can be founded by iteration procedure until no further increase can be found. Because the solutions of the numerical optimization are very sensitive to the initial values, GARCH(1,1) model starting parameters are defined. The number of iterations can be reduced using starting values close to the global maximum. Optimization procedure will be illustrated in framework of modeling volatility on daily basis of the most liquid stocks on Croatian capital market: Podravka stocks (food industry), Petrokemija stocks (fertilizer industry) and Ericsson Nikola Tesla stocks (information-s-communications industry).Keywords: Heteroscedasticity, Log-likelihood Maximization, Quasi-Newton iteration procedure, Volatility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26289226 Wet Polymeric Precipitation Synthesis for Monophasic Tricalcium Phosphate
Authors: I. Grigoraviciute-Puroniene, K. Tsuru, E. Garskaite, Z. Stankeviciute, A. Beganskiene, K. Ishikawa, A. Kareiva
Abstract:
Tricalcium phosphate (β-Ca3(PO4)2, β-TCP) powders were synthesized using wet polymeric precipitation method for the first time to our best knowledge. The results of X-ray diffraction analysis showed the formation of almost single a Ca-deficient hydroxyapatite (CDHA) phase of a poor crystallinity already at room temperature. With continuously increasing the calcination temperature up to 800 °C, the crystalline β-TCP was obtained as the main phase. It was demonstrated that infrared spectroscopy is very effective method to characterize the formation of β-TCP. The SEM results showed that β-TCP solids were homogeneous having a small particle size distribution. The β-TCP powders consisted of spherical particles varying in size from 100 to 300 nm. Fabricated β-TCP specimens were placed to the bones of the rats and maintained for 1-2 months.
Keywords: β-TCP, bone regeneration, wet chemical processing, polymeric precipitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10389225 Design and Analysis of Gauge R&R Studies: Making Decisions Based on ANOVA Method
Authors: Afrooz Moatari Kazerouni
Abstract:
In a competitive production environment, critical decision making are based on data resulted by random sampling of product units. Efficiency of these decisions depends on data quality and also their reliability scale. This point leads to the necessity of a reliable measurement system. Therefore, the conjecture process and analysing the errors contributes to a measurement system known as Measurement System Analysis (MSA). The aim of this research is on determining the necessity and assurance of extensive development in analysing measurement systems, particularly with the use of Repeatability and Reproducibility Gages (GR&R) to improve physical measurements. Nowadays in productive industries, repeatability and reproducibility gages released so well but they are not applicable as well as other measurement system analysis methods. To get familiar with this method and gain a feedback in improving measurement systems, this survey would be on “ANOVA" method as the most widespread way of calculating Repeatability and Reproducibility (R&R).Keywords: Analysis of Variance (ANOVA), MeasurementSystem Analysis (MSA), Part-Operator interaction effect, Repeatability and Reproducibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46479224 A Statistical Approach for Predicting and Optimizing Depth of Cut in AWJ Machining for 6063-T6 Al Alloy
Authors: Farhad Kolahan, A. Hamid Khajavi
Abstract:
In this paper, a set of experimental data has been used to assess the influence of abrasive water jet (AWJ) process parameters in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. The effects of these input parameters are studied on depth of cut (h); one of most important characteristics of AWJ. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. In the next stage, the proposed model is embedded into a Simulated Annealing (SA) algorithm to optimize the AWJ process parameters. The objective is to determine a suitable set of process parameters that can produce a desired depth of cut, considering the ranges of the process parameters. Computational results prove the effectiveness of the proposed model and optimization procedure.
Keywords: AWJ machining, Mathematical modeling, Simulated Annealing, Optimization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17579223 An Open-Label Pilot Study of Efficacy and Safety of 2% Curcuma aeruginosa Roxb. Extract Cream in the Treatment of Mild to Moderate Facial Seborrheic Dermatitis
Authors: Kulaya Wimolwat, Panlop Chakravitthamrong, Neti Waranuch
Abstract:
Background: Seborrheic dermatitis is a common chronic skin condition affecting the face, scalp, chest, and trunk. The cause of seborrheic dermatitis is still unknown. Sebum production, lipid composition, hormone levels, and Malassezia species have been suggested as important factors in the development of seborrheic dermatitis. Curcuma aeruginosa Roxb. extract-containing cream with anti-inflammatory and anti-androgenic properties may be beneficial for treating mild to moderate facial seborrheic dermatitis. Objectives: We evaluated the efficacy and safety of 2% C. aeruginosa Roxb. extract-containing cream in the treatment of mild to moderate seborrheic dermatitis. Methods: This was a prospective, open-label, and non-comparative study. Ten adult patients clinically diagnosed with mild to moderate seborrheic dermatitis were enrolled in a four-week study. The 2% C. aeruginosa Roxb. cream was applied twice daily to a lesional area on the face for four weeks. The Scoring Index (SI) ranking system on days 14 and 28 was compared with that at baseline to determine the efficacy of treatment. The adverse events (burning sensation and erythema) were evaluated on days 14 and 28 to determine the safety of the treatment. Results: Significant improvement was observed in the reduction of the mean SI at day 14 (2.9) and 28 (1.4) compared to that at baseline (4.9). An adverse reaction was observed on day 14 (mild erythema 20% and mild burning sensation 10%) and was resolved by the end of the study. Conclusion: This open-label pilot study has shown that there was a significant improvement in the severity in these seborrheic patients and most reported they were satisfied with it. Reported adverse events were all mild.
Keywords: Anti-androgenic, antifungals, anti-inflammatory, Curcuma aeruginosa, seborrheic dermatitis, efficacy, safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11379222 Optical Flow Technique for Supersonic Jet Measurements
Authors: H. D. Lim, Jie Wu, T. H. New, Shengxian Shi
Abstract:
This paper outlines the development of an experimental technique in quantifying supersonic jet flows, in an attempt to avoid seeding particle problems frequently associated with particle-image velocimetry (PIV) techniques at high Mach numbers. Based on optical flow algorithms, the idea behind the technique involves using high speed cameras to capture Schlieren images of the supersonic jet shear layers, before they are subjected to an adapted optical flow algorithm based on the Horn-Schnuck method to determine the associated flow fields. The proposed method is capable of offering full-field unsteady flow information with potentially higher accuracy and resolution than existing point-measurements or PIV techniques. Preliminary study via numerical simulations of a circular de Laval jet nozzle successfully reveals flow and shock structures typically associated with supersonic jet flows, which serve as useful data for subsequent validation of the optical flow based experimental results. For experimental technique, a Z-type Schlieren setup is proposed with supersonic jet operated in cold mode, stagnation pressure of 4 bar and exit Mach of 1.5. High-speed singleframe or double-frame cameras are used to capture successive Schlieren images. As implementation of optical flow technique to supersonic flows remains rare, the current focus revolves around methodology validation through synthetic images. The results of validation test offers valuable insight into how the optical flow algorithm can be further improved to improve robustness and accuracy. Despite these challenges however, this supersonic flow measurement technique may potentially offer a simpler way to identify and quantify the fine spatial structures within the shock shear layer.
Keywords: Schlieren, optical flow, supersonic jets, shock shear layer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18799221 Optimization of Protein Hydrolysate Production Process from Jatropha curcas Cake
Authors: Waraporn Apiwatanapiwat, Pilanee Vaithanomsat, Phanu Somkliang, Taweesiri Malapant
Abstract:
This was the first document revealing the investigation of protein hydrolysate production optimization from J. curcas cake. Proximate analysis of raw material showed 18.98% protein, 5.31% ash, 8.52% moisture and 12.18% lipid. The appropriate protein hydrolysate production process began with grinding the J. curcas cake into small pieces. Then it was suspended in 2.5% sodium hydroxide solution with ratio between solution/ J. curcas cake at 80:1 (v/w). The hydrolysis reaction was controlled at temperature 50 °C in water bath for 45 minutes. After that, the supernatant (protein hydrolysate) was separated using centrifuge at 8000g for 30 minutes. The maximum yield of resulting protein hydrolysate was 73.27 % with 7.34% moisture, 71.69% total protein, 7.12% lipid, 2.49% ash. The product was also capable of well dissolving in water.Keywords: Production, protein hydrolysate, Jatropha curcas cake, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19359220 Optimization Based Obstacle Avoidance
Authors: R. Dariani, S. Schmidt, R. Kasper
Abstract:
Based on a non-linear single track model which describes the dynamics of vehicle, an optimal path planning strategy is developed. Real time optimization is used to generate reference control values to allow leading the vehicle alongside a calculated lane which is optimal for different objectives such as energy consumption, run time, safety or comfort characteristics. Strict mathematic formulation of the autonomous driving allows taking decision on undefined situation such as lane change or obstacle avoidance. Based on position of the vehicle, lane situation and obstacle position, the optimization problem is reformulated in real-time to avoid the obstacle and any car crash.
Keywords: Autonomous driving, Obstacle avoidance, Optimal control, Path planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29919219 Optimization of Energy Consumption in Sequential Distillation Column
Authors: M.E. Masoumi, S. Kadkhodaie
Abstract:
Distillation column is one of the most common operations in process industries and is while the most expensive unit of the amount of energy consumption. Many ideas have been presented in the related literature for optimizing energy consumption in distillation columns. This paper studies the different heat integration methods in a distillation column which separate Benzene, Toluene, Xylene, and C9+. Three schemes of heat integration including, indirect sequence (IQ), indirect sequence with forward energy integration (IQF), and indirect sequence with backward energy integration (IQB) has been studied in this paper. Using shortcut method these heat integration schemes were simulated with Aspen HYSYS software and compared with each other with regarding economic considerations. The result shows that the energy consumption has been reduced 33% in IQF and 28% in IQB in comparison with IQ scheme. Also the economic result shows that the total annual cost has been reduced 12% in IQF and 8% in IQB regarding with IQ scheme. Therefore, the IQF scheme is most economic than IQB and IQ scheme.Keywords: Optimization, Distillation Column Sequence, Energy Savings
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29999218 Ant Colony Optimization for Optimal Distributed Generation in Distribution Systems
Authors: I. A. Farhat
Abstract:
The problem of optimal planning of multiple sources of distributed generation (DG) in distribution networks is treated in this paper using an improved Ant Colony Optimization algorithm (ACO). This objective of this problem is to determine the DG optimal size and location that in order to minimize the network real power losses. Considering the multiple sources of DG, both size and location are simultaneously optimized in a single run of the proposed ACO algorithm. The various practical constraints of the problem are taken into consideration by the problem formulation and the algorithm implementation. A radial power flow algorithm for distribution networks is adopted and applied to satisfy these constraints. To validate the proposed technique and demonstrate its effectiveness, the well-know 69-bus feeder standard test system is employed.cm.
Keywords: About Ant Colony Optimization (ACO), Distributed Generation (DG).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32579217 Deterministic Random Number Generator Algorithm for Cryptosystem Keys
Authors: Adi A. Maaita, Hamza A. A. Al_Sewadi
Abstract:
One of the crucial parameters of digital cryptographic systems is the selection of the keys used and their distribution. The randomness of the keys has a strong impact on the system’s security strength being difficult to be predicted, guessed, reproduced, or discovered by a cryptanalyst. Therefore, adequate key randomness generation is still sought for the benefit of stronger cryptosystems. This paper suggests an algorithm designed to generate and test pseudo random number sequences intended for cryptographic applications. This algorithm is based on mathematically manipulating a publically agreed upon information between sender and receiver over a public channel. This information is used as a seed for performing some mathematical functions in order to generate a sequence of pseudorandom numbers that will be used for encryption/decryption purposes. This manipulation involves permutations and substitutions that fulfill Shannon’s principle of “confusion and diffusion”. ASCII code characters were utilized in the generation process instead of using bit strings initially, which adds more flexibility in testing different seed values. Finally, the obtained results would indicate sound difficulty of guessing keys by attackers.Keywords: Cryptosystems, Information Security agreement, Key distribution, Random numbers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34099216 High Level Characterization and Optimization of Switched-Current Sigma-Delta Modulators with VHDL-AMS
Authors: A. Fakhfakh, N. Ksentini, M. Loulou, N. Masmoudi, J. J. Charlot
Abstract:
Today, design requirements are extending more and more from electronic (analogue and digital) to multidiscipline design. These current needs imply implementation of methodologies to make the CAD product reliable in order to improve time to market, study costs, reusability and reliability of the design process. This paper proposes a high level design approach applied for the characterization and the optimization of Switched-Current Sigma- Delta Modulators. It uses the new hardware description language VHDL-AMS to help the designers to optimize the characteristics of the modulator at a high level with a considerably reduced CPU time before passing to a transistor level characterization.Keywords: high level design, optimization, switched-Current Sigma-Delta Modulators, VHDL-AMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15009215 Solving Weighted Number of Operation Plus Processing Time Due-Date Assignment, Weighted Scheduling and Process Planning Integration Problem Using Genetic and Simulated Annealing Search Methods
Authors: Halil Ibrahim Demir, Caner Erden, Mumtaz Ipek, Ozer Uygun
Abstract:
Traditionally, the three important manufacturing functions, which are process planning, scheduling and due-date assignment, are performed separately and sequentially. For couple of decades, hundreds of studies are done on integrated process planning and scheduling problems and numerous researches are performed on scheduling with due date assignment problem, but unfortunately the integration of these three important functions are not adequately addressed. Here, the integration of these three important functions is studied by using genetic, random-genetic hybrid, simulated annealing, random-simulated annealing hybrid and random search techniques. As well, the importance of the integration of these three functions and the power of meta-heuristics and of hybrid heuristics are studied.
Keywords: Process planning, weighted scheduling, weighted due-date assignment, genetic search, simulated annealing, hybrid meta-heuristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15729214 Generalized Maximum Entropy Method for Cosmic Source Localization
Authors: Youssef Khmou, Said Safi, Miloud Frikel
Abstract:
The Maximum entropy principle in spectral analysis was used as an estimator of Direction of Arrival (DoA) of electromagnetic or acoustic sources impinging on an array of sensors, indeed the maximum entropy operator is very efficient when the signals of the radiating sources are ergodic and complex zero mean random processes which is the case for cosmic sources. In this paper, we present basic review of the maximum entropy method (MEM) which consists of rank one operator but not a projector, and we elaborate a new operator which is full rank and sum of all possible projectors. Two dimensional Simulation results based on Monte Carlo trials prove the resolution power of the new operator where the MEM presents some erroneous fluctuations.
Keywords: Maximum entropy, Cosmic source, Localization, operator, projector, azimuth, elevation, DoA, circular array.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21149213 Impact of Ship Traffic to PM2.5 and Particle Number Concentrations in Three Port-Cities of the Adriatic/Ionian Area
Authors: Daniele Contini, Antonio Donateo, Andrea Gambaro, Athanasios Argiriou, Dimitrios Melas, Daniela Cesari, Anastasia Poupkou, Athanasios Karagiannidis, Apostolos Tsakis, Eva Merico, Rita Cesari, Adelaide Dinoi
Abstract:
Emissions of atmospheric pollutants from ships and harbour activities are a growing concern at international level given their potential impacts on air quality and climate. These close-to-land emissions have potential impact on local communities in terms of air quality and health. Recent studies show that the impact of maritime traffic to atmospheric particulate matter concentrations in several coastal urban areas is comparable with the impact of road traffic of a medium size town. However, several different approaches have been used for these estimates making difficult a direct comparison of results. In this work, an integrated approach based on emission inventories and dedicated measurement campaigns has been applied to give a comparable estimate of the impact of maritime traffic to PM2.5 and particle number concentrations in three major harbours of the Adriatic/Ionian Seas. The influences of local meteorology and of the logistic layout of the harbours are discussed.
Keywords: Ship emissions, PM2.5, particle number concentrations, impact of shipping to atmospheric aerosol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23789212 Solid State Fermentation of Cassava Peel with Trichoderma viride (ATCC 36316) for Protein Enrichment
Authors: Olufunke O. Ezekiel, Ogugua C. Aworh
Abstract:
Solid state fermentation of cassava peel with emphasis on protein enrichment using Trichoderma viride was evaluated. The effect of five variables: moisture content, pH, particle size (p), nitrogen source and incubation temperature; on the true protein and total sugars of cassava peel was investigated. The optimum fermentation period was established to be 8 days. Total sugars were 5-fold higher at pH 6 relative to pH 4 and 7-fold higher when cassava peels were fermented at 30oC relative to 25oC as well as using ammonium sulfate as the nitrogen source relative to urea or a combination of both. Total sugars ranged between 123.21mg/g at 50% initial moisture content to 374mg/g at 60% and from 190.59mg/g with particle size range of 2.00>p>1.41mm to 310.10mg/g with 4.00>p>3.35mm.True protein ranged from 229.70 mg/g at pH 4 to 284.05 mg/g at pH 6; from 200.87 mg/g with urea as nitrogen source and to 254.50mg/g with ammonium sulfate; from 213.82mg/g at 50% initial moisture content to 254.50mg/g at 60% moisture content, from 205.75mg/g in cassava peel with 5.6>p> 4.75mm to 268.30 in cassava peel with particle size 4.00>p>3.35mm, from 207.57mg/g at 25oC to 254.50mg/g at 30oC Cassava peel with particle size 4.00>p>3.35 mm and initial moisture content of 60% at pH 6.0, 30oC incubation temperature with ammonium sulfate (10g N / kg substrate) was most suitable for protein enrichment with Trichoderma viride. Crude protein increased from 4.21 % in unfermented cassava peel samples to 10.43 % in fermented samples.
Keywords: Cassava peel, Solid state fermentation, Trichoderma viride, Total sugars, True protein.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33159211 Performance Evaluation of Data Mining Techniques for Predicting Software Reliability
Authors: Pradeep Kumar, Abdul Wahid
Abstract:
Accurate software reliability prediction not only enables developers to improve the quality of software but also provides useful information to help them for planning valuable resources. This paper examines the performance of three well-known data mining techniques (CART, TreeNet and Random Forest) for predicting software reliability. We evaluate and compare the performance of proposed models with Cascade Correlation Neural Network (CCNN) using sixteen empirical databases from the Data and Analysis Center for Software. The goal of our study is to help project managers to concentrate their testing efforts to minimize the software failures in order to improve the reliability of the software systems. Two performance measures, Normalized Root Mean Squared Error (NRMSE) and Mean Absolute Errors (MAE), illustrate that CART model is accurate than the models predicted using Random Forest, TreeNet and CCNN in all datasets used in our study. Finally, we conclude that such methods can help in reliability prediction using real-life failure datasets.
Keywords: Classification, Cascade Correlation Neural Network, Random Forest, Software reliability, TreeNet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18149210 Comparison between Post- and Oxy-Combustion Systems in a Petroleum Refinery Unit Using Modeling and Optimization
Authors: Farooq A. Al-Sheikh, Ali Elkamel, William A. Anderson
Abstract:
A fluidized catalytic cracking unit (FCCU) is one of the effective units in many refineries. Modeling and optimization of FCCU were done by many researchers in past decades, but in this research, comparison between post- and oxy-combustion was studied in the regenerator-FCCU. Therefore, a simplified mathematical model was derived by doing mass/heat balances around both reactor and regenerator. A state space analysis was employed to show effects of the flow rates variables such as air, feed, spent catalyst, regenerated catalyst and flue gas on the output variables. The main aim of studying dynamic responses is to figure out the most influencing variables that affect both reactor/regenerator temperatures; also, finding the upper/lower limits of the influencing variables to ensure that temperatures of the reactors and regenerator work within normal operating conditions. Therefore, those values will be used as side constraints in the optimization technique to find appropriate operating regimes. The objective functions were modeled to be maximizing the energy in the reactor while minimizing the energy consumption in the regenerator. In conclusion, an oxy-combustion process can be used instead of a post-combustion one.
Keywords: FCCU modeling, optimization, oxy-combustion post-combustion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9029209 A Multigrid Approach for Three-Dimensional Inverse Heat Conduction Problems
Authors: Jianhua Zhou, Yuwen Zhang
Abstract:
A two-step multigrid approach is proposed to solve the inverse heat conduction problem in a 3-D object under laser irradiation. In the first step, the location of the laser center is estimated using a coarse and uniform grid system. In the second step, the front-surface temperature is recovered in good accuracy using a multiple grid system in which fine mesh is used at laser spot center to capture the drastic temperature rise in this region but coarse mesh is employed in the peripheral region to reduce the total number of sensors required. The effectiveness of the two-step approach and the multiple grid system are demonstrated by the illustrative inverse solutions. If the measurement data for the temperature and heat flux on the back surface do not contain random error, the proposed multigrid approach can yield more accurate inverse solutions. When the back-surface measurement data contain random noise, accurate inverse solutions cannot be obtained if both temperature and heat flux are measured on the back surface.
Keywords: Conduction, inverse problems, conjugated gradient method, laser.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 820