Search results for: optimal debt ceiling
1203 Optimization of Lean Methodologies in the Textile Industry Using Design of Experiments
Authors: Ahmad Yame, Ahad Ali, Badih Jawad, Daw Al-Werfalli Mohamed Nasser, Sabah Abro
Abstract:
Industries in general have a lot of waste. Wool textile company, Baniwalid, Libya has many complex problems that led to enormous waste generated due to the lack of lean strategies, expertise, technical support and commitment. To successfully address waste at wool textile company, this study will attempt to develop a methodical approach that integrates lean manufacturing tools to optimize performance characteristics such as lead time and delivery. This methodology will utilize Value Stream Mapping (VSM) techniques to identify the process variables that affect production. Once these variables are identified, Design of Experiments (DOE) Methodology will be used to determine the significantly influential process variables, these variables are then controlled and set at their optimal to achieve optimal levels of productivity, quality, agility, efficiency and delivery to analyze the outputs of the simulation model for different lean configurations. The goal of this research is to investigate how the tools of lean manufacturing can be adapted from the discrete to the continuous manufacturing environment and to evaluate their benefits at a specific industrial.Keywords: Lean manufacturing, DOE, value stream mapping, textiles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19401202 Optimization of Surface Roughness and Vibration in Turning of Aluminum Alloy AA2024 Using Taguchi Technique
Authors: Vladimir Aleksandrovich Rogov, Ghorbani Siamak
Abstract:
Determination of optimal conditions of machining parameters is important to reduce the production cost and achieve the desired surface quality. This paper investigates the influence of cutting parameters on surface roughness and natural frequency in turning of aluminum alloy AA2024. The experiments were performed at the lathe machine using two different cutting tools made of AISI 5140 and carbide cutting insert coated with TiC. Turning experiments were planned by Taguchi method L9 orthogonal array.Three levels for spindle speed, feed rate, depth of cut and tool overhang were chosen as cutting variables. The obtained experimental data has been analyzed using signal to noise ratio and analysis of variance. The main effects have been discussed and percentage contributions of various parameters affecting surface roughness and natural frequency, and optimal cutting conditions have been determined. Finally, optimization of the cutting parameters using Taguchi method was verified by confirmation experiments.
Keywords: Turning, Cutting conditions, Surface roughness, Natural frequency, Taguchi method, ANOVA, S/N ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46141201 Optimal Document Archiving and Fast Information Retrieval
Authors: Hazem M. El-Bakry, Ahmed A. Mohammed
Abstract:
In this paper, an intelligent algorithm for optimal document archiving is presented. It is kown that electronic archives are very important for information system management. Minimizing the size of the stored data in electronic archive is a main issue to reduce the physical storage area. Here, the effect of different types of Arabic fonts on electronic archives size is discussed. Simulation results show that PDF is the best file format for storage of the Arabic documents in electronic archive. Furthermore, fast information detection in a given PDF file is introduced. Such approach uses fast neural networks (FNNs) implemented in the frequency domain. The operation of these networks relies on performing cross correlation in the frequency domain rather than spatial one. It is proved mathematically and practically that the number of computation steps required for the presented FNNs is less than that needed by conventional neural networks (CNNs). Simulation results using MATLAB confirm the theoretical computations.Keywords: Information Storage and Retrieval, Electronic Archiving, Fast Information Detection, Cross Correlation, Frequency Domain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15861200 A Case Study on Performance of Isolated Bridges under Near-Fault Ground Motion
Authors: Daniele Losanno, H. A. Hadad, Giorgio Serino
Abstract:
This paper presents a numerical investigation on the seismic performance of a benchmark bridge with different optimal isolation systems under near fault ground motion. Usually, very large displacements make seismic isolation an unfeasible solution due to boundary conditions, especially in case of existing bridges or high risk seismic regions. Hence, near-fault ground motions are most likely to affect either structures with long natural period range like isolated structures or structures sensitive to velocity content such as viscously damped structures. The work is aimed at analyzing the seismic performance of a three-span continuous bridge designed with different isolation systems having different levels of damping. The case study was analyzed in different configurations including: (a) simply supported, (b) isolated with lead rubber bearings (LRBs), (c) isolated with rubber isolators and 10% classical damping (HDLRBs), and (d) isolated with rubber isolators and 70% supplemental damping ratio. Case (d) represents an alternative control strategy that combines the effect of seismic isolation with additional supplemental damping trying to take advantages from both solutions. The bridge is modeled in SAP2000 and solved by time history direct-integration analyses under a set of six recorded near-fault ground motions. In addition to this, a set of analysis under Italian code provided seismic action is also conducted, in order to evaluate the effectiveness of the suggested optimal control strategies under far field seismic action. Results of the analysis demonstrated that an isolated bridge equipped with HDLRBs and a total equivalent damping ratio of 70% represents a very effective design solution for both mitigation of displacement demand at the isolation level and base shear reduction in the piers also in case of near fault ground motion.
Keywords: Isolated bridges, optimal design, near-fault motion, supplemental damping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12711199 Banking Union: A New Step towards Completing the Economic and Monetary Union
Authors: Marijana Ivanov, Roman Šubić
Abstract:
This study analyzes the critical gaps in the architecture of European stability and the expected role of the banking union as the new important step towards completing the Economic and Monetary Union that should enable the creation of safe and sound financial sector for the euro area market. The single rulebook together with the Single Supervisory Mechanism and the Single Resolution Mechanism - as two main pillars of the banking union, should provide a consistent application of common rules and administrative standards for supervision, recovery and resolution of banks – with the final aim of replacing the former bail-out practice with the bail-in system through which possible future bank failures would be resolved by their own funds, i.e. with minimal costs for taxpayers and real economy. In this way, the vicious circle between banks and sovereigns would be broken. It would also reduce the financial fragmentation recorded in the years of crisis as the result of divergent behaviors in risk premium, lending activities and interest rates between the core and the periphery. In addition, it should strengthen the effectiveness of monetary transmission channels, in particular the credit channels and overflows of liquidity on the money market which, due to the fragmentation of the common financial market, has been significantly disabled in period of crisis. However, contrary to all the positive expectations related to the future functioning of the banking union, major findings of this study indicate that characteristics of the economic system in which the banking union will operate should not be ignored. The euro area is an integration of strong and weak entities with large differences in economic development, wealth, assets of banking systems, growth rates and accountability of fiscal policy. The analysis indicates that low and unbalanced economic growth remains a challenge for the maintenance of financial stability and this problem cannot be resolved just by a single supervision. In many countries bank assets exceed their GDP by several times and large banks are still a matter of concern, because of their systemic importance for individual countries and the euro zone as a whole. The creation of the Single Supervisory Mechanism and the Single Resolution Mechanism is a response to the European crisis, which has particularly affected peripheral countries and caused the associated loop between the banking crisis and the sovereign debt crisis, but has also influenced banks’ balance sheets in the core countries, as the result of crossborder capital flows. The creation of the SSM and the SRM should prevent the similar episodes to happen again and should also provide a new opportunity for strengthening of economic and financial systems of the peripheral countries. On the other hand, there is a potential threat that future focus of the ECB, resolution mechanism and other relevant institutions will be extremely oriented towards large and significant banks (whereby one half of them operate in the core and most important euro area countries), and therefore it remains questionable to what extent will the common resolution funds will be used for rescue of less important institutions. Recent geopolitical developments will be the optimal indicator to show whether the previously established mechanisms are sufficient enough to maintain the adequate financial stability in the euro area market.Keywords: Banking Union, financial integration, single supervisory mechanism (SSM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16651198 Optimal Economic Restructuring Aimed at an Increase in GDP Constrained by a Decrease in Energy Consumption and CO2 Emissions
Authors: Alexander Y. Vaninsky
Abstract:
The objective of this paper is finding the way of economic restructuring - that is, change in the shares of sectoral gross outputs - resulting in the maximum possible increase in the gross domestic product (GDP) combined with decreases in energy consumption and CO2 emissions. It uses an input-output model for the GDP and factorial models for the energy consumption and CO2 emissions to determine the projection of the gradient of GDP, and the antigradients of the energy consumption and CO2 emissions, respectively, on a subspace formed by the structure-related variables. Since the gradient (antigradient) provides a direction of the steepest increase (decrease) of the objective function, and their projections retain this property for the functions' limitation to the subspace, each of the three directional vectors solves a particular problem of optimal structural change. In the next step, a type of factor analysis is applied to find a convex combination of the projected gradient and antigradients having maximal possible positive correlation with each of the three. This convex combination provides the desired direction of the structural change. The national economy of the United States is used as an example of applications.
Keywords: Economic restructuring, Input-Output analysis, Divisia index, Factorial decomposition, E3 models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16081197 Modern Seismic Design Approach for Buildings with Hysteretic Dampers
Authors: Vanessa A. Segovia, Sonia E. Ruiz
Abstract:
The use of energy dissipation systems for seismic applications has increased worldwide, thus it is necessary to develop practical and modern criteria for their optimal design. Here, a direct displacement-based seismic design approach for frame buildings with hysteretic energy dissipation systems (HEDS) is applied. The building is constituted by two individual structural systems consisting of: 1) a main elastic structural frame designed for service loads; and 2) a secondary system, corresponding to the HEDS, that controls the effects of lateral loads. The procedure implies to control two design parameters: a) the stiffness ratio (α=Kframe/Ktotal system), and b) the strength ratio (γ=Vdamper/Vtotal system). The proposed damage-controlled approach contributes to the design of a more sustainable and resilient building because the structural damage is concentrated on the HEDS. The reduction of the design displacement spectrum is done by means of a damping factor (recently published) for elastic structural systems with HEDS, located in Mexico City. Two limit states are verified: serviceability and near collapse. Instead of the traditional trial-error approach, a procedure that allows the designer to establish the preliminary sizes of the structural elements of both systems is proposed. The design methodology is applied to an 8-story steel building with buckling restrained braces, located in soft soil of Mexico City. With the aim of choosing the optimal design parameters, a parametric study is developed considering different values of હ and . The simplified methodology is for preliminary sizing, design, and evaluation of the effectiveness of HEDS, and it constitutes a modern and practical tool that enables the structural designer to select the best design parameters.
Keywords: Damage-controlled buildings, direct displacementbased seismic design, optimal hysteretic energy dissipation systems
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23531196 Aerodynamics and Optimization of Airfoil Under Ground Effect
Authors: Kyoungwoo Park, Byeong Sam Kim, Juhee Lee, Kwang Soo Kim
Abstract:
The Prediction of aerodynamic characteristics and shape optimization of airfoil under the ground effect have been carried out by integration of computational fluid dynamics and the multiobjective Pareto-based genetic algorithm. The main flow characteristics around an airfoil of WIG craft are lift force, lift-to-drag ratio and static height stability (H.S). However, they show a strong trade-off phenomenon so that it is not easy to satisfy the design requirements simultaneously. This difficulty can be resolved by the optimal design. The above mentioned three characteristics are chosen as the objective functions and NACA0015 airfoil is considered as a baseline model in the present study. The profile of airfoil is constructed by Bezier curves with fourteen control points and these control points are adopted as the design variables. For multi-objective optimization problems, the optimal solutions are not unique but a set of non-dominated optima and they are called Pareto frontiers or Pareto sets. As the results of optimization, forty numbers of non- dominated Pareto optima can be obtained at thirty evolutions.Keywords: Aerodynamics, Shape optimization, Airfoil on WIGcraft, Genetic algorithm, Computational fluid dynamics (CFD).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32301195 Rolling Element Bearing Diagnosis by Improved Envelope Spectrum: Optimal Frequency Band Selection
Authors: Juan David Arango, Alejandro Restrepo-Martinez
Abstract:
The Rolling Element Bearing (REB) vibration diagnosis is worth of special interest by the variety of REB and the wide necessity of those elements in industrial applications. The presence of a localized fault in a REB gives rise to a vibrational response, characterized by the modulation of a carrier signal. Frequency content of carrier signal (Spectral Frequency –f) is mainly related to resonance frequencies of the REB. This carrier signal is modulated by another signal, governed by the periodicity of the fault impact (Cyclic Frequency –α). In this sense, REB fault vibration response gives rise to a second-order cyclostationary signal. Second order cyclostationary signals could be represented in a bi-spectral map, where Spectral Coherence –SCoh are plotted against f and α. The Improved Envelope Spectrum –IES, is a useful approach to execute REB fault diagnosis. IES could be applied by the integration of SCoh over a predefined bandwidth on the f axis. Approaches to select f-bandwidth have been recently exposed by the definition of a metric which intends to evaluate the magnitude of the IES at the fault characteristics frequencies. This metric is represented in a 1/3-binary tree as a function of the frequency bandwidth and centre. Based on this binary tree the optimal frequency band is selected. However, some advantages have been seen if the metric is changed, which in fact tends to dictate different optimal f-bandwidth and so improve the IES representation. This paper evaluates the behaviour of the IES from a different metric optimization. This metric is based on the sample correlation coefficient, detecting high peaks in the selected frequencies while penalizing high peaks in the neighbours of the selected frequencies. Prior results indicate an improvement on the signal-noise ratio (SNR) on around 86% of samples analysed, which belong to IMS database.
Keywords: Sample Correlation IESFOgram, cyclostationary analysis, improved envelope spectrum, IES, rolling element bearing diagnosis, spectral coherence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7421194 Energy Efficient and Reliable Geographic Routing in Wireless Sensor Networks
Authors: Eunil Park, Kwangsu Cho
Abstract:
The wireless link can be unreliable in realistic wireless sensor networks (WSNs). Energy efficient and reliable data forwarding is important because each node has limited resources. Therefore, we must suggest an optimal solution that considers using the information of the node-s characteristics. Previous routing protocols were unsuited to realistic asymmetric WSNs. In this paper, we propose a Protocol that considers Both sides of Link-quality and Energy (PBLE), an optimal routing protocol that balances modified link-quality, distance and energy. Additionally, we propose a node scheduling method. PBLE achieves a longer lifetime than previous routing protocols and is more energy-efficient. PBLE uses energy, local information and both sides of PRR in a 1-hop distance. We explain how to send data packets to the destination node using the node's information. Simulation shows PBLE improves delivery rate and network lifetime compared to previous schemes. Moreover, we show the improvement in various WSN environments.Keywords: energy-efficient, lifetime, PBLE, unreliable
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16481193 A Hybrid Nature Inspired Algorithm for Generating Optimal Query Plan
Authors: R. Gomathi, D. Sharmila
Abstract:
The emergence of the Semantic Web technology increases day by day due to the rapid growth of multiple web pages. Many standard formats are available to store the semantic web data. The most popular format is the Resource Description Framework (RDF). Querying large RDF graphs becomes a tedious procedure with a vast increase in the amount of data. The problem of query optimization becomes an issue in querying large RDF graphs. Choosing the best query plan reduces the amount of query execution time. To address this problem, nature inspired algorithms can be used as an alternative to the traditional query optimization techniques. In this research, the optimal query plan is generated by the proposed SAPSO algorithm which is a hybrid of Simulated Annealing (SA) and Particle Swarm Optimization (PSO) algorithms. The proposed SAPSO algorithm has the ability to find the local optimistic result and it avoids the problem of local minimum. Experiments were performed on different datasets by changing the number of predicates and the amount of data. The proposed algorithm gives improved results compared to existing algorithms in terms of query execution time.
Keywords: Semantic web, RDF, Query optimization, Nature inspired algorithms, PSO, SA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22391192 A Proposed Optimized and Efficient Intrusion Detection System for Wireless Sensor Network
Authors: Abdulaziz Alsadhan, Naveed Khan
Abstract:
In recent years intrusions on computer network are the major security threat. Hence, it is important to impede such intrusions. The hindrance of such intrusions entirely relies on its detection, which is primary concern of any security tool like Intrusion detection system (IDS). Therefore, it is imperative to accurately detect network attack. Numerous intrusion detection techniques are available but the main issue is their performance. The performance of IDS can be improved by increasing the accurate detection rate and reducing false positive. The existing intrusion detection techniques have the limitation of usage of raw dataset for classification. The classifier may get jumble due to redundancy, which results incorrect classification. To minimize this problem, Principle component analysis (PCA), Linear Discriminant Analysis (LDA) and Local Binary Pattern (LBP) can be applied to transform raw features into principle features space and select the features based on their sensitivity. Eigen values can be used to determine the sensitivity. To further classify, the selected features greedy search, back elimination, and Particle Swarm Optimization (PSO) can be used to obtain a subset of features with optimal sensitivity and highest discriminatory power. This optimal feature subset is used to perform classification. For classification purpose, Support Vector Machine (SVM) and Multilayer Perceptron (MLP) are used due to its proven ability in classification. The Knowledge Discovery and Data mining (KDD’99) cup dataset was considered as a benchmark for evaluating security detection mechanisms. The proposed approach can provide an optimal intrusion detection mechanism that outperforms the existing approaches and has the capability to minimize the number of features and maximize the detection rates.
Keywords: Particle Swarm Optimization (PSO), Principle component analysis (PCA), Linear Discriminant Analysis (LDA), Local Binary Pattern (LBP), Support Vector Machine (SVM), Multilayer Perceptron (MLP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27651191 An Optimization of the New Die Design of Sheet Hydroforming by Taguchi Method
Authors: M. Hosseinzadeh, S. A. Zamani, A. Taheri
Abstract:
During the last few years, several sheet hydroforming processes have been introduced. Despite the advantages of these methods, they have some limitations. Of the processes, the two main ones are the standard hydroforming and hydromechanical deep drawing. A new sheet hydroforming die set was proposed that has the advantages of both processes and eliminates their limitations. In this method, a polyurethane plate was used as a part of the die-set to control the blank holder force. This paper outlines the Taguchi optimization methodology, which is applied to optimize the effective parameters in forming cylindrical cups by the new die set of sheet hydroforming process. The process parameters evaluated in this research are polyurethane hardness, polyurethane thickness, forming pressure path and polyurethane hole diameter. The design of experiments based upon L9 orthogonal arrays by Taguchi was used and analysis of variance (ANOVA) was employed to analyze the effect of these parameters on the forming pressure. The analysis of the results showed that the optimal combination for low forming pressure is harder polyurethane, bigger diameter of polyurethane hole and thinner polyurethane. Finally, the confirmation test was derived based on the optimal combination of parameters and it was shown that the Taguchi method is suitable to examine the optimization process.Keywords: Sheet Hydroforming, Optimization, Taguchi Method
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25961190 A Particle Swarm Optimal Control Method for DC Motor by Considering Energy Consumption
Authors: Yingjie Zhang, Ming Li, Ying Zhang, Jing Zhang, Zuolei Hu
Abstract:
In the actual start-up process of DC motors, the DC drive system often faces a conflict between energy consumption and acceleration performance. To resolve the conflict, this paper proposes a comprehensive performance index that energy consumption index is added on the basis of classical control performance index in the DC motor starting process. Taking the comprehensive performance index as the cost function, particle swarm optimization algorithm is designed to optimize the comprehensive performance. Then it conducts simulations on the optimization of the comprehensive performance of the DC motor on condition that the weight coefficient of the energy consumption index should be properly designed. The simulation results show that as the weight of energy consumption increased, the energy efficiency was significantly improved at the expense of a slight sacrifice of fastness indicators with the comprehensive performance index method. The energy efficiency was increased from 63.18% to 68.48% and the response time reduced from 0.2875s to 0.1736s simultaneously compared with traditional proportion integrals differential controller in energy saving.
Keywords: Comprehensive performance index, energy consumption, acceleration performance, particle swarm optimal control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6421189 Investigation of the Emulsifying Properties of Bambara Groundnut Flour and Starch
Authors: Ebunoluwa G. Gabriel, Victoria A. Jideani, Daniel I. O. Ikhu-Omoregbe
Abstract:
The current desire in food and industrial emulsification is the use of natural emulsifiers. Bambara groundnut flour (BGNF) and its starch (BGNS) will serve both emulsifying and nutritional purposes if found suitable. This current study was aimed at investigating the emulsifying properties of BGNF/BGNS. BGNS was extracted from the BGNF. Emulsions were prepared using a wide range of flour-oil-water and starch-oil-water composition as generated through the application of Response Surface (D-optimal) design. Prepared emulsions were investigated for stability to creaming/sedimentation (using the kinetic information from turbiscan) and flocculation/coalescence (by monitoring the droplet diameter growth using optical microscope) over 5 days. The most stable emulsions (one BGNF-stabilized and the other BGNS-stabilized) were determined. The optimal emulsifier/oil composition was 9g/39g for BGNF and 5g/30g for BGNS. The two emulsions had only 30% and 50% growth in oil droplet diameter respectively by day 5, compared to over 3000% in the unstable ones. The BGNF-stabilized emulsions were more stable than the BGNS-stabilized ones. Emulsions were successfully stabilized with BGNF and BGNS.
Keywords: Bambara groundnut, coalescence, creaming, emulsification, emulsion, emulsion stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34491188 Effect of Sintering Temperature Curve in Wick Manufactured for Loop Heat Pipe
Authors: Shen-Chun Wu, Chuo-Jeng Huang, Wun-Hong Yang, Jy-Cheng Chang, Chien-Chun Kung
Abstract:
This investigation examines the effect of the sintering temperature curve in manufactured nickel powder capillary structure (wick) for a loop heat pipe (LHP). The sintering temperature curve is composed of a region of increasing temperature; a region of constant temperature and a region of declining temperature. The most important region is that in which the temperature increases, as an index in the stage in which the temperature increases. The wick of nickel powder is manufactured in the stage of fixed sintering temperature and the time between the stage of constant temperature and the stage of falling temperature. When the slope of the curve in the region of increasing temperature is unity (equivalent to 10 °C/min), the structure of the wick is complete and the heat transfer performance is optimal. The result of experiment test demonstrates that the heat transfer performance is optimal at 320W; the minimal total thermal resistance is approximately 0.18°C/W, and the heat flux is 17W/cm2; the internal parameters of the wick are an effective pore radius of 3.1 μm, a permeability of 3.25×10-13m2 and a porosity of 71%.Keywords: Loop heat pipe (LHP), capillary structure (wick), sintered temperature curve.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20941187 The Power of “Merkiavelli”: Representations of Angela Merkel in the Portuguese Press (2008-2015)
Authors: Ana Mouro, Ana Ramalheira
Abstract:
Since 1989, with the Fall of the Berlin Wall, Germany has undergone a profound restructuring political and economic process. When the Euro Crisis broke out, Germany was no longer the “sick man” of Europe. Instead, it had recovered its dominance as the strongest and wealthiest economy within the European Union. With the European Debt Crisis, that has been taking place in the European Union since the end of 2009, Germany´s Chancellor Angela Merkel has gained the power of deciding, so to say, on the fate of the debtor nations, but she neither stands for binding German commitments, nor refuses assistance. A debate on whether Merkel’s hesitation has been deliberated and used as a means of coercion has arisen on international print media, and the Portuguese Press has been no exception. This study, which was conducted by using news reporting, opinion articles, interviews and editorials, published in the Portuguese weekly Expresso and the daily Público, from 2008 to 2015, tries to show how Merkel’s hesitation, depicted in the press by the term “Merkiavelli”, was perceived in Portugal, a country that had to embrace the austerity measures, imposed by the European Central Bank, but defined under Angela Merkel´s leading role.
Keywords: Euro crisis, “Merkiavelli”, cultural studies, Portuguese quality press.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8521186 Optimal Design of Airfoil with High Aspect Ratio in Unmanned Aerial Vehicles
Authors: Kyoungwoo Park, Ji-Won Han, Hyo-Jae Lim, Byeong-Sam Kim, Juhee Lee
Abstract:
Shape optimization of the airfoil with high aspect ratio of long endurance unmanned aerial vehicle (UAV) is performed by the multi-objective optimization technology coupled with computational fluid dynamics (CFD). For predicting the aerodynamic characteristics around the airfoil the high-fidelity Navier-Stokes solver is employed and SMOGA (Simple Multi-Objective Genetic Algorithm), which is developed by authors, is used for solving the multi-objective optimization problem. To obtain the optimal solutions of the design variable (i.e., sectional airfoil profile, wing taper ratio and sweep) for high performance of UAVs, both the lift and lift-to-drag ratio are maximized whereas the pitching moment should be minimized, simultaneously. It is found that the lift force and lift-to-drag ratio are linearly dependent and a unique and dominant solution are existed. However, a trade-off phenomenon is observed between the lift-to-drag ratio and pitching moment. As the result of optimization, sixty-five (65) non-dominated Pareto individuals at the cutting edge of design spaces that is decided by airfoil shapes can be obtained.Keywords: Unmanned aerial vehicle (UAV), Airfoil, CFD, Shape optimization, Lift-to-drag ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 64361185 Simulation Model of an Ultra-Light Overhead Conveyor System; Analysis of the Process in the Warehouse
Authors: Batin Latif Aylak, Bernd Noche, M. Baran Cantepe, Aydin Karakaya
Abstract:
Ultra-light overhead conveyor systems are rope-based conveying systems with individually driven vehicles. The vehicles can move automatically on the rope and this can be realized by energy and signals. The ultra-light overhead conveyor systems always must be integrated with a logistical process by finding a best way for a cheaper material flow in order to guarantee precise and fast workflows. This paper analyzes the process of an ultra-light overhead conveyor system using necessary assumptions. The analysis consists of three scenarios. These scenarios are based on raising the vehicle speeds with equal increments at each case. The correlation between the vehicle speed and system throughput is investigated. A discrete-event simulation model of an ultra-light overhead conveyor system is constructed using DOSIMIS-3 software to implement three scenarios. According to simulation results; the optimal scenario, hence the optimal vehicle speed, is found out among three scenarios. This simulation model demonstrates the effect of increased speed on the system throughput.
Keywords: Logistics, material flow, simulation, ultra-light overhead conveyor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25641184 Multiobjective Optimization Solution for Shortest Path Routing Problem
Authors: C. Chitra, P. Subbaraj
Abstract:
The shortest path routing problem is a multiobjective nonlinear optimization problem with constraints. This problem has been addressed by considering Quality of service parameters, delay and cost objectives separately or as a weighted sum of both objectives. Multiobjective evolutionary algorithms can find multiple pareto-optimal solutions in one single run and this ability makes them attractive for solving problems with multiple and conflicting objectives. This paper uses an elitist multiobjective evolutionary algorithm based on the Non-dominated Sorting Genetic Algorithm (NSGA), for solving the dynamic shortest path routing problem in computer networks. A priority-based encoding scheme is proposed for population initialization. Elitism ensures that the best solution does not deteriorate in the next generations. Results for a sample test network have been presented to demonstrate the capabilities of the proposed approach to generate well-distributed pareto-optimal solutions of dynamic routing problem in one single run. The results obtained by NSGA are compared with single objective weighting factor method for which Genetic Algorithm (GA) was applied.Keywords: Multiobjective optimization, Non-dominated SortingGenetic Algorithm, Routing, Weighted sum.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32651183 Contribution to the Study and Optimal Exploitation of a Solar Power System for a Semi-Arid Zone (Case Study: Ferkene, Algeria)
Authors: D. Dib, W. Guebabi, M. B. Guesmi
Abstract:
The objective of this paper is a contribution to a study of power supply by solar energy system called a common Ferkène north of Algerian desert in the semi-arid area. The optimal exploitation of the system, goes through stages of study and essential design, the choice of the model of the photovoltaic panel, the study of behavior with all the parameters involved in simulation before fixing the trajectory tracking the maximum point the power to extract (MPPT), form the essential platform to shape the design of the solar system set up to supply the town Ferkène without considering the grid. The identification of the common Ferkène by the collection of geographical, meteorological, demographic and electrical provides a basis uniform and important data. The results reflect a valid fictive model for any attempt to study and design a solar system to supply an arid or semi-arid zone by electrical energy from photovoltaic panels.
Keywords: Solar power, photovoltaic panel, Boost converter, supply, design, electric power, Ferkène, Algeria.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17521182 Estimating the Traffic Impacts of Green Light Optimal Speed Advisory Systems Using Microsimulation
Authors: C. B. Masera, M. Imprialou, L. Budd, C. Morton
Abstract:
Even though signalised intersections are necessary for urban road traffic management, they can act as bottlenecks and disrupt traffic operations. Interrupted traffic flow causes congestion, delays, stop-and-go conditions (i.e. excessive acceleration/deceleration) and longer journey times. Vehicle and infrastructure connectivity offers the potential to provide improved new services with additional functions of assisting drivers. This paper focuses on one of the applications of vehicle-to-infrastructure communication namely Green Light Optimal Speed Advisory (GLOSA). To assess the effectiveness of GLOSA in the urban road network, an integrated microscopic traffic simulation framework is built into VISSIM software. Vehicle movements and vehicle-infrastructure communications are simulated through the interface of External Driver Model. A control algorithm is developed for recommending an optimal speed that is continuously updated in every time step for all vehicles approaching a signal-controlled point. This algorithm allows vehicles to pass a traffic signal without stopping or to minimise stopping times at a red phase. This study is performed with all connected vehicles at 100% penetration rate. Conventional vehicles are also simulated in the same network as a reference. A straight road segment composed of two opposite directions with two traffic lights per lane is studied. The simulation is implemented under 150 vehicles per hour and 200 per hour traffic volume conditions to identify how different traffic densities influence the benefits of GLOSA. The results indicate that traffic flow is improved by the application of GLOSA. According to this study, vehicles passed through the traffic lights more smoothly, and waiting times were reduced by up to 28 seconds. Average delays decreased for the entire network by 86.46% and 83.84% under traffic densities of 150 vehicles per hour per lane and 200 vehicles per hour per lane, respectively.
Keywords: Connected vehicles, GLOSA, intelligent transportation systems, infrastructure-to-vehicle communication.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16131181 A New Heuristic Approach for Large Size Zero-One Multi Knapsack Problem Using Intercept Matrix
Authors: K. Krishna Veni, S. Raja Balachandar
Abstract:
This paper presents a heuristic to solve large size 0-1 Multi constrained Knapsack problem (01MKP) which is NP-hard. Many researchers are used heuristic operator to identify the redundant constraints of Linear Programming Problem before applying the regular procedure to solve it. We use the intercept matrix to identify the zero valued variables of 01MKP which is known as redundant variables. In this heuristic, first the dominance property of the intercept matrix of constraints is exploited to reduce the search space to find the optimal or near optimal solutions of 01MKP, second, we improve the solution by using the pseudo-utility ratio based on surrogate constraint of 01MKP. This heuristic is tested for benchmark problems of sizes upto 2500, taken from literature and the results are compared with optimum solutions. Space and computational complexity of solving 01MKP using this approach are also presented. The encouraging results especially for relatively large size test problems indicate that this heuristic can successfully be used for finding good solutions for highly constrained NP-hard problems.
Keywords: 0-1 Multi constrained Knapsack problem, heuristic, computational complexity, NP-Hard problems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18571180 Bayesian Network Model for Students- Laboratory Work Performance Assessment: An Empirical Investigation of the Optimal Construction Approach
Authors: Ifeyinwa E. Achumba, Djamel Azzi, Rinat Khusainov
Abstract:
There are three approaches to complete Bayesian Network (BN) model construction: total expert-centred, total datacentred, and semi data-centred. These three approaches constitute the basis of the empirical investigation undertaken and reported in this paper. The objective is to determine, amongst these three approaches, which is the optimal approach for the construction of a BN-based model for the performance assessment of students- laboratory work in a virtual electronic laboratory environment. BN models were constructed using all three approaches, with respect to the focus domain, and compared using a set of optimality criteria. In addition, the impact of the size and source of the training, on the performance of total data-centred and semi data-centred models was investigated. The results of the investigation provide additional insight for BN model constructors and contribute to literature providing supportive evidence for the conceptual feasibility and efficiency of structure and parameter learning from data. In addition, the results highlight other interesting themes.Keywords: Bayesian networks, model construction, parameterlearning, structure learning, performance index, model comparison.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17271179 Linear Phase High Pass FIR Filter Design using Improved Particle Swarm Optimization
Authors: Sangeeta Mondal, Vasundhara, Rajib Kar, Durbadal Mandal, S. P. Ghoshal
Abstract:
This paper presents an optimal design of linear phase digital high pass finite impulse response (FIR) filter using Improved Particle Swarm Optimization (IPSO). In the design process, the filter length, pass band and stop band frequencies, feasible pass band and stop band ripple sizes are specified. FIR filter design is a multi-modal optimization problem. An iterative method is introduced to find the optimal solution of FIR filter design problem. Evolutionary algorithms like real code genetic algorithm (RGA), particle swarm optimization (PSO), improved particle swarm optimization (IPSO) have been used in this work for the design of linear phase high pass FIR filter. IPSO is an improved PSO that proposes a new definition for the velocity vector and swarm updating and hence the solution quality is improved. A comparison of simulation results reveals the optimization efficacy of the algorithm over the prevailing optimization techniques for the solution of the multimodal, nondifferentiable, highly non-linear, and constrained FIR filter design problems.Keywords: FIR Filter, IPSO, GA, PSO, Parks and McClellan Algorithm, Evolutionary Optimization, High Pass Filter
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30991178 Experimental Measurements of Mean and Turbulence Quantities behind the Circular Cylinder by Attaching Different Number of Tripping Wires
Authors: Amir Bak Khoshnevis, Mahdieh Khodadadi, Aghil Lotfi
Abstract:
For a bluff body, roughness elements in simulating a turbulent boundary layer, leading to delayed flow separation, a smaller wake, and lower form drag. In the present work, flow past a circular cylinder with using tripping wires is studied experimentally. The wind tunnel used for modeling free stream is open blow circuit (maximum speed = 30m/s and maximum turbulence of free stream = 0.1%). The selected Reynolds number for all tests was constant (Re = 25000). The circular cylinder selected for this experiment is 20 and 400mm in diameter and length, respectively. The aim of this research is to find the optimal operation mode. In this study installed some tripping wires 1mm in diameter, with a different number of wires on the circular cylinder and the wake characteristics of the circular cylinder is studied. Results showed that by increasing number of tripping wires attached to the circular cylinder (6, 8, and 10, respectively), The optimal angle for the tripping wires with 1mm in diameter to be installed on the cylinder is 60̊ (or 6 wires required at angle difference of 60̊). Strouhal number for the cylinder with tripping wires 1mm in diameter at angular position 60̊ showed the maximum value.
Keywords: Wake of a circular cylinder, trip wire, velocity defect, Strouhal number.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9881177 A Nondominated Sorting Genetic Algorithm for Shortest Path Routing Problem
Authors: C. Chitra, P. Subbaraj
Abstract:
The shortest path routing problem is a multiobjective nonlinear optimization problem with constraints. This problem has been addressed by considering Quality of service parameters, delay and cost objectives separately or as a weighted sum of both objectives. Multiobjective evolutionary algorithms can find multiple pareto-optimal solutions in one single run and this ability makes them attractive for solving problems with multiple and conflicting objectives. This paper uses an elitist multiobjective evolutionary algorithm based on the Non-dominated Sorting Genetic Algorithm (NSGA), for solving the dynamic shortest path routing problem in computer networks. A priority-based encoding scheme is proposed for population initialization. Elitism ensures that the best solution does not deteriorate in the next generations. Results for a sample test network have been presented to demonstrate the capabilities of the proposed approach to generate well-distributed pareto-optimal solutions of dynamic routing problem in one single run. The results obtained by NSGA are compared with single objective weighting factor method for which Genetic Algorithm (GA) was applied.
Keywords: Multiobjective optimization, Non-dominated Sorting Genetic Algorithm, Routing, Weighted sum.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19271176 Optimization of Strategies and Models Review for Optimal Technologies - Based On Fuzzy Schemes for Green Architecture
Authors: Ghada Elshafei, Abdelazim Negm
Abstract:
Recently, the green architecture becomes a significant way to a sustainable future. Green building designs involve finding the balance between comfortable homebuilding and sustainable environment. Moreover, the utilization of the new technologies such as artificial intelligence techniques are used to complement current practices in creating greener structures to keep the built environment more sustainable. The most common objectives in green buildings should be designed to minimize the overall impact of the built environment that effect on ecosystems in general and in particularly human health and natural environment. This will lead to protecting occupant health, improving employee productivity, reducing pollution and sustaining the environmental. In green building design, multiple parameters which may be interrelated, contradicting, vague and of qualitative/quantitative nature are broaden to use. This paper presents a comprehensive critical state- ofart- review of current practices based on fuzzy and its combination techniques. Also, presented how green architecture/building can be improved using the technologies that been used for analysis to seek optimal green solutions strategies and models to assist in making the best possible decision out of different alternatives.
Keywords: Green architecture/building, technologies, optimization, strategies, fuzzy techniques and models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25231175 Relation of Optimal Pilot Offsets in the Shifted Constellation-Based Method for the Detection of Pilot Contamination Attacks
Authors: Dimitriya A. Mihaylova, Zlatka V. Valkova-Jarvis, Georgi L. Iliev
Abstract:
One possible approach for maintaining the security of communication systems relies on Physical Layer Security mechanisms. However, in wireless time division duplex systems, where uplink and downlink channels are reciprocal, the channel estimate procedure is exposed to attacks known as pilot contamination, with the aim of having an enhanced data signal sent to the malicious user. The Shifted 2-N-PSK method involves two random legitimate pilots in the training phase, each of which belongs to a constellation, shifted from the original N-PSK symbols by certain degrees. In this paper, legitimate pilots’ offset values and their influence on the detection capabilities of the Shifted 2-N-PSK method are investigated. As the implementation of the technique depends on the relation between the shift angles rather than their specific values, the optimal interconnection between the two legitimate constellations is investigated. The results show that no regularity exists in the relation between the pilot contamination attacks (PCA) detection probability and the choice of offset values. Therefore, an adversary who aims to obtain the exact offset values can only employ a brute-force attack but the large number of possible combinations for the shifted constellations makes such a type of attack difficult to successfully mount. For this reason, the number of optimal shift value pairs is also studied for both 100% and 98% probabilities of detecting pilot contamination attacks. Although the Shifted 2-N-PSK method has been broadly studied in different signal-to-noise ratio scenarios, in multi-cell systems the interference from the signals in other cells should be also taken into account. Therefore, the inter-cell interference impact on the performance of the method is investigated by means of a large number of simulations. The results show that the detection probability of the Shifted 2-N-PSK decreases inversely to the signal-to-interference-plus-noise ratio.
Keywords: Channel estimation, inter-cell interference, pilot contamination attacks, wireless communications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6771174 Statistical Distributions of the Lapped Transform Coefficients for Images
Authors: Vijay Kumar Nath, Deepika Hazarika, Anil Mahanta,
Abstract:
Discrete Cosine Transform (DCT) based transform coding is very popular in image, video and speech compression due to its good energy compaction and decorrelating properties. However, at low bit rates, the reconstructed images generally suffer from visually annoying blocking artifacts as a result of coarse quantization. Lapped transform was proposed as an alternative to the DCT with reduced blocking artifacts and increased coding gain. Lapped transforms are popular for their good performance, robustness against oversmoothing and availability of fast implementation algorithms. However, there is no proper study reported in the literature regarding the statistical distributions of block Lapped Orthogonal Transform (LOT) and Lapped Biorthogonal Transform (LBT) coefficients. This study performs two goodness-of-fit tests, the Kolmogorov-Smirnov (KS) test and the 2- test, to determine the distribution that best fits the LOT and LBT coefficients. The experimental results show that the distribution of a majority of the significant AC coefficients can be modeled by the Generalized Gaussian distribution. The knowledge of the statistical distribution of transform coefficients greatly helps in the design of optimal quantizers that may lead to minimum distortion and hence achieve optimal coding efficiency.
Keywords: Lapped orthogonal transform, Lapped biorthogonal transform, Image compression, KS test,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1605