Search results for: Complexity reduction approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6778

Search results for: Complexity reduction approach

6568 Prediction of Bath Temperature Using Neural Networks

Authors: H. Meradi, S. Bouhouche, M. Lahreche

Abstract:

In this work, we consider an application of neural networks in LD converter. Application of this approach assumes a reliable prediction of steel temperature and reduces a reblow ratio in steel work. It has been applied a conventional model to charge calculation, the obtained results by this technique are not always good, this is due to the process complexity. Difficulties are mainly generated by the noisy measurement and the process non linearities. Artificial Neural Networks (ANNs) have become a powerful tool for these complex applications. It is used a backpropagation algorithm to learn the neural nets. (ANNs) is used to predict the steel bath temperature in oxygen converter process for the end condition. This model has 11 inputs process variables and one output. The model was tested in steel work, the obtained results by neural approach are better than the conventional model.

Keywords: LD converter, bath temperature, neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1789
6567 Adaptive Square-Rooting Companding Technique for PAPR Reduction in OFDM Systems

Authors: Wisam F. Al-Azzo, Borhanuddin Mohd. Ali

Abstract:

This paper addresses the problem of peak-to-average power ratio (PAPR) in orthogonal frequency division multiplexing (OFDM) systems. It also introduces a new PAPR reduction technique based on adaptive square-rooting (SQRT) companding process. The SQRT process of the proposed technique changes the statistical characteristics of the OFDM output signals from Rayleigh distribution to Gaussian-like distribution. This change in statistical distribution results changes of both the peak and average power values of OFDM signals, and consequently reduces significantly the PAPR. For the 64QAM OFDM system using 512 subcarriers, up to 6 dB reduction in PAPR was achieved by square-rooting technique with fixed degradation in bit error rate (BER) equal to 3 dB. However, the PAPR is reduced at the expense of only -15 dB out-ofband spectral shoulder re-growth below the in-band signal level. The proposed adaptive SQRT technique is superior in terms of BER performance than the original, non-adaptive, square-rooting technique when the required reduction in PAPR is no more than 5 dB. Also, it provides fixed amount of PAPR reduction in which it is not available in the original SQRT technique.

Keywords: complementary cumulative distribution function(CCDF), OFDM, peak-to-average power ratio (PAPR), adaptivesquare-rooting PAPR reduction technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2153
6566 Towards a Simulation Model to Ensure the Availability of Machines in Maintenance Activities

Authors: Maryam Gallab, Hafida Bouloiz, Youness Chater, Mohamed Tkiouat

Abstract:

The aim of this paper is to present a model based on multi-agent systems in order to manage the maintenance activities and to ensure the reliability and availability of machines just with the required resources (operators, tools). The interest of the simulation is to solve the complexity of the system and to find results without cost or wasting time. An implementation of the model is carried out on the AnyLogic platform to display the defined performance indicators.

Keywords: Maintenance, complexity, simulation, multi-agent systems, AnyLogic platform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1463
6565 A New Distribution Network Reconfiguration Approach using a Tree Model

Authors: E. Dolatdar, S. Soleymani, B. Mozafari

Abstract:

Power loss reduction is one of the main targets in power industry and so in this paper, the problem of finding the optimal configuration of a radial distribution system for loss reduction is considered. Optimal reconfiguration involves the selection of the best set of branches to be opened ,one each from each loop, for reducing resistive line losses , and reliving overloads on feeders by shifting the load to adjacent feeders. However ,since there are many candidate switching combinations in the system ,the feeder reconfiguration is a complicated problem. In this paper a new approach is proposed based on a simple optimum loss calculation by determining optimal trees of the given network. From graph theory a distribution network can be represented with a graph that consists a set of nodes and branches. In fact this problem can be viewed as a problem of determining an optimal tree of the graph which simultaneously ensure radial structure of each candidate topology .In this method the refined genetic algorithm is also set up and some improvements of algorithm are made on chromosome coding. In this paper an implementation of the algorithm presented by [7] is applied by modifying in load flow program and a comparison of this method with the proposed method is employed. In [7] an algorithm is proposed that the choice of the switches to be opened is based on simple heuristic rules. This algorithm reduce the number of load flow runs and also reduce the switching combinations to a fewer number and gives the optimum solution. To demonstrate the validity of these methods computer simulations with PSAT and MATLAB programs are carried out on 33-bus test system. The results show that the performance of the proposed method is better than [7] method and also other methods.

Keywords: Distribution System, Reconfiguration, Loss Reduction , Graph Theory , Optimization , Genetic Algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3739
6564 Synthesis of Gold Nanoparticles Stabilized in Na-Montmorillonite for Nitrophenol Reduction

Authors: F. Ammari, M. Chenouf

Abstract:

Synthesis of gold nanoparticles has attracted much attention since the pioneering discovery of the high catalytic activity of supported gold nanoparticles in the reaction of CO oxidation at low temperature. In this research field, we used Na-montmorillonite for gold nanoparticles stabilization; various gold loading percentage 1, 2 and 5% were used for gold nanoparticles preparation. The gold nanoparticles were obtained using chemical reduction method using NaBH4 as reductant agent. The obtained gold nanoparticles stabilized in Na-montmorillonite were used as catalysts for the reduction of 4- nitrophenol to aminophenol with sodium borohydride at room temperature. The UV-Vis results confirmed directly the gold nanoparticles formation. The XRD and N2 adsorption results showed the formation of gold nanoparticles in the pores of montmorillonite with an average size of 5 nm obtained on samples with 2% gold loading percentage. The gold particles size increased with the increase of gold loading percentage. The reduction reaction of 4- nitrophenol into 4-aminophenol with NaBH4 catalyzed by Au-Namontmorillonite catalyst exhibits remarkably a high activity; the reaction was completed within 9 min for 1%Au-Na-montmorillonite and within 3 min for 2%Au-Na-montmorillonite.

Keywords: Chemical reduction, gold, montmorillonite, nanoparticles, 4-nitrophenol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2066
6563 Promoting Complex Systems Learning through the use of Computer Modeling

Authors: Kamel Hashem, David Mioduser

Abstract:

This paper describes part of a project about Learningby- Modeling (LbM). Studying complex systems is increasingly important in teaching and learning many science domains. Many features of complex systems make it difficult for students to develop deep understanding. Previous research indicates that involvement with modeling scientific phenomena and complex systems can play a powerful role in science learning. Some researchers argue with this view indicating that models and modeling do not contribute to understanding complexity concepts, since these increases the cognitive load on students. This study will investigate the effect of different modes of involvement in exploring scientific phenomena using computer simulation tools, on students- mental model from the perspective of structure, behavior and function. Quantitative and qualitative methods are used to report about 121 freshmen students that engaged in participatory simulations about complex phenomena, showing emergent, self-organized and decentralized patterns. Results show that LbM plays a major role in students' concept formation about complexity concepts.

Keywords: Complexity, Educational technology, Learning by modeling, Mental models

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535
6562 The Knapsack Sharing Problem: A Tree Search Exact Algorithm

Authors: Mhand Hifi, Hedi Mhalla

Abstract:

In this paper, we study the knapsack sharing problem, a variant of the well-known NP-Hard single knapsack problem. We investigate the use of a tree search for optimally solving the problem. The used method combines two complementary phases: a reduction interval search phase and a branch and bound procedure one. First, the reduction phase applies a polynomial reduction strategy; that is used for decomposing the problem into a series of knapsack problems. Second, the tree search procedure is applied in order to attain a set of optimal capacities characterizing the knapsack problems. Finally, the performance of the proposed optimal algorithm is evaluated on a set of instances of the literature and its runtime is compared to the best exact algorithm of the literature.

Keywords: Branch and bound, combinatorial optimization, knap¬sack, knapsack sharing, heuristics, interval reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1519
6561 Evaluating Complexity – Ethical Challenges in Computational Design Processes

Authors: J.Partanen

Abstract:

Complexity, as a theoretical background has made it easier to understand and explain the features and dynamic behavior of various complex systems. As the common theoretical background has confirmed, borrowing the terminology for design from the natural sciences has helped to control and understand urban complexity. Phenomena like self-organization, evolution and adaptation are appropriate to describe the formerly inaccessible characteristics of the complex environment in unpredictable bottomup systems. Increased computing capacity has been a key element in capturing the chaotic nature of these systems. A paradigm shift in urban planning and architectural design has forced us to give up the illusion of total control in urban environment, and consequently to seek for novel methods for steering the development. New methods using dynamic modeling have offered a real option for more thorough understanding of complexity and urban processes. At best new approaches may renew the design processes so that we get a better grip on the complex world via more flexible processes, support urban environmental diversity and respond to our needs beyond basic welfare by liberating ourselves from the standardized minimalism. A complex system and its features are as such beyond human ethics. Self-organization or evolution is either good or bad. Their mechanisms are by nature devoid of reason. They are common in urban dynamics in both natural processes and gas. They are features of a complex system, and they cannot be prevented. Yet their dynamics can be studied and supported. The paradigm of complexity and new design approaches has been criticized for a lack of humanity and morality, but the ethical implications of scientific or computational design processes have not been much discussed. It is important to distinguish the (unexciting) ethics of the theory and tools from the ethics of computer aided processes based on ethical decisions. Urban planning and architecture cannot be based on the survival of the fittest; however, the natural dynamics of the system cannot be impeded on grounds of being “non-human". In this paper the ethical challenges of using the dynamic models are contemplated in light of a few examples of new architecture and dynamic urban models and literature. It is suggested that ethical challenges in computational design processes could be reframed under the concepts of responsibility and transparency.

Keywords: urban planning, architecture, dynamic modeling, ethics, complexity theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1849
6560 A Two-Stage Expert System for Diagnosis of Leukemia Based on Type-2 Fuzzy Logic

Authors: Ali Akbar Sadat Asl

Abstract:

Diagnosis and deciding about diseases in medical fields is facing innate uncertainty which can affect the whole process of treatment. This decision is made based on expert knowledge and the way in which an expert interprets the patient's condition, and the interpretation of the various experts from the patient's condition may be different. Fuzzy logic can provide mathematical modeling for many concepts, variables, and systems that are unclear and ambiguous and also it can provide a framework for reasoning, inference, control, and decision making in conditions of uncertainty. In systems with high uncertainty and high complexity, fuzzy logic is a suitable method for modeling. In this paper, we use type-2 fuzzy logic for uncertainty modeling that is in diagnosis of leukemia. The proposed system uses an indirect-direct approach and consists of two stages: In the first stage, the inference of blood test state is determined. In this step, we use an indirect approach where the rules are extracted automatically by implementing a clustering approach. In the second stage, signs of leukemia, duration of disease until its progress and the output of the first stage are combined and the final diagnosis of the system is obtained. In this stage, the system uses a direct approach and final diagnosis is determined by the expert. The obtained results show that the type-2 fuzzy expert system can diagnose leukemia with the average accuracy about 97%.

Keywords: Expert system, leukemia, medical diagnosis, type-2 fuzzy logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 976
6559 An Optimal Feature Subset Selection for Leaf Analysis

Authors: N. Valliammal, S.N. Geethalakshmi

Abstract:

This paper describes an optimal approach for feature subset selection to classify the leaves based on Genetic Algorithm (GA) and Kernel Based Principle Component Analysis (KPCA). Due to high complexity in the selection of the optimal features, the classification has become a critical task to analyse the leaf image data. Initially the shape, texture and colour features are extracted from the leaf images. These extracted features are optimized through the separate functioning of GA and KPCA. This approach performs an intersection operation over the subsets obtained from the optimization process. Finally, the most common matching subset is forwarded to train the Support Vector Machine (SVM). Our experimental results successfully prove that the application of GA and KPCA for feature subset selection using SVM as a classifier is computationally effective and improves the accuracy of the classifier.

Keywords: Optimization, Feature extraction, Feature subset, Classification, GA, KPCA, SVM and Computation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2186
6558 Ductility, Rμ, and Overstrength Factors for V Braced Reinforced Concrete Buildings

Authors: Birendra Kumar Bohara

Abstract:

Steel bracings are used to improve the seismic behaviors of the structures. In this study, 8, 12 and 16 story reinforced concrete (RC) buildings with steel bracings are used in three base shear contributions (25%, 50% and 75%) in the columns. With the help of pushover analysis and capacity curves, the overstrength factors, ductility factors and ductility reduction factors are investigated for braced RC buildings. It is observed that when the base shear contribution in the columns increases the ductility reduction factor also increases. The results show that when the time period of the structures increases, the ductility reduction factors of the structures decrease.

Keywords: Steel bracing, overstrength factor, ductility, ductility reduction factors, base shear contributions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 329
6557 Methane versus Carbon Dioxide: Mitigation Prospects

Authors: Alexander J. Severinsky, Allen L. Sessoms

Abstract:

Atmospheric carbon dioxide (CO2) has dominated the discussion around the causes of climate change. This is a reflection of a 100-year time horizon for all greenhouse gases that became a norm.  The 100-year time horizon is much too long – and yet, almost all mitigation efforts, including those set in the near-term frame of within 30 years, are still geared toward it. In this paper, we show that for a 30-year time horizon, methane (CH4) is the greenhouse gas whose radiative forcing exceeds that of CO2. In our analysis, we use the radiative forcing of greenhouse gases in the atmosphere, because they directly affect the rise in temperature on Earth. We found that in 2019, the radiative forcing (RF) of methane was ~2.5 W/m2 and that of carbon dioxide was ~2.1 W/m2. Under a business-as-usual (BAU) scenario until 2050, such forcing would be ~2.8 W/m2 and ~3.1 W/m2 respectively. There is a substantial spread in the data for anthropogenic and natural methane (CH4) emissions, along with natural gas, (which is primarily CH4), leakages from industrial production to consumption. For this reason, we estimate the minimum and maximum effects of a reduction of these leakages, and assume an effective immediate reduction by 80%. Such action may serve to reduce the annual radiative forcing of all CH4 emissions by ~15% to ~30%. This translates into a reduction of RF by 2050 from ~2.8 W/m2 to ~2.5 W/m2 in the case of the minimum effect that can be expected, and to ~2.15 W/m2 in the case of the maximum effort to reduce methane leakages. Under the BAU, we find that the RF of CO2 will increase from ~2.1 W/m2 now to ~3.1 W/m2 by 2050. We assume a linear reduction of 50% in anthropogenic emission over the course of the next 30 years, which would reduce the radiative forcing of CO2 from ~3.1 W/m2 to ~2.9 W/m2. In the case of "net zero," the other 50% of only anthropogenic CO2 emissions reduction would be limited to being either from sources of emissions or directly from the atmosphere. In this instance, the total reduction would be from ~3.1 W/m2 to ~2.7 W/m2, or ~0.4 W/m2. To achieve the same radiative forcing as in the scenario of maximum reduction of methane leakages of ~2.15 W/m2, an additional reduction of radiative forcing of CO2 would be approximately 2.7 -2.15 = 0.55 W/m2. In total, one would need to remove ~660 GT of CO2 from the atmosphere in order to match the maximum reduction of current methane leakages, and ~270 GT of CO2 from emitting sources, to reach "negative emissions". This amounts to over 900 GT of CO2.

Keywords: Methane Leakages, Methane Radiative Forcing, Methane Mitigation, Methane Net Zero.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 534
6556 The Effects of Different Level Cluster Tip Reduction and Foliar Boric Acid Applications on Yield and Yield Components of Italia Grape Cultivar

Authors: A. Akin

Abstract:

This study was carried out on Italia grape variety (Vitis vinifera L.) in Konya province, Turkey in 2016. The cultivar is five years old and grown on 1103 Paulsen rootstock. It was determined the effects of applications of the Control (C), 1/3 Cluster Tip Reduction (1/3 CTR), 1/6 Cluster Tip Reduction (1/6 CTR), 1/9 Cluster Tip Reduction (1/9 CTR), 1/3 CTR+Boric Acid (BA), 1/6 CTR+BA, 1/9 CTR+BA, on yield and yield components of the Italia grape variety. The results were obtained as the highest fresh grape yield (4.74 g) with 1/9 CTR+BA application; the highest cluster weight (220.08 g) with 1/3 CTR application; the highest 100 berry weight (565.85 g) with 1/9 CTR+BA application; as the highest maturity index (49.28) with 1/9 CTR+BA application; as the highest must yield (685.33 ml/kg) with 1/3 CTR+BA and (685.33 ml/kg) with 1/9 CTR+BA applications. To increase the fresh grape yield, 100 berry weight and maturity index in the Italia grape variety, the 1/9 CTR+BA application can be recommended.

Keywords: Italia grape variety, boric acid, cluster tip reduction, yield, yield components.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 960
6555 Low Complexity Multi Mode Interleaver Core for WiMAX with Support for Convolutional Interleaving

Authors: Rizwan Asghar, Dake Liu

Abstract:

A hardware efficient, multi mode, re-configurable architecture of interleaver/de-interleaver for multiple standards, like DVB, WiMAX and WLAN is presented. The interleavers consume a large part of silicon area when implemented by using conventional methods as they use memories to store permutation patterns. In addition, different types of interleavers in different standards cannot share the hardware due to different construction methodologies. The novelty of the work presented in this paper is threefold: 1) Mapping of vital types of interleavers including convolutional interleaver onto a single architecture with flexibility to change interleaver size; 2) Hardware complexity for channel interleaving in WiMAX is reduced by using 2-D realization of the interleaver functions; and 3) Silicon cost overheads reduced by avoiding the use of small memories. The proposed architecture consumes 0.18mm2 silicon area for 0.12μm process and can operate at a frequency of 140 MHz. The reduced complexity helps in minimizing the memory utilization, and at the same time provides strong support to on-the-fly computation of permutation patterns.

Keywords: Hardware interleaver implementation, WiMAX, DVB, block interleaver, convolutional interleaver, hardwaremultiplexing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1987
6554 Robust Numerical Scheme for Pricing American Options under Jump Diffusion Models

Authors: Salah Alrabeei, Mohammad Yousuf

Abstract:

The goal of option pricing theory is to help the investors to manage their money, enhance returns and control their financial future by theoretically valuing their options. However, most of the option pricing models have no analytical solution. Furthermore, not all the numerical methods are efficient to solve these models because they have nonsmoothing payoffs or discontinuous derivatives at the exercise price. In this paper, we solve the American option under jump diffusion models by using efficient time-dependent numerical methods. several techniques are integrated to reduced the overcome the computational complexity. Fast Fourier Transform (FFT) algorithm is used as a matrix-vector multiplication solver, which reduces the complexity from O(M2) into O(M logM). Partial fraction decomposition technique is applied to rational approximation schemes to overcome the complexity of inverting polynomial of matrices. The proposed method is easy to implement on serial or parallel versions. Numerical results are presented to prove the accuracy and efficiency of the proposed method.

Keywords: Integral differential equations, American options, jump–diffusion model, rational approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 497
6553 Stereoselective Reduction of Amino Ketone with Sodium Borohydride in the Presence of Metal Chloride. A Simple Pathway to S-Propranolol

Authors: R. Inkum, A. Teerawutgulrag, P. Puangsombat, N. Rakariyatham

Abstract:

Propranolol is worldwide hypertension drug that is active in S-isomer. Patients must use this drug throughout their lives, and this action employsa significant level of expenditure. A simpler synthesis and lower cost can reduce the price for the patient. A sis pathway of S-propranolol starting from protection of (R,S)-propranolol with di-t-butyldicarbonate and then the product is oxidized with pyridiniumchlorochromate. The selective reduction of ketone occurrs with sodiumborohydride in the presence of metal chloride provided S-propranolol.

Keywords: S-propranolol, selective reduction, sodium borohydride, metal chloride

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2107
6552 Loss Reduction and Reliability Improvement of Industrial Distribution System through Network Reconfiguration

Authors: Ei Ei Phyu, Kyaw Myo Lin, Thin Thin Moe

Abstract:

The paper presents an approach to improve the reliability and reduce line losses of practical distribution system applying network reconfiguration. The change of the topology redirects the power flow within the distribution network to obtain better performance of the system. Practical distribution network (Pyigyitagon Industrial Zone (I)) is used as the case study network. The detailed calculations of the reliability indices are done by using analytical method and power flow calculation is performed by Newton-Rephason solver. The comparison of various network reconfiguration techniques are described with respect to power loss and reliability index levels. Finally, the optimal reconfigured network is selected among difference cases based on the two factors: the most reliable network and the least loss minimization.

Keywords: Distribution system reliability, loss reduction, network reconfiguration, reliability enhancement, reliability indices.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 842
6551 Model-Based Software Regression Test Suite Reduction

Authors: Shiwei Deng, Yang Bao

Abstract:

In this paper, we present a model-based regression test suite reducing approach that uses EFSM model dependence analysis and probability-driven greedy algorithm to reduce software regression test suites. The approach automatically identifies the difference between the original model and the modified model as a set of elementary model modifications. The EFSM dependence analysis is performed for each elementary modification to reduce the regression test suite, and then the probability-driven greedy algorithm is adopted to select the minimum set of test cases from the reduced regression test suite that cover all interaction patterns. Our initial experience shows that the approach may significantly reduce the size of regression test suites.

Keywords: Dependence analysis, EFSM model, greedy algorithm, regression test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1879
6550 Optimal Solution of Constraint Satisfaction Problems

Authors: Jeffrey L. Duffany

Abstract:

An optimal solution for a large number of constraint satisfaction problems can be found using the technique of substitution and elimination of variables analogous to the technique that is used to solve systems of equations. A decision function f(A)=max(A2) is used to determine which variables to eliminate. The algorithm can be expressed in six lines and is remarkable in both its simplicity and its ability to find an optimal solution. However it is inefficient in that it needs to square the updated A matrix after each variable elimination. To overcome this inefficiency the algorithm is analyzed and it is shown that the A matrix only needs to be squared once at the first step of the algorithm and then incrementally updated for subsequent steps, resulting in significant improvement and an algorithm complexity of O(n3).

Keywords: Algorithm, complexity, constraint, np-complete.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1391
6549 An Efficient Multi Join Algorithm Utilizing a Lattice of Double Indices

Authors: Hanan A. M. Abd Alla, Lilac A. E. Al-Safadi

Abstract:

In this paper, a novel multi join algorithm to join multiple relations will be introduced. The novel algorithm is based on a hashed-based join algorithm of two relations to produce a double index. This is done by scanning the two relations once. But instead of moving the records into buckets, a double index will be built. This will eliminate the collision that can happen from a complete hash algorithm. The double index will be divided into join buckets of similar categories from the two relations. The algorithm then joins buckets with similar keys to produce joined buckets. This will lead at the end to a complete join index of the two relations. without actually joining the actual relations. The time complexity required to build the join index of two categories is Om log m where m is the size of each category. Totaling time complexity to O n log m for all buckets. The join index will be used to materialize the joined relation if required. Otherwise, it will be used along with other join indices of other relations to build a lattice to be used in multi-join operations with minimal I/O requirements. The lattice of the join indices can be fitted into the main memory to reduce time complexity of the multi join algorithm.

Keywords: Multi join, Relation, Lattice, Join indices.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1251
6548 SLM Using Riemann Sequence Combined with DCT Transform for PAPR Reduction in OFDM Communication Systems

Authors: Pepin Magnangana Zoko Goyoro, Ibrahim James Moumouni, Sroy Abouty

Abstract:

Orthogonal Frequency Division Multiplexing (OFDM) is an efficient method of data transmission for high speed communication systems. However, the main drawback of OFDM systems is that, it suffers from the problem of high Peak-to-Average Power Ratio (PAPR) which causes inefficient use of the High Power Amplifier and could limit transmission efficiency. OFDM consist of large number of independent subcarriers, as a result of which the amplitude of such a signal can have high peak values. In this paper, we propose an effective reduction scheme that combines DCT and SLM techniques. The scheme is composed of the DCT followed by the SLM using the Riemann matrix to obtain phase sequences for the SLM technique. The simulation results show PAPR can be greatly reduced by applying the proposed scheme. In comparison with OFDM, while OFDM had high values of PAPR –about 10.4dB our proposed method achieved about 4.7dB reduction of the PAPR with low complexities computation. This approach also avoids randomness in phase sequence selection, which makes it simpler to decode at the receiver. As an added benefit, the matrices can be generated at the receiver end to obtain the data signal and hence it is not required to transmit side information (SI).

Keywords: DCT transform, OFDM, PAPR, Riemann matrix, SLM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2595
6547 Estimation of Component Reusability through Reusability Metrics

Authors: Aditya Pratap Singh, Pradeep Tomar

Abstract:

Software reusability is an essential characteristic of Component-Based Software (CBS). The component reusability is an important assess for the effective reuse of components in CBS. The attributes of reusability proposed by various researchers are studied and four of them are identified as potential factors affecting reusability. This paper proposes metric for reusability estimation of black-box software component along with metrics for Interface Complexity, Understandability, Customizability and Reliability. An experiment is performed for estimation of reusability through a case study on a sample web application using a real world component.

Keywords: Component-based software, component reusability, customizability, interface complexity, reliability, understandability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3016
6546 Reduction of Search Space by Applying Controlled Genetic Operators for Weight Constrained Shortest Path Problem

Authors: A.K.M. Khaled Ahsan Talukder, Taibun Nessa, Kaushik Roy

Abstract:

The weight constrained shortest path problem (WCSPP) is one of most several known basic problems in combinatorial optimization. Because of its importance in many areas of applications such as computer science, engineering and operations research, many researchers have extensively studied the WCSPP. This paper mainly concentrates on the reduction of total search space for finding WCSP using some existing Genetic Algorithm (GA). For this purpose, some controlled schemes of genetic operators are adopted on list chromosome representation. This approach gives a near optimum solution with smaller elapsed generation than classical GA technique. From further analysis on the matter, a new generalized schema theorem is also developed from the philosophy of Holland-s theorem.

Keywords: Genetic Algorithm, Evolutionary Optimization, Multi Objective Optimization, Non-linear Schema Theorem, WCSPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1373
6545 Numerical Study of Base Drag Reduction Using Locked Vortex Flow Management Technique for Lower Subsonic Regime

Authors: Kailas S. Jagtap, Karthik Sundarraj, Nirmal Kumar, S. Rajnarasimha, Prakash S. Kulkarni

Abstract:

The issue of turbulence base streams and the drag related to it have been of important attention for rockets, missiles, and aircraft. Different techniques are used for base drag reduction. This paper presents the numerical study of numerous drag reduction technique. The base drag or afterbody drag of bluff bodies can be reduced easily using locked vortex drag reduction technique. For bluff bodies having a cylindrical shape, the base drag is much larger compared to streamlined bodies. For such bodies using splitter plates, the vortex can be trapped between the base and the plate, which results in smooth flow. Splitter plate with round and curved corner shapes has influence in drag reduction. In this paper, the comparison is done between single splitter plate as different positions and with the bluff body. Base drag for the speed of 30m/s can be reduced about 20% to 30% by using single splitter plate as compared to the bluff body.

Keywords: Base drag, bluff body, splitter plate, vortex flow, ANSYS, Fluent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 872
6544 Characteristics of Wall Thickness Increase in Pipe Reduction Process using Planetary Rolls

Authors: Yuji Kotani, Shunsuke Kanai, Hisaki Watari

Abstract:

In recent years, global warming has become a worldwide problem. The reduction of carbon dioxide emissions is a top priority for many companies in the manufacturing industry. In the automobile industry as well, the reduction of carbon dioxide emissions is one of the most important issues. Technology to reduce the weight of automotive parts improves the fuel economy of automobiles, and is an important technology for reducing carbon dioxide. Also, even if this weight reduction technology is applied to electric automobiles rather than gasoline automobiles, reducing energy consumption remains an important issue. Plastic processing of hollow pipes is one important technology for realizing the weight reduction of automotive parts. Ohashi et al. [1],[2] present an example of research on pipe formation in which a process was carried out to enlarge a pipe diameter using a lost core, achieving the suppression of wall thickness reduction and greater pipe expansion than hydroforming. In this study, we investigated a method to increase the wall thickness of a pipe through pipe compression using planetary rolls. The establishment of a technology whereby the wall thickness of a pipe can be controlled without buckling the pipe is an important technology for the weight reduction of products. Using the finite element analysis method, we predicted that it would be possible to increase the compression of an aluminum pipe with a 3mm wall thickness by approximately 20%, and wall thickness by approximately 20% by pressing the hollow pipe with planetary rolls.

Keywords: Pipe-Forming, Wall Thickness, Finite-element-method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2939
6543 Reasons for the Slow Uptake of Embodied Carbon Estimation in the Sri Lankan Building Sector

Authors: Amalka Nawarathna, Nirodha Fernando, Zaid Alwan

Abstract:

Global carbon reduction is not merely a responsibility of environmentally advanced developed countries, but also a responsibility of developing countries regardless of their less impact on global carbon emissions. In recognition of that, Sri Lanka as a developing country has initiated promoting green building construction as one reduction strategy. However, notwithstanding the increasing attention on Embodied Carbon (EC) reduction in the global building sector, they still mostly focus on Operational Carbon (OC) reduction (through improving operational energy). An adequate attention has not yet been given on EC estimation and reduction. Therefore, this study aims to identify the reasons for the slow uptake of EC estimation in the Sri Lankan building sector. To achieve this aim, 16 numbers of global barriers to estimate EC were identified through existing literature. They were then subjected to a pilot survey to identify the significant reasons for the slow uptake of EC estimation in the Sri Lankan building sector. A questionnaire with a three-point Likert scale was used to this end. The collected data were analysed using descriptive statistics. The findings revealed that 11 out of 16 challenges/ barriers are highly relevant as reasons for the slow uptake in estimating EC in buildings in Sri Lanka while the other five challenges/ barriers remain as moderately relevant reasons. Further, the findings revealed that there are no low relevant reasons. Eventually, the paper concluded that all the known reasons are significant to the Sri Lankan building sector and it is necessary to address them in order to upturn the attention on EC reduction.

Keywords: Embodied carbon emissions, embodied carbon estimation, global carbon reduction, Sri Lankan building sector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 787
6542 An Approach for Blind Source Separation using the Sliding DFT and Time Domain Independent Component Analysis

Authors: Koji Yamanouchi, Masaru Fujieda, Takahiro Murakami, Yoshihisa Ishida

Abstract:

''Cocktail party problem'' is well known as one of the human auditory abilities. We can recognize the specific sound that we want to listen by this ability even if a lot of undesirable sounds or noises are mixed. Blind source separation (BSS) based on independent component analysis (ICA) is one of the methods by which we can separate only a special signal from their mixed signals with simple hypothesis. In this paper, we propose an online approach for blind source separation using the sliding DFT and the time domain independent component analysis. The proposed method can reduce calculation complexity in comparison with conventional methods, and can be applied to parallel processing by using digital signal processors (DSPs) and so on. We evaluate this method and show its availability.

Keywords: Cocktail party problem, blind Source Separation(BSS), independent component analysis, sliding DFT, onlineprocessing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1596
6541 M-ary Chaotic Sequence Based SLM-OFDM System for PAPR Reduction without Side-Information

Authors: A.Goel, M. Agrawal, P. Gupta Poddar

Abstract:

Selected Mapping (SLM) is a PAPR reduction technique, which converts the OFDM signal into several independent signals by multiplication with the phase sequence set and transmits one of the signals with lowest PAPR. But it requires the index of the selected signal i.e. side information (SI) to be transmitted with each OFDM symbol. The PAPR reduction capability of the SLM scheme depends on the selection of phase sequence set. In this paper, we have proposed a new phase sequence set generation scheme based on M-ary chaotic sequence and a mapping scheme to map quaternary data to concentric circle constellation (CCC) is used. It is shown that this method does not require SI and provides better SER performance with good PAPR reduction capability as compared to existing SLMOFDM methods.

Keywords: Orthogonal frequency division multiplexing (OFDM), Peak-to-average power ratio (PAPR), Selected mapping (SLM), Side information (SI)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1915
6540 Assessment of Pollution Reduction

Authors: Katarzyna Strzała-Osuch

Abstract:

Environmental investments, including ecological projects, relating to the protection of atmosphere are today a need. However, investing in the environment should be based on rational management rules. This comes across a problem of selecting a method to assess substances reduced during projects. Therefore, a method allowing for the assessment of decision rationality has to be found. The purpose of this article is to present and systematise pollution reduction assessment methods and illustrate theoretical analyses with empirical data. Empirical results confirm theoretical considerations, which proved that the only method for judging pollution reduction, free of apparent disadvantages, is the Eco 99-ratio method. To make decisions on environmental projects, financing institutions should take into account a rationality rule. Therefore the Eco 99-ratio method could be applied to make decisions relating to environmental investments in the area of air protection.

Keywords: Assessment of pollution reduction, costs of environmental protection, efficiency of environmental investments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1286
6539 Motion Planning and Posture Control of the General 3-Trailer System

Authors: K. Raghuwaiya, B. Sharma, J. Vanualailai

Abstract:

This paper presents a set of artificial potential field functions that improves upon, in general, the motion planning and posture control, with theoretically guaranteed point and posture stabilities, convergence and collision avoidance properties of the general3-trailer system in a priori known environment. We basically design and inject two new concepts; ghost walls and the distance optimization technique (DOT) to strengthen point and posture stabilities, in the sense of Lyapunov, of our dynamical model. This new combination of techniques emerges as a convenient mechanism for obtaining feasible orientations at the target positions with an overall reduction in the complexity of the navigation laws. Simulations are provided to demonstrate the effectiveness of the controls laws.

Keywords: Artificial potential fields, 3-trailer systems, motion planning, posture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2109