Search results for: finite element method (fem)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20813

Search results for: finite element method (fem)

17123 Performance Estimation of Small Scale Wind Turbine Rotor for Very Low Wind Regime Condition

Authors: Vilas Warudkar, Dinkar Janghel, Siraj Ahmed

Abstract:

Rapid development experienced by India requires huge amount of energy. Actual supply capacity additions have been consistently lower than the targets set by the government. According to World Bank 40% of residences are without electricity. In 12th five year plan 30 GW grid interactive renewable capacity is planned in which 17 GW is Wind, 10 GW is from solar and 2.1 GW from small hydro project, and rest is compensated by bio gas. Renewable energy (RE) and energy efficiency (EE) meet not only the environmental and energy security objectives, but also can play a crucial role in reducing chronic power shortages. In remote areas or areas with a weak grid, wind energy can be used for charging batteries or can be combined with a diesel engine to save fuel whenever wind is available. India according to IEC 61400-1 belongs to class IV Wind Condition; it is not possible to set up wind turbine in large scale at every place. So, the best choice is to go for small scale wind turbine at lower height which will have good annual energy production (AEP). Based on the wind characteristic available at MANIT Bhopal, rotor for small scale wind turbine is designed. Various Aero foil data is reviewed for selection of airfoil in the Blade Profile. Airfoil suited of Low wind conditions i.e. at low Reynold’s number is selected based on Coefficient of Lift, Drag and angle of attack. For designing of the rotor blade, standard Blade Element Momentum (BEM) Theory is implanted. Performance of the Blade is estimated using BEM theory in which axial induction factor and angular induction factor is optimized using iterative technique. Rotor performance is estimated for particular designed blade specifically for low wind Conditions. Power production of rotor is determined at different wind speeds for particular pitch angle of the blade. At pitch 15o and velocity 5 m/sec gives good cut in speed of 2 m/sec and power produced is around 350 Watts. Tip speed of the Blade is considered as 6.5 for which Coefficient of Performance of the rotor is calculated 0.35, which is good acceptable value for Small scale Wind turbine. Simple Load Model (SLM, IEC 61400-2) is also discussed to improve the structural strength of the rotor. In SLM, Edge wise Moment and Flap Wise moment is considered which cause bending stress at the root of the blade. Various Load case mentioned in the IEC 61400-2 is calculated and checked for the partial safety factor of the wind turbine blade.

Keywords: annual energy production, Blade Element Momentum Theory, low wind Conditions, selection of airfoil

Procedia PDF Downloads 328
17122 A Ground Structure Method to Minimize the Total Installed Cost of Steel Frame Structures

Authors: Filippo Ranalli, Forest Flager, Martin Fischer

Abstract:

This paper presents a ground structure method to optimize the topology and discrete member sizing of steel frame structures in order to minimize total installed cost, including material, fabrication and erection components. The proposed method improves upon existing cost-based ground structure methods by incorporating constructability considerations well as satisfying both strength and serviceability constraints. The architecture for the method is a bi-level Multidisciplinary Feasible (MDF) architecture in which the discrete member sizing optimization is nested within the topology optimization process. For each structural topology generated, the sizing optimization process seek to find a set of discrete member sizes that result in the lowest total installed cost while satisfying strength (member utilization) and serviceability (node deflection and story drift) criteria. To accurately assess cost, the connection details for the structure are generated automatically using accurate site-specific cost information obtained directly from fabricators and erectors. Member continuity rules are also applied to each node in the structure to improve constructability. The proposed optimization method is benchmarked against conventional weight-based ground structure optimization methods resulting in an average cost savings of up to 30% with comparable computational efficiency.

Keywords: cost-based structural optimization, cost-based topology and sizing, optimization, steel frame ground structure optimization, multidisciplinary optimization of steel structures

Procedia PDF Downloads 334
17121 The Formation of Thin Copper Films on Graphite Surface Using Magnetron Sputtering Method

Authors: Zydrunas Kavaliauskas, Aleksandras Iljinas, Liutauras Marcinauskas, Mindaugas Milieska, Vitas Valincius

Abstract:

The magnetron sputtering deposition method is often used to obtain thin film coatings. The main advantage of magnetron vaporization compared to other deposition methods is the high rate erosion of the cathode material (e.g., copper, aluminum, etc.) and the ability to operate under low-pressure conditions. The structure of the formed coatings depends on the working parameters of the magnetron deposition system, which is why it is possible to influence the properties of the growing film, such as morphology, crystal orientation, and dimensions, stresses, adhesion, etc. The properties of these coatings depend on the distance between the substrate and the magnetron surface, the vacuum depth, the gas used, etc. Using this deposition technology, substrates are most often placed near the anode. The magnetic trap of the magnetrons for localization of electrons in the cathode region is formed using a permanent magnet system that is on the side of the cathode. The scientific literature suggests that, after insertion of a small amount of copper into graphite, the electronic conductivity of graphite increase. The aim of this work is to create thin (up to 300 nm) layers on a graphite surface using a magnetron evaporation method, to investigate the formation peculiarities and microstructure of thin films, as well as the mechanism of copper diffusion into graphite inner layers at different thermal treatment temperatures. The electron scanning microscope was used to investigate the microrelief of the coating surface. The chemical composition is determined using the EDS method, which shows that, with an increase of the thermal treatment of the copper-carbon layer from 200 °C to 400 °C, the copper content is reduced from 8 to 4 % in atomic mass units. This is because the EDS method captures only the amount of copper on the graphite surface, while the temperature of the heat treatment increases part of the copper because of the diffusion processes penetrates into the inner layers of the graphite. The XRD method shows that the crystalline copper structure is not affected by thermal treatment.

Keywords: carbon, coatings, copper, magnetron sputtering

Procedia PDF Downloads 281
17120 Estimation of Slab Depth, Column Size and Rebar Location of Concrete Specimen Using Impact Echo Method

Authors: Y. T. Lee, J. H. Na, S. H. Kim, S. U. Hong

Abstract:

In this study, an experimental research for estimation of slab depth, column size and location of rebar of concrete specimen is conducted using the Impact Echo Method (IE) based on stress wave among non-destructive test methods. Estimation of slab depth had total length of 1800×300 and 6 different depths including 150 mm, 180 mm, 210 mm, 240 mm, 270 mm and 300 mm. The concrete column specimen was manufactured by differentiating the size into 300×300×300 mm, 400×400×400 mm and 500×500×500 mm. In case of the specimen for estimation of rebar, rebar of ∅22 mm was used in a specimen of 300×370×200 and arranged at 130 mm and 150 mm from the top to the rebar top. As a result of error rate of slab depth was overall mean of 3.1%. Error rate of column size was overall mean of 1.7%. Mean error rate of rebar location was 1.72% for top, 1.19% for bottom and 1.5% for overall mean showing relative accuracy.

Keywords: impact echo method, estimation, slab depth, column size, rebar location, concrete

Procedia PDF Downloads 341
17119 Using Priority Order of Basic Features for Circumscribed Masses Detection in Mammograms

Authors: Minh Dong Le, Viet Dung Nguyen, Do Huu Viet, Nguyen Huu Tu

Abstract:

In this paper, we present a new method for circumscribed masses detection in mammograms. Our method is evaluated on 23 mammographic images of circumscribed masses and 20 normal mammograms from public Mini-MIAS database. The method is quite sanguine with sensitivity (SE) of 95% with only about 1 false positive per image (FPpI). To achieve above results we carry out a progression following: Firstly, the input images are preprocessed with the aim to enhance key information of circumscribed masses; Next, we calculate and evaluate statistically basic features of abnormal regions on training database; Then, mammograms on testing database are divided into equal blocks which calculated corresponding features. Finally, using priority order of basic features to classify blocks as an abnormal or normal regions.

Keywords: mammograms, circumscribed masses, evaluated statistically, priority order of basic features

Procedia PDF Downloads 326
17118 Optimizing Human Diet Problem Using Linear Programming Approach: A Case Study

Authors: P. Priyanka, S. Shruthi, N. Guruprasad

Abstract:

Health is a common theme in most cultures. In fact all communities have their concepts of health, as part of their culture. Health continues to be a neglected entity. Planning of Human diet should be done very careful by selecting the food items or groups of food items also the composition involved. Low price and good taste of foods are regarded as two major factors for optimal human nutrition. Linear programming techniques have been extensively used for human diet formulation for quiet good number of years. Through the process, we mainly apply “The Simplex Method” which is a very useful statistical tool based on the theorem of Elementary Row Operation from Linear Algebra and also incorporate some other necessary rules set by the Simplex Method to help solve the problem. The study done by us is an attempt to develop a programming model for optimal planning and best use of nutrient ingredients.

Keywords: diet formulation, linear programming, nutrient ingredients, optimization, simplex method

Procedia PDF Downloads 552
17117 Laser Keratoplasty in Human Eye Considering the Fluid Aqueous Humor and Vitreous Humor Fluid Flow

Authors: Dara Singh, Keikhosrow Firouzbakhsh, Mohammad Taghi Ahmadian

Abstract:

In this paper, conventional laser Keratoplasty surgeries in the human eye are studied. For this purpose, a validated 3D finite volume model of the human eye is introduced. In this model the fluid flow has also been considered. The discretized domain of the human eye incorporates a bio-heat transfer equation coupled with a Boussinesq equation. Both continuous and pulsed lasers have been modeled and the results are compared. Moreover, two different conventional surgical positions that are upright and recumbent are compared for these laser therapies. The simulation results show that in these conventional surgeries, the temperature rises above the critical values at the laser insertion areas. However, due to the short duration and the localized nature, the potential damages are restricted to very small regions and can be ignored. The conclusion is that the present day lasers are acceptably safe to the human eye.

Keywords: eye, heat-transfer, keratoplasty laser, surgery

Procedia PDF Downloads 264
17116 A Robust and Efficient Segmentation Method Applied for Cardiac Left Ventricle with Abnormal Shapes

Authors: Peifei Zhu, Zisheng Li, Yasuki Kakishita, Mayumi Suzuki, Tomoaki Chono

Abstract:

Segmentation of left ventricle (LV) from cardiac ultrasound images provides a quantitative functional analysis of the heart to diagnose disease. Active Shape Model (ASM) is a widely used approach for LV segmentation but suffers from the drawback that initialization of the shape model is not sufficiently close to the target, especially when dealing with abnormal shapes in disease. In this work, a two-step framework is proposed to improve the accuracy and speed of the model-based segmentation. Firstly, a robust and efficient detector based on Hough forest is proposed to localize cardiac feature points, and such points are used to predict the initial fitting of the LV shape model. Secondly, to achieve more accurate and detailed segmentation, ASM is applied to further fit the LV shape model to the cardiac ultrasound image. The performance of the proposed method is evaluated on a dataset of 800 cardiac ultrasound images that are mostly of abnormal shapes. The proposed method is compared to several combinations of ASM and existing initialization methods. The experiment results demonstrate that the accuracy of feature point detection for initialization was improved by 40% compared to the existing methods. Moreover, the proposed method significantly reduces the number of necessary ASM fitting loops, thus speeding up the whole segmentation process. Therefore, the proposed method is able to achieve more accurate and efficient segmentation results and is applicable to unusual shapes of heart with cardiac diseases, such as left atrial enlargement.

Keywords: hough forest, active shape model, segmentation, cardiac left ventricle

Procedia PDF Downloads 333
17115 A Study of Flow near the Leading Edge of a Flat Plate by New Idea in Analytical Methods

Authors: M. R. Akbari, S. Akbari, L. Abdollahpour

Abstract:

The present paper is concerned with calculating the 2-dimensional velocity profile of a viscous flow for an incompressible fluid along the leading edge of a flat plate by using the continuity and motion equations with a simple and innovative approach. A Comparison between Numerical method and AGM has been made and the results have been revealed that AGM is very accurate and easy and can be applied for a wide variety of nonlinear problems. It is notable that most of the differential equations can be solved in this approach which in the other approaches they do not have this capability. Moreover, there are some valuable benefits in this method of solving differential equations, for instance: Without any dimensionless procedure, we can solve many differential equation(s), that is, differential equations are directly solvable by this method. In addition, it is not necessary to convert variables into new ones. According to the afore-mentioned expressions which will be proved in this literature, the process of solving nonlinear differential equation(s) will be very simple and convenient in contrast to the other approaches.

Keywords: leading edge, new idea, flat plate, incompressible fluid

Procedia PDF Downloads 277
17114 SiC Particulate-Reinforced SiC Composites Fabricated by PIP Method Using Highly Concentrated SiC Slurry

Authors: Jian Gu, Sea-Hoon Lee, Jun-Seop Kim

Abstract:

SiC particulate-reinforced SiC ceramic composites (SiCp/SiC) were successfully fabricated using polymer impregnation and pyrolysis (PIP) method. The effects of green density, infiltrated method, pyrolytic temperature, and heating rate on the densification behavior of the composites were investigated. SiCp/SiC particulate reinforced composites with high relative density up to 88.06% were fabricated after 4 PIP cycles using SiC pellets with high green density. The pellets were prepared by drying 62-70 vol.% aqueous SiC slurries, and the maximum relative density of the pellets was 75.5%. The hardness of the as-fabricated SiCp/SiCs was 21.05 GPa after 4 PIP cycles, which value increased to 23.99 GPa after a heat treatment at 2000℃. Excellent mechanical properties, thermal stability, and short processing time render the SiCp/SiC composite as a challenging candidate for the high-temperature application.

Keywords: high green density, mechanical property, polymer impregnation and pyrolysis, structural application

Procedia PDF Downloads 131
17113 Date Palm Compreg: A High Quality Bio-Composite of Date Palm Wood

Authors: Mojtaba Soltani, Edi Suhaimi Bakar, Hamid Reza Naji

Abstract:

Date Palm Wood (D.P.W) specimens were impregnated with Phenol formaldehyde (PF) resin at 15% level, using vacuum/pressure method. Three levels of moisture content (MC) (50%, 60%, and 70% ) before pressing stage and three hot pressing times (15, 20, and 30 minutes) were the variables. The boards were prepared at 20% compression rate. The physical properties of specimens such as spring back, thickness swelling and water absorption, and mechanical properties including MOR, MOE were studied and compared between variables. The results indicated that the percentage of MC levels before compression set was the main factor on the properties of the Date Palm Compreg. Also, the results showed that this compregnation method can be used as a good method for making high-quality bio-composite from Date Palm Wood.

Keywords: Date palm, phenol formaldehyde resin, high-quality bio-composite, physical and mechanical properties

Procedia PDF Downloads 339
17112 Reactive Dyed Superhydrophobic Cotton Fabric Production by Sol-Gel Method

Authors: Kuddis Büyükakıllı

Abstract:

The pretreated and bleached mercerized cotton fabric was dyed with reactive Everzol Brilliant Yellow 4GR (C.I. Yellow 160) dyestuff. Superhydrophobicity is provided to white and reactive dyed fabrics by using a nanotechnological sol-gel method with tetraethoxysilane and fluorcarbon water repellent agents by the two-step method. The effect of coating on color yield, fastness and functional properties of fabric was investigated. It was observed that water drop contact angles were higher in colorless coated fabrics compared to colored coated fabrics, there was no significant color change in colored superhydrophobic fabric and high color fastness values. Although there are no significant color losses in the fabrics after multiple washing and dry cleaning processes, water drop contact angles are greatly reduced.

Keywords: fluorcarbon water repellent agent, colored cotton fabric, sol-gel, superhydrophobic

Procedia PDF Downloads 109
17111 Systematic Identification of Noncoding Cancer Driver Somatic Mutations

Authors: Zohar Manber, Ran Elkon

Abstract:

Accumulation of somatic mutations (SMs) in the genome is a major driving force of cancer development. Most SMs in the tumor's genome are functionally neutral; however, some cause damage to critical processes and provide the tumor with a selective growth advantage (termed cancer driver mutations). Current research on functional significance of SMs is mainly focused on finding alterations in protein coding sequences. However, the exome comprises only 3% of the human genome, and thus, SMs in the noncoding genome significantly outnumber those that map to protein-coding regions. Although our understanding of noncoding driver SMs is very rudimentary, it is likely that disruption of regulatory elements in the genome is an important, yet largely underexplored mechanism by which somatic mutations contribute to cancer development. The expression of most human genes is controlled by multiple enhancers, and therefore, it is conceivable that regulatory SMs are distributed across different enhancers of the same target gene. Yet, to date, most statistical searches for regulatory SMs have considered each regulatory element individually, which may reduce statistical power. The first challenge in considering the cumulative activity of all the enhancers of a gene as a single unit is to map enhancers to their target promoters. Such mapping defines for each gene its set of regulating enhancers (termed "set of regulatory elements" (SRE)). Considering multiple enhancers of each gene as one unit holds great promise for enhancing the identification of driver regulatory SMs. However, the success of this approach is greatly dependent on the availability of comprehensive and accurate enhancer-promoter (E-P) maps. To date, the discovery of driver regulatory SMs has been hindered by insufficient sample sizes and statistical analyses that often considered each regulatory element separately. In this study, we analyzed more than 2,500 whole-genome sequence (WGS) samples provided by The Cancer Genome Atlas (TCGA) and The International Cancer Genome Consortium (ICGC) in order to identify such driver regulatory SMs. Our analyses took into account the combinatorial aspect of gene regulation by considering all the enhancers that control the same target gene as one unit, based on E-P maps from three genomics resources. The identification of candidate driver noncoding SMs is based on their recurrence. We searched for SREs of genes that are "hotspots" for SMs (that is, they accumulate SMs at a significantly elevated rate). To test the statistical significance of recurrence of SMs within a gene's SRE, we used both global and local background mutation rates. Using this approach, we detected - in seven different cancer types - numerous "hotspots" for SMs. To support the functional significance of these recurrent noncoding SMs, we further examined their association with the expression level of their target gene (using gene expression data provided by the ICGC and TCGA for samples that were also analyzed by WGS).

Keywords: cancer genomics, enhancers, noncoding genome, regulatory elements

Procedia PDF Downloads 99
17110 Numerical Solutions of Fredholm Integral Equations by B-Spline Wavelet Method

Authors: Ritu Rani

Abstract:

In this paper, we apply minimalistically upheld linear semi-orthogonal B-spline wavelets, exceptionally developed for the limited interim to rough the obscure function present in the integral equations. Semi-orthogonal wavelets utilizing B-spline uniquely developed for the limited interim and these wavelets can be spoken to in a shut frame. This gives a minimized help. Semi-orthogonal wavelets frame the premise in the space L²(R). Utilizing this premise, an arbitrary function in L²(R) can be communicated as the wavelet arrangement. For the limited interim, the wavelet arrangement cannot be totally introduced by utilizing this premise. This is on the grounds that backings of some premise are truncated at the left or right end purposes of the interim. Subsequently, an uncommon premise must be brought into the wavelet development on the limited interim. These functions are alluded to as the limit scaling functions and limit wavelet functions. B-spline wavelet method has been connected to fathom linear and nonlinear integral equations and their systems. The above method diminishes the integral equations to systems of algebraic equations and afterward these systems can be illuminated by any standard numerical methods. Here, we have connected Newton's method with suitable starting speculation for solving these systems.

Keywords: semi-orthogonal, wavelet arrangement, integral equations, wavelet development

Procedia PDF Downloads 173
17109 Chromatography Study of Fundamental Properties of Medical Radioisotope Astatine-211

Authors: Evgeny E. Tereshatov

Abstract:

Astatine-211 is considered one of the most promising radionuclides for Targeted Alpha Therapy. In order to develop reliable procedures to label biomolecules and utilize efficient delivery vehicle principles, one should understand the main chemical characteristics of astatine. The short half-life of 211At (~7.2 h) and absence of any stable isotopes of this element are limiting factors towards studying the behavior of astatine. Our team has developed a procedure for rapid and efficient isolation of astatine from irradiated bismuth material in nitric acid media based on 3-octanone and 1-octanol extraction chromatography resins. This process has been automated and it takes 20 min from the beginning of the target dissolution to the At-211 fraction elution. Our next step is to consider commercially available chromatography resins and their applicability in astatine purification in the same media. Results obtained along with the corresponding sorption mechanisms will be discussed.

Keywords: astatine-211, chromatography, automation, mechanism, radiopharmaceuticals

Procedia PDF Downloads 86
17108 Evaluation of Newly Synthesized Steroid Derivatives Using In silico Molecular Descriptors and Chemometric Techniques

Authors: Milica Ž. Karadžić, Lidija R. Jevrić, Sanja Podunavac-Kuzmanović, Strahinja Z. Kovačević, Anamarija I. Mandić, Katarina Penov-Gaši, Andrea R. Nikolić, Aleksandar M. Oklješa

Abstract:

This study considered selection of the in silico molecular descriptors and the models for newly synthesized steroid derivatives description and their characterization using chemometric techniques. Multiple linear regression (MLR) models were established and gave the best molecular descriptors for quantitative structure-retention relationship (QSRR) modeling of the retention of the investigated molecules. MLR models were without multicollinearity among the selected molecular descriptors according to the variance inflation factor (VIF) values. Used molecular descriptors were ranked using generalized pair correlation method (GPCM). In this method, the significant difference between independent variables can be noticed regardless almost equal correlation between dependent variable. Generated MLR models were statistically and cross-validated and the best models were kept. Models were ranked using sum of ranking differences (SRD) method. According to this method, the most consistent QSRR model can be found and similarity or dissimilarity between the models could be noticed. In this study, SRD was performed using average values of experimentally observed data as a golden standard. Chemometric analysis was conducted in order to characterize newly synthesized steroid derivatives for further investigation regarding their potential biological activity and further synthesis. This article is based upon work from COST Action (CM1105), supported by COST (European Cooperation in Science and Technology).

Keywords: generalized pair correlation method, molecular descriptors, regression analysis, steroids, sum of ranking differences

Procedia PDF Downloads 342
17107 Design and Implementation of Low-code Model-building Methods

Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu

Abstract:

This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.

Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment

Procedia PDF Downloads 15
17106 Identification of Nonlinear Systems Using Radial Basis Function Neural Network

Authors: C. Pislaru, A. Shebani

Abstract:

This paper uses the radial basis function neural network (RBFNN) for system identification of nonlinear systems. Five nonlinear systems are used to examine the activity of RBFNN in system modeling of nonlinear systems; the five nonlinear systems are dual tank system, single tank system, DC motor system, and two academic models. The feed forward method is considered in this work for modelling the non-linear dynamic models, where the K-Means clustering algorithm used in this paper to select the centers of radial basis function network, because it is reliable, offers fast convergence and can handle large data sets. The least mean square method is used to adjust the weights to the output layer, and Euclidean distance method used to measure the width of the Gaussian function.

Keywords: system identification, nonlinear systems, neural networks, radial basis function, K-means clustering algorithm

Procedia PDF Downloads 463
17105 Fault Diagnosis in Induction Motors Using Discrete Wavelet Transform

Authors: K. Yahia, A. Titaouine, A. Ghoggal, S. E. Zouzou, F. Benchabane

Abstract:

This paper deals with the problem of stator faults diagnosis in induction motors. Using the discrete wavelet transform (DWT) for the current Park’s vector modulus (CPVM) analysis, the inter-turn short-circuit faults diagnosis can be achieved. This method is based on the decomposition of the CPVM signal, where wavelet approximation and detail coefficients of this signal have been extracted. The energy evaluation of a known bandwidth detail permits to define a fault severity factor (FSF). This method has been tested through the simulation of an induction motor using a mathematical model based on the winding-function approach. Simulation, as well as experimental, results show the effectiveness of the used method.

Keywords: Induction Motors (IMs), inter-turn short-circuits diagnosis, Discrete Wavelet Transform (DWT), Current Park’s Vector Modulus (CPVM)

Procedia PDF Downloads 545
17104 A Practical Protection Method for Parallel Transmission-Lines Based on the Fault Travelling-Waves

Authors: Mohammad Reza Ebrahimi

Abstract:

In new restructured power systems, swift fault detection is very important. The parallel transmission-lines are vastly used in this kind of power systems because of high amount of energy transferring. In this paper, a method based on the comparison of two schemes, i.e., i) maximum magnitude of travelling-wave (TW) energy ii) the instants of maximum energy occurrence at the circuits of parallel transmission-line is proposed. Using the travelling-wave of fault in order to faulted line identification this method has noticeable operation time. Moreover, the algorithm can cover for identification of faults as external or internal faults. For an internal fault, the exact location of the fault can be estimated confidently. A lot of simulations have been done with PSCAD/EMTDC to verify the performance of the proposed algorithm.

Keywords: travelling-wave, maximum energy, parallel transmission-line, fault location

Procedia PDF Downloads 178
17103 Identification of Spam Keywords Using Hierarchical Category in C2C E-Commerce

Authors: Shao Bo Cheng, Yong-Jin Han, Se Young Park, Seong-Bae Park

Abstract:

Consumer-to-Consumer (C2C) E-commerce has been growing at a very high speed in recent years. Since identical or nearly-same kinds of products compete one another by relying on keyword search in C2C E-commerce, some sellers describe their products with spam keywords that are popular but are not related to their products. Though such products get more chances to be retrieved and selected by consumers than those without spam keywords, the spam keywords mislead the consumers and waste their time. This problem has been reported in many commercial services like e-bay and taobao, but there have been little research to solve this problem. As a solution to this problem, this paper proposes a method to classify whether keywords of a product are spam or not. The proposed method assumes that a keyword for a given product is more reliable if the keyword is observed commonly in specifications of products which are the same or the same kind as the given product. This is because that a hierarchical category of a product in general determined precisely by a seller of the product and so is the specification of the product. Since higher layers of the hierarchical category represent more general kinds of products, a reliable degree is differently determined according to the layers. Hence, reliable degrees from different layers of a hierarchical category become features for keywords and they are used together with features only from specifications for classification of the keywords. Support Vector Machines are adopted as a basic classifier using the features, since it is powerful, and widely used in many classification tasks. In the experiments, the proposed method is evaluated with a golden standard dataset from Yi-han-wang, a Chinese C2C e-commerce, and is compared with a baseline method that does not consider the hierarchical category. The experimental results show that the proposed method outperforms the baseline in F1-measure, which proves that spam keywords are effectively identified by a hierarchical category in C2C e-commerce.

Keywords: spam keyword, e-commerce, keyword features, spam filtering

Procedia PDF Downloads 291
17102 Impact Assessment of Lean Practices on Social Sustainability Indicators: An Approach Using ISM Method

Authors: Aline F. Marcon, Eduardo F. da Silva, Marina Bouzon

Abstract:

The impact of lean management on environmental sustainability is the research line that receives the most attention from academicians. Therefore, the social dimension of sustainable development has so far received less attention. This paper aims to evaluate the impact of intra-plant lean manufacturing practices on social sustainability indicators extracted from the Global Reporting Initiative (GRI) parameters. The method is two-phased, including MCDM approach to uncover the most relevant practices regarding social performance and Interpretive Structural Modeling (ISM) method to reveal the structural relationship among lean practices. Professionals from the academic and industrial fields answered the questionnaires. From the results of this paper, it is possible to verify that practices such as “Safety Improvement Programs”, “Total Quality Management” and “Cross-functional Workforce” are the ones which have the most positive influence on the set of GRI social indicators.

Keywords: indicators, ISM, lean, social, sustainability

Procedia PDF Downloads 139
17101 Calcium Silicate Bricks – Ultrasonic Pulse Method: Effects of Natural Frequency of Transducers on Measurement Results

Authors: Jiri Brozovsky

Abstract:

Modulus of elasticity is one of the important parameters of construction materials, which considerably influence their deformation properties and which can also be determined by means of non-destructive test methods like ultrasonic pulse method. However, measurement results of ultrasonic pulse methods are influenced by various factors, one of which is the natural frequency of the transducers. The paper states knowledge about influence of natural frequency of the transducers (54; 82 and 150kHz) on ultrasonic pulse velocity and dynamic modulus of elasticity (Young's Dynamic modulus of elasticity). Differences between ultrasonic pulse velocity and dynamic modulus of elasticity were found with the same smallest dimension of test specimen in the direction of sounding and density their value decreases as the natural frequency of transducers grew.

Keywords: calcium silicate brick, ultrasonic pulse method, ultrasonic pulse velocity, dynamic modulus of elasticity

Procedia PDF Downloads 411
17100 Examples of Parameterization of Stabilizing Controllers with One-Side Coprime Factorization

Authors: Kazuyoshi Mori

Abstract:

Examples of parameterization of stabilizing controllers that require only one of right-/left-coprime factorizations are presented. One parameterization method requires one side coprime factorization. The other requires no coprime factorization. The methods are based on the factorization approach so that a number of models can be applied the method we use in this paper.

Keywords: parametrization, coprime factorization, factorization approach, linear systems

Procedia PDF Downloads 367
17099 After-Cooling Analysis of RC Structural Members Exposed to High Temperature by Using Numerical Approach

Authors: Ju-Young Hwang, Hyo-Gyoung Kwak

Abstract:

This paper introduces a numerical analysis method for reinforced-concrete (RC) structures exposed to fire and compares the result with experimental results. The proposed analysis method for RC structure under the high temperature consists of two procedures. First step is to decide the temperature distribution across the section through the heat transfer analysis by using the time-temperature curve. After determination of the temperature distribution, the nonlinear analysis is followed. By considering material and geometrical nonlinearity with the temperature distribution, nonlinear analysis predicts the behavior of RC structure under the fire by the exposed time. The proposed method is validated by the comparison with the experimental results. Finally, prediction model to describe the status of after-cooling concrete can also be introduced based on the results of additional experiment. The product of this study is expected to be embedded for smart structure monitoring system against fire in u-City.

Keywords: RC, high temperature, after-cooling analysis, nonlinear analysis

Procedia PDF Downloads 406
17098 Task Scheduling and Resource Allocation in Cloud-based on AHP Method

Authors: Zahra Ahmadi, Fazlollah Adibnia

Abstract:

Scheduling of tasks and the optimal allocation of resources in the cloud are based on the dynamic nature of tasks and the heterogeneity of resources. Applications that are based on the scientific workflow are among the most widely used applications in this field, which are characterized by high processing power and storage capacity. In order to increase their efficiency, it is necessary to plan the tasks properly and select the best virtual machine in the cloud. The goals of the system are effective factors in scheduling tasks and resource selection, which depend on various criteria such as time, cost, current workload and processing power. Multi-criteria decision-making methods are a good choice in this field. In this research, a new method of work planning and resource allocation in a heterogeneous environment based on the modified AHP algorithm is proposed. In this method, the scheduling of input tasks is based on two criteria of execution time and size. Resource allocation is also a combination of the AHP algorithm and the first-input method of the first client. Resource prioritization is done with the criteria of main memory size, processor speed and bandwidth. What is considered in this system to modify the AHP algorithm Linear Max-Min and Linear Max normalization methods are the best choice for the mentioned algorithm, which have a great impact on the ranking. The simulation results show a decrease in the average response time, return time and execution time of input tasks in the proposed method compared to similar methods (basic methods).

Keywords: hierarchical analytical process, work prioritization, normalization, heterogeneous resource allocation, scientific workflow

Procedia PDF Downloads 138
17097 Quantum Algebra from Generalized Q-Algebra

Authors: Muna Tabuni

Abstract:

The paper contains an investigation of the notion of Q algebras. A brief introduction to quantum mechanics is given, in that systems the state defined by a vector in a complex vector space H which have Hermitian inner product property. H may be finite or infinite-dimensional. In quantum mechanics, operators must be hermitian. These facts are saved by Lie algebra operators but not by those of quantum algebras. A Hilbert space H consists of a set of vectors and a set of scalars. Lie group is a differentiable topological space with group laws given by differentiable maps. A Lie algebra has been introduced. Q-algebra has been defined. A brief introduction to BCI-algebra is given. A BCI sub algebra is introduced. A brief introduction to BCK=BCH-algebra is given. Every BCI-algebra is a BCH-algebra. Homomorphism maps meanings are introduced. Homomorphism maps between two BCK algebras are defined. The mathematical formulations of quantum mechanics can be expressed using the theory of unitary group representations. A generalization of Q algebras has been introduced, and their properties have been considered. The Q- quantum algebra has been studied, and various examples have been given.

Keywords: Q-algebras, BCI, BCK, BCH-algebra, quantum mechanics

Procedia PDF Downloads 192
17096 A Tuning Method for Microwave Filter via Complex Neural Network and Improved Space Mapping

Authors: Shengbiao Wu, Weihua Cao, Min Wu, Can Liu

Abstract:

This paper presents an intelligent tuning method of microwave filter based on complex neural network and improved space mapping. The tuning process consists of two stages: the initial tuning and the fine tuning. At the beginning of the tuning, the return loss of the filter is transferred to the passband via the error of phase. During the fine tuning, the phase shift caused by the transmission line and the higher order mode is removed by the curve fitting. Then, an Cauchy method based on the admittance parameter (Y-parameter) is used to extract the coupling matrix. The influence of the resonant cavity loss is eliminated during the parameter extraction process. By using processed data pairs (the amount of screw variation and the variation of the coupling matrix), a tuning model is established by the complex neural network. In view of the improved space mapping algorithm, the mapping relationship between the actual model and the ideal model is established, and the amplitude and direction of the tuning is constantly updated. Finally, the tuning experiment of the eight order coaxial cavity filter shows that the proposed method has a good effect in tuning time and tuning precision.

Keywords: microwave filter, scattering parameter, coupling matrix, intelligent tuning

Procedia PDF Downloads 293
17095 Application of Geosynthetics for the Recovery of Located Road on Geological Failure

Authors: Rideci Farias, Haroldo Paranhos

Abstract:

The present work deals with the use of drainage geo-composite as a deep drainage and geogrid element to reinforce the base of the body of the landfill destined to the road pavement on geological faults in the stretch of the TO-342 Highway, between the cities of Miracema and Miranorte, in the State of Tocantins / TO, Brazil, which for many years was the main link between TO-010 and BR-153, after the city of Palmas, also in the state of Tocantins / TO, Brazil. For this application, geotechnical and geological studies were carried out by means of SPT percussion drilling, drilling and rotary drilling, to understand the problem, identifying the type of faults, filling material and the definition of the water table. According to the geological and geotechnical studies carried out, the area where the route was defined, passes through a zone of longitudinal fault to the runway, with strong breaking / fracturing, with presence of voids, intense alteration and with advanced argilization of the rock and with the filling up parts of the faults by organic and compressible soils leachate from other horizons. This geology presents as a geotechnical aggravating agent a medium of high hydraulic load and very low resistance to penetration. For more than 20 years, the region presented constant excessive deformations in the upper layers of the pavement, which after routine services of regularization, reconformation, re-compaction of the layers and application of the asphalt coating. The faults were quickly propagated to the surface of the asphalt pavement, generating a longitudinal shear, forming steps (unevenness), close to 40 cm, causing numerous accidents and discomfort to the drivers, since the geometric positioning was in a horizontal curve. Several projects were presented to the region's highway department to solve the problem. Due to the need for partial closure of the runway, the short time for execution, the use of geosynthetics was proposed and the most adequate solution for the problem was taken into account the movement of existing geological faults and the position of the water level in relation to several Layers of pavement and failure. In order to avoid any flow of water in the body of the landfill and in the filling material of the faults, a drainage curtain solution was used, carried out at 4.0 meters depth, with drainage geo-composite and as reinforcement element and inhibitor of the possible A geogrid of 200 kN / m of resistance was inserted at the base of the reconstituted landfill. Recent evaluations, after 13 years of application of the solution, show the efficiency of the technique used, supported by the geotechnical studies carried out in the area.

Keywords: geosynthetics, geocomposite, geogrid, road, recovery, geological failure

Procedia PDF Downloads 163
17094 High-pressure Crystallographic Characterization of f-block Element Complexes

Authors: Nicholas B. Beck, Thomas E. Albrecht-Schönzart

Abstract:

High-pressure results in decreases in the bond lengths of metal-ligand bonds, which has proven to be incredibly informative in uncovering differences in bonding between lanthanide and actinide complexes. The degree of f-electron contribution to the metal ligand bonds has been observed to increase under pressure by a far greater degree in the actinides than the lanthanides, as revealed by spectroscopic studies. However, the actual changes in bond lengths have yet to be quantified, although computationally predicted. By using high-pressure crystallographic techniques, crystal structures of lanthanide complexes have been obtained at pressures up to 5 GPa for both hard and soft-donor ligands. These studies have revealed some unpredicted changes in the coordination environment as well as provided experimental support to computational results

Keywords: crystallography, high-pressure, lanthanide, materials

Procedia PDF Downloads 95