Search results for: Large Volume Swiss-Cheese Calabi-Yau's
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32817

Search results for: Large Volume Swiss-Cheese Calabi-Yau's

32637 Scatter Analysis of Fatigue Life and Pore Size Data of Die-Cast AM60B Magnesium Alloy

Authors: S. Mohd, Y. Mutoh, Y. Otsuka, Y. Miyashita, T. Koike, T. Suzuki

Abstract:

Scatter behavior of fatigue life in die-cast AM60B alloy was investigated. For comparison, those in rolled AM60B alloy and die-cast A365-T5 aluminum alloy were also studied. Scatter behavior of pore size was also investigated to discuss dominant factors for fatigue life scatter in die-cast materials. Three-parameter Weibull function was suitable to explain the scatter behavior of both fatigue life and pore size. The scatter of fatigue life in die-cast AM60B alloy was almost comparable to that in die-cast A365-T5 alloy, while it was significantly large compared to that in the rolled AM60B alloy. Scatter behavior of pore size observed at fracture nucleation site on the fracture surface was comparable to that observed on the specimen cross-section and also to that of fatigue life. Therefore, the dominant factor for large scatter of fatigue life in die-cast alloys would be the large scatter of pore size. This speculation was confirmed by the fracture mechanics fatigue life prediction, where the pore observed at fatigue crack nucleation site was assumed as the pre-existing crack.

Keywords: Fatigue life, Pore size, Scatter, Weibull distribution, Die-cast magnesium alloy

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2338
32636 Parallelization of Ensemble Kalman Filter (EnKF) for Oil Reservoirs with Time-lapse Seismic Data

Authors: Md Khairullah, Hai-Xiang Lin, Remus G. Hanea, Arnold W. Heemink

Abstract:

In this paper we describe the design and implementation of a parallel algorithm for data assimilation with ensemble Kalman filter (EnKF) for oil reservoir history matching problem. The use of large number of observations from time-lapse seismic leads to a large turnaround time for the analysis step, in addition to the time consuming simulations of the realizations. For efficient parallelization it is important to consider parallel computation at the analysis step. Our experiments show that parallelization of the analysis step in addition to the forecast step has good scalability, exploiting the same set of resources with some additional efforts.

Keywords: EnKF, Data assimilation, Parallel computing, Parallel efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2227
32635 Application of Model Free Adaptive Control in Main Steam Temperature System of Thermal Power Plant

Authors: Khaing Yadana Swe, Lillie Dewan

Abstract:

At present, the cascade PID control is widely used to control the superheating temperature (main steam temperature). As Main Steam Temperature has the characteristics of large inertia, large time-delay and time varying, etc., conventional PID control strategy cannot achieve good control performance. In order to overcome the bad performance and deficiencies of main steam temperature control system, Model Free Adaptive Control (MFAC) - P cascade control system is proposed in this paper. By substituting MFAC in PID of the main control loop of the main steam temperature control, it can overcome time delays, non-linearity, disturbance and time variation.

Keywords: Model free Adaptive Control, Cascade Control, Adaptive Control, PID.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2749
32634 Production Planning and Scheduling and SME

Authors: M. Heck, H. Vettiger

Abstract:

Small and medium-sized enterprises (SME) are the backbone of central Europe’s economies and have a significant contribution to the gross domestic product. Production planning and scheduling (PPS) is still a crucial element in manufacturing industries of the 21st century even though this area of research is more than a century old. The topic of PPS is well researched especially in the context of large enterprises in the manufacturing industry. However the implementation of PPS methodologies within SME is mostly unobserved. This work analyzes how PPS is implemented in SME with the geographical focus on Switzerland and its vicinity. Based on restricted resources compared to large enterprises, SME have to face different challenges. The real problem areas of selected enterprises in regards of PPS are identified and evaluated. For the identified real-life problem areas of SME clear and detailed recommendations are created, covering concepts and best practices and the efficient usage of PPS. Furthermore the economic and entrepreneurial value for companies is lined out and why the implementation of the introduced recommendations is advised.

Keywords: Central Europe, PPS, Production Planning, SME.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2120
32633 An Intelligent Approach of Rough Set in Knowledge Discovery Databases

Authors: Hrudaya Ku. Tripathy, B. K. Tripathy, Pradip K. Das

Abstract:

Knowledge Discovery in Databases (KDD) has evolved into an important and active area of research because of theoretical challenges and practical applications associated with the problem of discovering (or extracting) interesting and previously unknown knowledge from very large real-world databases. Rough Set Theory (RST) is a mathematical formalism for representing uncertainty that can be considered an extension of the classical set theory. It has been used in many different research areas, including those related to inductive machine learning and reduction of knowledge in knowledge-based systems. One important concept related to RST is that of a rough relation. In this paper we presented the current status of research on applying rough set theory to KDD, which will be helpful for handle the characteristics of real-world databases. The main aim is to show how rough set and rough set analysis can be effectively used to extract knowledge from large databases.

Keywords: Data mining, Data tables, Knowledge discovery in database (KDD), Rough sets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2290
32632 Improving Convergence of Parameter Tuning Process of the Additive Fuzzy System by New Learning Strategy

Authors: Thi Nguyen, Lee Gordon-Brown, Jim Peterson, Peter Wheeler

Abstract:

An additive fuzzy system comprising m rules with n inputs and p outputs in each rule has at least t m(2n + 2 p + 1) parameters needing to be tuned. The system consists of a large number of if-then fuzzy rules and takes a long time to tune its parameters especially in the case of a large amount of training data samples. In this paper, a new learning strategy is investigated to cope with this obstacle. Parameters that tend toward constant values at the learning process are initially fixed and they are not tuned till the end of the learning time. Experiments based on applications of the additive fuzzy system in function approximation demonstrate that the proposed approach reduces the learning time and hence improves convergence speed considerably.

Keywords: Additive fuzzy system, improving convergence, parameter learning process, unsupervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1473
32631 Advantages of Large Strands in Precast/Prestressed Concrete Highway Application

Authors: Amin Akhnoukh

Abstract:

The objective of this research is to investigate the advantages of using large-diameter 0.7 inch prestressing strands in pretention applications. The advantages of large-diameter strands are mainly beneficial in the heavy construction applications. Bridges and tunnels are subjected to a higher daily traffic with an exponential increase in trucks ultimate weight, which raise the demand for higher structural capacity of bridges and tunnels. In this research, precast prestressed I-girders were considered as a case study. Flexure capacities of girders fabricated using 0.7 inch strands and different concrete strengths were calculated and compared to capacities of 0.6 inch strands girders fabricated using equivalent concrete strength. The effect of bridge deck concrete strength on composite deck-girder section capacity was investigated due to its possible effect on final section capacity. Finally, a comparison was made to compare the bridge cross-section of girders designed using regular 0.6 inch strands and the large-diameter 0.7 inch. The research findings showed that structural advantages of 0.7 inch strands allow for using fewer bridge girders, reduced material quantity, and light-weight members. The structural advantages of 0.7 inch strands are maximized when high strength concrete (HSC) are used in girder fabrication, and concrete of minimum 5ksi compressive strength is used in pouring bridge decks. The use of 0.7 inch strands in bridge industry can partially contribute to the improvement of bridge conditions, minimize construction cost, and reduce the construction duration of the project.

Keywords: 0.7 Inch Strands, I-Girders, Pretension, Flexure Capacity

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2696
32630 Hydrodynamic Simulation of Fixed Bed GTL Reactor Using CFD

Authors: Sh. Shahhosseini, S. Alinia, M. Irani

Abstract:

In this work, axisymetric CFD simulation of fixed bed GTL reactor has been conducted, using computational fluid dynamics (CFD). In fixed bed CFD modeling, when N (tube-to-particle diameter ratio) has a large value, it is common to consider the packed bed as a porous media. Synthesis gas (a mixture of predominantly carbon monoxide and hydrogen) was fed to the reactor. The reactor length was 20 cm, divided to three sections. The porous zone was in the middle section of the reactor. The model equations were solved employing finite volume method. The effects of particle diameter, bed voidage, fluid velocity and bed length on pressure drop have been investigated. Simulation results showed these parameters could have remarkable impacts on the reactor pressure drop.

Keywords: GTL Process, Fixed bed reactor, Pressure drop, CFDsimulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2331
32629 A Robust and Adaptive Unscented Kalman Filter for the Air Fine Alignment of the Strapdown Inertial Navigation System/GPS

Authors: Jian Shi, Baoguo Yu, Haonan Jia, Meng Liu, Ping Huang

Abstract:

Adapting to the flexibility of war, a large number of guided weapons launch from aircraft. Therefore, the inertial navigation system loaded in the weapon needs to undergo an alignment process in the air. This article proposes the following methods to the problem of inaccurate modeling of the system under large misalignment angles, the accuracy reduction of filtering caused by outliers, and the noise changes in GPS signals: first, considering the large misalignment errors of Strapdown Inertial Navigation System (SINS)/GPS, a more accurate model is made rather than to make a small-angle approximation, and the Unscented Kalman Filter (UKF) algorithms are used to estimate the state; then, taking into account the impact of GPS noise changes on the fine alignment algorithm, the innovation adaptive filtering algorithm is introduced to estimate the GPS’s noise in real-time; at the same time, in order to improve the anti-interference ability of the air fine alignment algorithm, a robust filtering algorithm based on outlier detection is combined with the air fine alignment algorithm to improve the robustness of the algorithm. The algorithm can improve the alignment accuracy and robustness under interference conditions, which is verified by simulation.

Keywords: Air alignment, fine alignment, inertial navigation system, integrated navigation system, UKF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 471
32628 Investigation of Some Technical Indexes inStock Forecasting Using Neural Networks

Authors: Myungsook Klassen

Abstract:

Training neural networks to capture an intrinsic property of a large volume of high dimensional data is a difficult task, as the training process is computationally expensive. Input attributes should be carefully selected to keep the dimensionality of input vectors relatively small. Technical indexes commonly used for stock market prediction using neural networks are investigated to determine its effectiveness as inputs. The feed forward neural network of Levenberg-Marquardt algorithm is applied to perform one step ahead forecasting of NASDAQ and Dow stock prices.

Keywords: Stock Market Prediction, Neural Networks, Levenberg-Marquadt Algorithm, Technical Indexes

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1903
32627 Design of the Large Dimension Cold Shield Cooled by G-M Cryocooler

Authors: Gong Jie, Yu Qianxu, Liu Min, Shan Weiwei

Abstract:

The design of methods of the 20 K large dimension cold shield used for infrared radiation demarcating in space environment simulation test were introduced in this paper. The cold shield were cooled by five G-M cryocoolers , and the dimension of the cold shield is the largest in our country.Cold shield installation and distribution and compensator for contraction on cooling were introduced detailedly. The temperature distribution and cool-down time of cold shield surface were also calculated and analysed in this paper. The design of cold shield resolves the difficulty of compensator for contraction on cooling successfully. Test results show that the actual technical performance indicators of cold shield met and exceeded the design requirements.

Keywords: cold shield, G-M cryocooler,infrared radiometer demarcating, satellite, space environment simulation equipments

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1643
32626 Research on the Protection and Reuse Model of Historical Buildings in Chinese Airports

Authors: Jie Ouyang, Chen Nie

Abstract:

China had constructed a large number of military and civilian airports before and after World War II, and then began large-scale repairs, reconstructions or relocation of airports after the baptism of wars after World War I and World War II. The airport's historical area and its historical buildings such as terminals, hangars, and towers have adopted different protection strategies and reuse application strategies. This paper is based on the judgment of the value of airport historical buildings to study different protection and reuse strategies. The protection and reuse models of historical buildings are classified in three dimensions: the airport historical area, the airport historical building complex and its individual buildings, and combined with specific examples to discuss and summarize the technical characteristics, protection strategies and successful experiences of different modes of protection and reuse of historical areas and historical buildings of airports.

Keywords: Airport, airport area, historic airport building, protection, reuse model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 599
32625 Reducing the Number of Constraints in Non Safe Petri Net

Authors: M. Zareiee, A. Dideban

Abstract:

This paper addresses the problem of forbidden states in non safe Petri Nets. In the system, for preventing it from entering the forbidden states, some linear constraints can be assigned to them. Then these constraints can be enforced on the system using control places. But when the number of constraints in the system is large, a large number of control places must be added to the model of system. This concept complicates the model of system. There are some methods for reducing the number of constraints in safe Petri Nets. But there is no a systematic method for non safe Petri Nets. In this paper we propose a method for reducing the number of constraints in non safe Petri Nets which is based on solving an integer linear programming problem.

Keywords: discrete event system, Supervisory control, Petri Net, Constraint

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1274
32624 Modeling and Control of Direct Driven PMSG for Ultra Large Wind Turbines

Authors: Ahmed M. Hemeida, Wael A. Farag, Osama A. Mahgoub

Abstract:

This paper focuses on developing an integrated reliable and sophisticated model for ultra large wind turbines And to study the performance and analysis of vector control on large wind turbines. With the advance of power electronics technology, direct driven multi-pole radial flux PMSG (Permanent Magnet Synchronous Generator) has proven to be a good choice for wind turbines manufacturers. To study the wind energy conversion systems, it is important to develop a wind turbine simulator that is able to produce realistic and validated conditions that occur in real ultra MW wind turbines. Three different packages are used to simulate this model, namely, Turbsim, FAST and Simulink. Turbsim is a Full field wind simulator developed by National Renewable Energy Laboratory (NREL). The wind turbine mechanical parts are modeled by FAST (Fatigue, Aerodynamics, Structures and Turbulence) code which is also developed by NREL. Simulink is used to model the PMSG, full scale back to back IGBT converters, and the grid.

Keywords: FAST, Permanent Magnet Synchronous Generator(PMSG), TurbSim, Vector Control and Pitch Control

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5545
32623 A Software Tool Design for Cerebral Infarction of MR Images

Authors: Kyoung-Jong Park, Woong-Gi Jeon, Hee-Cheol Kim, Dong-Eog Kim, Heung-Kook Choi

Abstract:

The brain MR imaging-based clinical research and analysis system were specifically built and the development for a large-scale data was targeted. We used the general clinical data available for building large-scale data. Registration period for the selection of the lesion ROI and the region growing algorithm was used and the Mesh-warp algorithm for matching was implemented. The accuracy of the matching errors was modified individually. Also, the large ROI research data can accumulate by our developed compression method. In this way, the correctly decision criteria to the research result was suggested. The experimental groups were age, sex, MR type, patient ID and smoking which can easily be queries. The result data was visualized of the overlapped images by a color table. Its data was calculated by the statistical package. The evaluation for the utilization of this system in the chronic ischemic damage in the area has done from patients with the acute cerebral infarction. This is the cause of neurologic disability index location in the center portion of the lateral ventricle facing. The corona radiate was found in the position. Finally, the system reliability was measured both inter-user and intra-user registering correlation.

Keywords: Software tool design, Cerebral infarction, Brain MR image, Registration

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1616
32622 Multidimensional Data Mining by Means of Randomly Travelling Hyper-Ellipsoids

Authors: Pavel Y. Tabakov, Kevin Duffy

Abstract:

The present study presents a new approach to automatic data clustering and classification problems in large and complex databases and, at the same time, derives specific types of explicit rules describing each cluster. The method works well in both sparse and dense multidimensional data spaces. The members of the data space can be of the same nature or represent different classes. A number of N-dimensional ellipsoids are used for enclosing the data clouds. Due to the geometry of an ellipsoid and its free rotation in space the detection of clusters becomes very efficient. The method is based on genetic algorithms that are used for the optimization of location, orientation and geometric characteristics of the hyper-ellipsoids. The proposed approach can serve as a basis for the development of general knowledge systems for discovering hidden knowledge and unexpected patterns and rules in various large databases.

Keywords: Classification, clustering, data minig, genetic algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1733
32621 Particle Size Effect on Shear Strength of Granular Materials in Direct Shear Test

Authors: R. Alias, A. Kasa, M. R. Taha

Abstract:

The effect of particle size on shear strength of granular materials are investigated using direct shear tests. Small direct shear test (60 mm by 60 mm by 24 mm deep) were conducted for particles passing the sieves with opening size of 2.36 mm. Meanwhile, particles passing the standard 20 mm sieves were tested using large direct shear test (300 mm by 300 mm by 200 mm deep). The large direct shear tests and the small direct shear tests carried out using the same shearing rate of 0.09 mm/min and similar normal stresses of 100, 200 and 300 kPa. The results show that the peak and residual shear strength increases as particle size increases.

Keywords: Particle size, shear strength, granular material, direct shear test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5187
32620 A Genetic Algorithm for Clustering on Image Data

Authors: Qin Ding, Jim Gasvoda

Abstract:

Clustering is the process of subdividing an input data set into a desired number of subgroups so that members of the same subgroup are similar and members of different subgroups have diverse properties. Many heuristic algorithms have been applied to the clustering problem, which is known to be NP Hard. Genetic algorithms have been used in a wide variety of fields to perform clustering, however, the technique normally has a long running time in terms of input set size. This paper proposes an efficient genetic algorithm for clustering on very large data sets, especially on image data sets. The genetic algorithm uses the most time efficient techniques along with preprocessing of the input data set. We test our algorithm on both artificial and real image data sets, both of which are of large size. The experimental results show that our algorithm outperforms the k-means algorithm in terms of running time as well as the quality of the clustering.

Keywords: Clustering, data mining, genetic algorithm, image data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2002
32619 Lightweight Mirrors for Space X-Ray Telescopes

Authors: M. Mika, L. Pina, M. Landova, L. Sveda, R. Havlikova, V. Marsikova, R. Hudec, A. Inneman

Abstract:

Future astronomical projects on large space x-ray imaging telescopes require novel substrates and technologies for the construction of their reflecting mirrors. The mirrors must be lightweight and precisely shaped to achieve large collecting area with high angular resolution. The new materials and technologies must be cost-effective. Currently, the most promising materials are glass or silicon foils. We focused on precise shaping these foils by thermal forming process. We studied free and forced slumping in the temperature region of hot plastic deformation and compared the shapes obtained by the different slumping processes. We measured the shapes and the surface quality of the foils. In the experiments, we varied both heat-treatment temperature and time following our experiment design. The obtained data and relations we can use for modeling and optimizing the thermal forming procedure.

Keywords: Glass, silicon, thermal forming, x-ray

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1340
32618 Active Intra-ONU Scheduling with Cooperative Prediction Mechanism in EPONs

Authors: Chuan-Ching Sue, Shi-Zhou Chen, Ting-Yu Huang

Abstract:

Dynamic bandwidth allocation in EPONs can be generally separated into inter-ONU scheduling and intra-ONU scheduling. In our previous work, the active intra-ONU scheduling (AS) utilizes multiple queue reports (QRs) in each report message to cooperate with the inter-ONU scheduling and makes the granted bandwidth fully utilized without leaving unused slot remainder (USR). This scheme successfully solves the USR problem originating from the inseparability of Ethernet frame. However, without proper setting of threshold value in AS, the number of QRs constrained by the IEEE 802.3ah standard is not enough, especially in the unbalanced traffic environment. This limitation may be solved by enlarging the threshold value. The large threshold implies the large gap between the adjacent QRs, thus resulting in the large difference between the best granted bandwidth and the real granted bandwidth. In this paper, we integrate AS with a cooperative prediction mechanism and distribute multiple QRs to reduce the penalty brought by the prediction error. Furthermore, to improve the QoS and save the usage of queue reports, the highest priority (EF) traffic which comes during the waiting time is granted automatically by OLT and is not considered in the requested bandwidth of ONU. The simulation results show that the proposed scheme has better performance metrics in terms of bandwidth utilization and average delay for different classes of packets.

Keywords: EPON, Inter-ONU and Intra-ONU scheduling, Prediction, Unused slot remainder

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1560
32617 MONARC: A Case Study on Simulation Analysis for LHC Activities

Authors: Ciprian Dobre

Abstract:

The scale, complexity and worldwide geographical spread of the LHC computing and data analysis problems are unprecedented in scientific research. The complexity of processing and accessing this data is increased substantially by the size and global span of the major experiments, combined with the limited wide area network bandwidth available. We present the latest generation of the MONARC (MOdels of Networked Analysis at Regional Centers) simulation framework, as a design and modeling tool for large scale distributed systems applied to HEP experiments. We present simulation experiments designed to evaluate the capabilities of the current real-world distributed infrastructure to support existing physics analysis processes and the means by which the experiments bands together to meet the technical challenges posed by the storage, access and computing requirements of LHC data analysis within the CMS experiment.

Keywords: Modeling and simulation, evaluation, large scale distributed systems, LHC experiments, CMS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1762
32616 Entropy Based Spatial Design: A Genetic Algorithm Approach (Case Study)

Authors: Abbas Siefi, Mohammad Javad Karimifar

Abstract:

We study the spatial design of experiment and we want to select a most informative subset, having prespecified size, from a set of correlated random variables. The problem arises in many applied domains, such as meteorology, environmental statistics, and statistical geology. In these applications, observations can be collected at different locations and possibly at different times. In spatial design, when the design region and the set of interest are discrete then the covariance matrix completely describe any objective function and our goal is to choose a feasible design that minimizes the resulting uncertainty. The problem is recast as that of maximizing the determinant of the covariance matrix of the chosen subset. This problem is NP-hard. For using these designs in computer experiments, in many cases, the design space is very large and it's not possible to calculate the exact optimal solution. Heuristic optimization methods can discover efficient experiment designs in situations where traditional designs cannot be applied, exchange methods are ineffective and exact solution not possible. We developed a GA algorithm to take advantage of the exploratory power of this algorithm. The successful application of this method is demonstrated in large design space. We consider a real case of design of experiment. In our problem, design space is very large and for solving the problem, we used proposed GA algorithm.

Keywords: Spatial design of experiments, maximum entropy sampling, computer experiments, genetic algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1601
32615 Strengthening of RC Beams with Large Openings in Shear by CFRP Laminates: 2D Nonlinear FE Analysis

Authors: S.C. Chin, N. Shafiq, M.F. Nuruddin

Abstract:

To date, theoretical studies concerning the Carbon Fiber Reinforced Polymer (CFRP) strengthening of RC beams with openings have been rather limited. In addition, various numerical analyses presented so far have effectively simulated the behaviour of solid beam strengthened by FRP material. In this paper, a two dimensional nonlinear finite element analysis is presented to validate against the laboratory test results of six RC beams. All beams had the same rectangular cross-section geometry and were loaded under four point bending. The crack pattern results of the finite element model show good agreement with the crack pattern of the experimental beams. The load midspan deflection curves of the finite element models exhibited a stiffer result compared to the experimental beams. The possible reason may be due to the perfect bond assumption used between the concrete and steel reinforcement.

Keywords: CFRP, large opening, RC beam, strengthening

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765
32614 The Location of Park and Ride Facilities Using the Fuzzy Inference Model

Authors: Anna Lower, Michal Lower, Robert Masztalski, Agnieszka Szumilas

Abstract:

The paper presents a method in which the expert knowledge is applied to fuzzy inference model. Even a less experienced person could benefit from the use of such a system, e.g. urban planners, officials. The analysis result is obtained in a very short time, so a large number of the proposed locations can also be verified in a short time. The proposed method is intended for testing of locations of car parks in a city. The paper shows selected examples of locations of the P&R facilities in cities planning to introduce the P&R. The analyses of existing objects are also shown in the paper and they are confronted with the opinions of the system users, with particular emphasis on unpopular locations. The results of the analyses are compared to expert analysis of the P&R facilities location that was outsourced by the city and the opinions about existing facilities users that were expressed on social networking sites. The obtained results are consistent with actual users’ feedback. The proposed method proves to be good, but does not require the involvement of a large experts team and large financial contributions for complicated research. The method also provides an opportunity to show the alternative location of P&R facilities. Although the results of the method are approximate, they are not worse than results of analysis of employed experts. The advantage of this method is ease of use, which simplifies the professional expert analysis. The ability of analyzing a large number of alternative locations gives a broader view on the problem. It is valuable that the arduous analysis of the team of people can be replaced by the model's calculation. According to the authors, the proposed method is also suitable for implementation on a GIS platform.

Keywords: Fuzzy logic inference, P&R facilities, P&R location.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614
32613 Identifying Corporate Managerial Topics with Web Pages

Authors: Juan Llopis, Reyes Gonzalez, Jose Gasco

Abstract:

This paper has as its main aim to analyse how corporate web pages can become an essential tool in order to detect strategic trends by firms or sectors, and even a primary source for benchmarking. This technique has made it possible to identify the key issues in the strategic management of the most excellent large Spanish firms and also to describe trends in their long-range planning, a way of working that can be generalised to any country or firm group. More precisely, two objectives were sought. The first one consisted in showing the way in which corporate websites make it possible to obtain direct information about the strategic variables which can define firms. This tool is dynamic (since web pages are constantly updated) as well as direct and reliable, since the information comes from the firm itself, not from comments of third parties (such as journalists, academicians, consultants...). When this information is analysed for a group of firms, one can observe their characteristics in terms of both managerial tasks and business management. As for the second objective, the methodology proposed served to describe the corporate profile of the large Spanish enterprises included in the Ibex35 (the Ibex35 or Iberia Index is the reference index in the Spanish Stock Exchange and gathers periodically the 35 most outstanding Spanish firms). An attempt is therefore made to define the long-range planning that would be characteristic of the largest Spanish firms.

Keywords: Web Pages, Strategic Management, Corporate Description, Large Firms, Spain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1537
32612 Learning Monte Carlo Data for Circuit Path Length

Authors: Namal A. Senanayake, A. Beg, Withana C. Prasad

Abstract:

This paper analyzes the patterns of the Monte Carlo data for a large number of variables and minterms, in order to characterize the circuit path length behavior. We propose models that are determined by training process of shortest path length derived from a wide range of binary decision diagram (BDD) simulations. The creation of the model was done use of feed forward neural network (NN) modeling methodology. Experimental results for ISCAS benchmark circuits show an RMS error of 0.102 for the shortest path length complexity estimation predicted by the NN model (NNM). Use of such a model can help reduce the time complexity of very large scale integrated (VLSI) circuitries and related computer-aided design (CAD) tools that use BDDs.

Keywords: Monte Carlo data, Binary decision diagrams, Neural network modeling, Shortest path length estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1555
32611 The Surface Adsorption of Nano-pore Template

Authors: M. J. Kao, S.F. Chang, C.C. Chen, C.G. Kuo

Abstract:

This paper aims to fabricated high quality anodic aluminum oxide (AAO) film by anodization method. AAO pore size, pore density, and film thickness can be controlled in 10~500 nm, 108~1011 pore.cm-2, and 1~100 μm. AAO volume and surface area can be computed based on structural parameters such as thickness, pore size, pore density, and sample size. Base on the thetorical calculation, AAO has 100 μm thickness with 15 nm, 60 nm, and 500 nm pore diameters AAO surface areas are 1225.2 cm2, 3204.4 cm2, and 549.7 cm2, respectively. The large unit surface area which is useful for adsorption application. When AAO adsorbed pH indictor of bromphenol blue presented a sensitive pH detection of solution change. This testing method can further be used for the precise measurement of biotechnology, convenience measurement of industrial engineering.

Keywords: AAO, Pore, Surface area, Adsorption, Indicator

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2089
32610 A Systematic Review for the Latest Development in Requirement Engineering

Authors: Ishraq A. Al-Fataftah, Ayman A. Issa

Abstract:

Requirement engineering has been the subject of large volume of researches due to the significant role it plays in the software development life cycle. However, dynamicity of software industry is much faster than advances in requirements engineering approaches. Therefore, this paper aims to systematically review and evaluate the current research in requirement engineering and identify new research trends and direction in this field. In addition, various research methods associated with the Evaluation-based techniques and empirical study are highlighted for the requirements engineering field. Finally, challenges and recommendations on future directions research are presented based on the research team observations during this study.

Keywords: Current trends, Future trends, Requirement Engineering

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2336
32609 Stability Enhancement of a Large-Scale Power System Using Power System Stabilizer Based on Adaptive Neuro Fuzzy Inference System

Authors: Agung Budi Muljono, I Made Ginarsa, I Made Ari Nrartha

Abstract:

A large-scale power system (LSPS) consists of two or more sub-systems connected by inter-connecting transmission. Loading pattern on an LSPS always changes from time to time and varies depend on consumer need. The serious instability problem is appeared in an LSPS due to load fluctuation in all of the bus. Adaptive neuro-fuzzy inference system (ANFIS)-based power system stabilizer (PSS) is presented to cover the stability problem and to enhance the stability of an LSPS. The ANFIS control is presented because the ANFIS control is more effective than Mamdani fuzzy control in the computation aspect. Simulation results show that the presented PSS is able to maintain the stability by decreasing peak overshoot to the value of −2.56 × 10−5 pu for rotor speed deviation Δω2−3. The presented PSS also makes the settling time to achieve at 3.78 s on local mode oscillation. Furthermore, the presented PSS is able to improve the peak overshoot and settling time of Δω3−9 to the value of −0.868 × 10−5 pu and at the time of 3.50 s for inter-area oscillation.

Keywords: ANFIS, large-scale, power system, PSS, stability enhancement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1148
32608 Detection of Linkages Between Extreme Flow Measures and Climate Indices

Authors: Mohammed Sharif, Donald Burn

Abstract:

Large scale climate signals and their teleconnections can influence hydro-meteorological variables on a local scale. Several extreme flow and timing measures, including high flow and low flow measures, from 62 hydrometric stations in Canada are investigated to detect possible linkages with several large scale climate indices. The streamflow data used in this study are derived from the Canadian Reference Hydrometric Basin Network and are characterized by relatively pristine and stable land-use conditions with a minimum of 40 years of record. A composite analysis approach was used to identify linkages between extreme flow and timing measures and climate indices. The approach involves determining the 10 highest and 10 lowest values of various climate indices from the data record. Extreme flow and timing measures for each station were examined for the years associated with the 10 largest values and the years associated with the 10 smallest values. In each case, a re-sampling approach was applied to determine if the 10 values of extreme flow measures differed significantly from the series mean. Results indicate that several stations are impacted by the large scale climate indices considered in this study. The results allow the determination of any relationship between stations that exhibit a statistically significant trend and stations for which the extreme measures exhibit a linkage with the climate indices.

Keywords: flood analysis, low-flow events, climate change, trend analysis, Canada

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1536