Search results for: time series models
8529 Interoperability in Component Based Software Development
Authors: M. Madiajagan, B. Vijayakumar
Abstract:
The ability of information systems to operate in conjunction with each other encompassing communication protocols, hardware, software, application, and data compatibility layers. There has been considerable work in industry on the development of component interoperability models, such as CORBA, (D)COM and JavaBeans. These models are intended to reduce the complexity of software development and to facilitate reuse of off-the-shelf components. The focus of these models is syntactic interface specification, component packaging, inter-component communications, and bindings to a runtime environment. What these models lack is a consideration of architectural concerns – specifying systems of communicating components, explicitly representing loci of component interaction, and exploiting architectural styles that provide well-understood global design solutions. The development of complex business applications is now focused on an assembly of components available on a local area network or on the net. These components must be localized and identified in terms of available services and communication protocol before any request. The first part of the article introduces the base concepts of components and middleware while the following sections describe the different up-todate models of communication and interaction and the last section shows how different models can communicate among themselves.
Keywords: Interoperability, component packaging, communication technology, heterogeneous platform, component interface, middleware.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27878528 A Discrete-Event-Simulation Approach for Logistic Systems with Real Time Resource Routing and VR Integration
Authors: Gerrit Alves, Jürgen Roßmann, Roland Wischnewski
Abstract:
Today, transport and logistic systems are often tightly integrated in the production. Lean production and just-in-time delivering create multiple constraints that have to be fulfilled. As transport networks often have evolved over time they are very expensive to change. This paper describes a discrete-event-simulation system which simulates transportation models using real time resource routing and collision avoidance. It allows for the specification of own control algorithms and validation of new strategies. The simulation is integrated into a virtual reality (VR) environment and can be displayed in 3-D to show the progress. Simulation elements can be selected through VR metaphors. All data gathered during the simulation can be presented as a detailed summary afterwards. The included cost-benefit calculation can help to optimize the financial outcome. The operation of this approach is shown by the example of a timber harvest simulation.Keywords: Discrete-Event-Simulation, Logistic, Simulation, Virtual Reality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18818527 Small Signal Stability Assessment Employing PSO Based TCSC Controller with Comparison to GA Based Design
Authors: D. Mondal, A. Chakrabarti, A. Sengupta
Abstract:
This paper aims to select the optimal location and setting parameters of TCSC (Thyristor Controlled Series Compensator) controller using Particle Swarm Optimization (PSO) and Genetic Algorithm (GA) to mitigate small signal oscillations in a multimachine power system. Though Power System Stabilizers (PSSs) are prime choice in this issue, installation of FACTS device has been suggested here in order to achieve appreciable damping of system oscillations. However, performance of any FACTS devices highly depends upon its parameters and suitable location in the power network. In this paper PSO as well as GA based techniques are used separately and compared their performances to investigate this problem. The results of small signal stability analysis have been represented employing eigenvalue as well as time domain response in face of two common power system disturbances e.g., varying load and transmission line outage. It has been revealed that the PSO based TCSC controller is more effective than GA based controller even during critical loading condition.Keywords: Genetic Algorithm, Particle Swarm Optimization, Small Signal Stability, Thyristor Controlled Series Compensator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19568526 Combined Effect of Heat Stimulation and Delay Addition of Superplasticizer with Slag on Fresh and Hardened Property of Mortar
Authors: Antoni Wibowo, Harry Pujianto, Dewi Retno Sari Saputro
Abstract:
The stock market can provide huge profits in a relatively short time in financial sector; however, it also has a high risk for investors and traders if they are not careful to look the factors that affect the stock market. Therefore, they should give attention to the dynamic fluctuations and movements of the stock market to optimize profits from their investment. In this paper, we present a nonlinear autoregressive exogenous model (NARX) to predict the movements of stock market; especially, the movements of the closing price index. As case study, we consider to predict the movement of the closing price in Indonesia composite index (IHSG) and choose the best structures of NARX for IHSG’s prediction.
Keywords: NARX, prediction, stock market, time series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8178525 Improvement Approach on Rotor Time Constant Adaptation with Optimum Flux in IFOC for Induction Machines Drives
Authors: S. Grouni, R. Ibtiouen, M. Kidouche, O. Touhami
Abstract:
Induction machine models used for steady-state and transient analysis require machine parameters that are usually considered design parameters or data. The knowledge of induction machine parameters is very important for Indirect Field Oriented Control (IFOC). A mismatched set of parameters will degrade the response of speed and torque control. This paper presents an improvement approach on rotor time constant adaptation in IFOC for Induction Machines (IM). Our approach tends to improve the estimation accuracy of the fundamental model for flux estimation. Based on the reduced order of the IM model, the rotor fluxes and rotor time constant are estimated using only the stator currents and voltages. This reduced order model offers many advantages for real time identification parameters of the IM.Keywords: Indirect Field Oriented Control (IFOC), InductionMachine (IM), Rotor Time Constant, Parameters ApproachAdaptation. Optimum rotor flux.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17078524 Synthesis, Structure and Functional Characteristics of Solid Electrolytes Based on Lanthanum Niobates
Authors: Maria V. Morozova, Yulia V. Emelyanova, Anastasia A. Levina, Elena S. Buyanova, Zoya A. Mikhaylovskaya, Sofia A. Petrova
Abstract:
The solid solutions of lanthanum niobates substituted by yttrium, bismuth and tungsten were synthesized. The structure of the solid solutions is either LaNbO4-based monoclinic or BiNbO4-based triclinic. The series where niobium is substituted by tungsten on B site reveals phase-modulated structure. The values of cell parameters decrease with increasing the dopant concentration for all samples except the tungsten series although the latter show higher total conductivity.Keywords: Impedance spectroscopy, LaNbO4, lanthanum ortho-niobates.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15898523 Partial Derivatives and Optimization Problem on Time Scales
Authors: Francisco Miranda
Abstract:
The optimization problem using time scales is studied. Time scale is a model of time. The language of time scales seems to be an ideal tool to unify the continuous-time and the discrete-time theories. In this work we present necessary conditions for a solution of an optimization problem on time scales. To obtain that result we use properties and results of the partial diamond-alpha derivatives for continuous-multivariable functions. These results are also presented here.Keywords: Lagrange multipliers, mathematical programming, optimization problem, time scales.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17258522 Prediction of Post Underwater Shock Properties of Polymer - Clay/Silica Hybrid Nanocomposites through Regression Models
Authors: D. Lingaraju, K. Ramji, M. Pramiladevi, U. Rajyalakshmi
Abstract:
Exploding concentrated underwater charges to damage underwater structures such as ship hulls is a part of naval warfare strategies. Adding small amounts of foreign particles (like clay or silica) of nanosize significantly improves the engineering properties of the polymers. In the present work the clay in terms 1, 2 and 3 percent by weight was surface treated with a suitable silane agent. The hybrid nanocomposite was prepared by the hand lay-up technique. Mathematical regression models have been employed for theoretical prediction. This will result in considerable savings in terms of project time, effort and cost.Keywords: ANOVA, clay, halloysite, nanocomposites, underwater shock, regression, silica.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21888521 Transient Hydrodynamic and Thermal Behaviors of Fluid Flow in a Vertical Porous Microchannel under the Effect of Hyperbolic Heat Conduction Model
Authors: A. F. Khadrawi
Abstract:
The transient hydrodynamics and thermal behaviors of fluid flow in open-ended vertical parallel-plate porous microchannel are investigated semi-analytically under the effect of the hyperbolic heat conduction model. The model that combines both the continuum approach and the possibility of slip at the boundary is adopted in the study. The Effects of Knudsen number , Darcy number , and thermal relaxation time on the microchannel hydrodynamics and thermal behaviors are investigated using the hyperbolic heat conduction models. It is found that as increases the slip in the hydrodynamic and thermal boundary condition increases. This slip in the hydrodynamic boundary condition increases as increases. Also, the slip in the thermal boundary condition increases as decreases especially the early stage of time.Keywords: free convection, hyperbolic heat conduction, macroscopic heat conduction models in microchannel, porous media, vertical microchannel, microchannel thermal, hydrodynamic behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19278520 Parametric Cost Estimating Relationships for Design Effort Estimation
Authors: Adil Salam, Nadia Bhuiyan, Gerard J. Gouw
Abstract:
The Canadian aerospace industry faces many challenges. One of them is the difficulty in estimating costs. In particular, the design effort required in a project impacts resource requirements and lead-time, and consequently the final cost. This paper presents the findings of a case study conducted for recognized global leader in the design and manufacturing of aircraft engines. The study models parametric cost estimation relationships to estimate the design effort of integrated blade-rotor low-pressure compressor fans. Several effort drivers are selected to model the relationship. Comparative analyses of three types of models are conducted. The model with the best accuracy and significance in design estimation is retained.
Keywords: Effort estimation, design, aerospace.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25788519 Specialized Reduced Models of Dynamic Flows in 2-Stroke Engines
Authors: S. Cagin, X. Fischer, E. Delacourt, N. Bourabaa, C. Morin, D. Coutellier, B. Carré, S. Loumé
Abstract:
The complexity of scavenging by ports and its impact on engine efficiency create the need to understand and to model it as realistically as possible. However, there are few empirical scavenging models and these are highly specialized. In a design optimization process, they appear very restricted and their field of use is limited. This paper presents a comparison of two methods to establish and reduce a model of the scavenging process in 2-stroke diesel engines. To solve the lack of scavenging models, a CFD model has been developed and is used as the referent case. However, its large size requires a reduction. Two techniques have been tested depending on their fields of application: The NTF method and neural networks. They both appear highly appropriate drastically reducing the model’s size (over 90% reduction) with a low relative error rate (under 10%). Furthermore, each method produces a reduced model which can be used in distinct specialized fields of application: the distribution of a quantity (mass fraction for example) in the cylinder at each time step (pseudo-dynamic model) or the qualification of scavenging at the end of the process (pseudo-static model).
Keywords: Diesel engine, Design optimization, Model reduction, Neural network, NTF algorithm, Scavenging.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13308518 Increased Signal to Noise Ratio in P300 Potentials by the Method of Coherent Self-Averaging in BCI Systems
Authors: Ricardo Espinosa
Abstract:
The coherent Self-Averaging (CSA), is a new method proposed in this work; applied to simulated signals evoked potentials related to events (ERP) to find the wave P300, useful systems in the brain computer interface (BCI). The CSA method cleans signal in the time domain of white noise through of successive averaging of a single signal. The method is compared with the traditional method, coherent averaging or synchronized (CA), showing optimal results in the improvement of the signal to noise ratio (SNR). The method of CSA is easy to implement, robust and applicable to any physiological time series contaminated with white noise
Keywords: Evoked potentials, wave P300, Coherent Self-averaging, brain - computer interface (BCI).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21448517 Grid Coordination with Marketmaker Agents
Authors: Xin Bai, Kresimir Sivoncik, Damla Turgut, Ladislau Bölöni
Abstract:
Market based models are frequently used in the resource allocation on the computational grid. However, as the size of the grid grows, it becomes difficult for the customer to negotiate directly with all the providers. Middle agents are introduced to mediate between the providers and customers and facilitate the resource allocation process. The most frequently deployed middle agents are the matchmakers and the brokers. The matchmaking agent finds possible candidate providers who can satisfy the requirements of the consumers, after which the customer directly negotiates with the candidates. The broker agents are mediating the negotiation with the providers in real time. In this paper we present a new type of middle agent, the marketmaker. Its operation is based on two parallel operations - through the investment process the marketmaker is acquiring resources and resource reservations in large quantities, while through the resale process it sells them to the customers. The operation of the marketmaker is based on the fact that through its global view of the grid it can perform a more efficient resource allocation than the one possible in one-to-one negotiations between the customers and providers. We present the operation and algorithms governing the operation of the marketmaker agent, contrasting it with the matchmaker and broker agents. Through a series of simulations in the task oriented domain we compare the operation of the three agents types. We find that the use of marketmaker agent leads to a better performance in the allocation of large tasks and a significant reduction of the messaging overhead.Keywords: grid computing, autonomous agents, market-basedgrid
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15308516 Simulation of Reactive Distillation: Comparison of Equilibrium and Nonequilibrium Stage Models
Authors: Asfaw Gezae Daful
Abstract:
In the present study, two distinctly different approaches are followed for modeling of reactive distillation column, the equilibrium stage model and the nonequilibrium stage model. These models are simulated with a computer code developed in the present study using MATLAB programming. In the equilibrium stage models, the vapor and liquid phases are assumed to be in equilibrium and allowance is made for finite reaction rates, where as in the nonequilibrium stage models simultaneous mass transfer and reaction rates are considered. These simulated model results are validated from the experimental data reported in the literature. The simulated results of equilibrium and nonequilibrium models are compared for concentration, temperature and reaction rate profiles in a reactive distillation column for Methyl Tert Butyle Ether (MTBE) production. Both the models show similar trend for the concentration, temperature and reaction rate profiles but the nonequilibrium model predictions are higher and closer to the experimental values reported in the literature.
Keywords: Reactive Distillation, Equilibrium model, Nonequilibrium model, Methyl Tert-Butyl Ether
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42068515 Classification of Business Models of Italian Bancassurance by Balance Sheet Indicators
Authors: Andrea Bellucci, Martina Tofi
Abstract:
The aim of paper is to analyze business models of bancassurance in Italy for life business. The life insurance business is very developed in the Italian market and banks branches have 80% of the market share. Given its maturity, the life insurance market needs to consolidate its organizational form to allow for the development of non-life business, which nowadays collects few premiums but represents a great opportunity to enlarge the market share of bancassurance using its strength in the distribution channel while the market share of independent agents is decreasing. Starting with the main business model of bancassurance for life business, this paper will analyze the performances of life companies in the Italian market by balance sheet indicators and by main discriminant variables of business models. The study will observe trends from 2013 to 2015 for the Italian market by exploiting a database managed by Associazione Nazionale delle Imprese di Assicurazione (ANIA). The applied approach is based on a bottom-up analysis starting with variables and indicators to define business models’ classification. The statistical classification algorithm proposed by Ward is employed to design business models’ profiles. Results from the analysis will be a representation of the main business models built by their profile related to indicators. In that way, an unsupervised analysis is developed that has the limit of its judgmental dimension based on research opinion, but it is possible to obtain a design of effective business models.
Keywords: Balance sheet indicators, Bancassurance, business models, ward algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12618514 Experimentation on Piercing with Abrasive Waterjet
Authors: Johan Fredin, Anders Jönsson
Abstract:
Abrasive waterjet cutting (AWJ) is a highly efficient method for cutting almost any type of material. When holes shall be cut the waterjet first needs to pierce the material.This paper presents a vast experimental analysis of piercing parameters effect on piercing time. Results from experimentation on feed rates, work piece thicknesses, abrasive flow rates, standoff distances and water pressure are also presented as well as studies on three methods for dynamic piercing. It is shown that a large amount of time and resources can be saved by choosing the piercing parameters in a correct way. The large number of experiments puts demands on the experimental setup. An automated experimental setup including piercing detection is presented to enable large series of experiments to be carried out efficiently.Keywords: Waterjet cutting, Piercing, Experimentation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24478513 The Strengths and Limitations of the Statistical Modeling of Complex Social Phenomenon: Focusing on SEM, Path Analysis, or Multiple Regression Models
Authors: Jihye Jeon
Abstract:
This paper analyzes the conceptual framework of three statistical methods, multiple regression, path analysis, and structural equation models. When establishing research model of the statistical modeling of complex social phenomenon, it is important to know the strengths and limitations of three statistical models. This study explored the character, strength, and limitation of each modeling and suggested some strategies for accurate explaining or predicting the causal relationships among variables. Especially, on the studying of depression or mental health, the common mistakes of research modeling were discussed.Keywords: Multiple regression, path analysis, structural equation models, statistical modeling, social and psychological phenomenon.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 92508512 A Study of Classification Models to Predict Drill-Bit Breakage Using Degradation Signals
Authors: Bharatendra Rai
Abstract:
Cutting tools are widely used in manufacturing processes and drilling is the most commonly used machining process. Although drill-bits used in drilling may not be expensive, their breakage can cause damage to expensive work piece being drilled and at the same time has major impact on productivity. Predicting drill-bit breakage, therefore, is important in reducing cost and improving productivity. This study uses twenty features extracted from two degradation signals viz., thrust force and torque. The methodology used involves developing and comparing decision tree, random forest, and multinomial logistic regression models for classifying and predicting drill-bit breakage using degradation signals.
Keywords: Degradation signal, drill-bit breakage, random forest, multinomial logistic regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22428511 The Classification Model for Hard Disk Drive Functional Tests under Sparse Data Conditions
Authors: S. Pattanapairoj, D. Chetchotsak
Abstract:
This paper proposed classification models that would be used as a proxy for hard disk drive (HDD) functional test equitant which required approximately more than two weeks to perform the HDD status classification in either “Pass" or “Fail". These models were constructed by using committee network which consisted of a number of single neural networks. This paper also included the method to solve the problem of sparseness data in failed part, which was called “enforce learning method". Our results reveal that the constructed classification models with the proposed method could perform well in the sparse data conditions and thus the models, which used a few seconds for HDD classification, could be used to substitute the HDD functional tests.Keywords: Sparse data, Classifications, Committee network
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17368510 Laplace Decomposition Approximation Solution for a System of Multi-Pantograph Equations
Authors: M. A. Koroma, C. Zhan, A. F. Kamara, A. B. Sesay
Abstract:
In this work we adopt a combination of Laplace transform and the decomposition method to find numerical solutions of a system of multi-pantograph equations. The procedure leads to a rapid convergence of the series to the exact solution after computing a few terms. The effectiveness of the method is demonstrated in some examples by obtaining the exact solution and in others by computing the absolute error which decreases as the number of terms of the series increases.
Keywords: Laplace decomposition, pantograph equations, exact solution, numerical solution, approximate solution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16498509 A GPU Based Texture Mapping Technique for 3D Models Using Multi-View Images
Authors: In Lee, Kyung-Kyu Kang, Jaewoon Lee, Dongho Kim
Abstract:
Previous the 3D model texture generation from multi-view images and mapping algorithms has issues in the texture chart generation which are the self-intersection and the concentration of the texture in texture space. Also we may suffer from some problems due to the occluded areas, such as inside parts of thighs. In this paper we propose a texture mapping technique for 3D models using multi-view images on the GPU. We do texture mapping directly on the GPU fragment shader per pixel without generation of the texture map. And we solve for the occluded area using the 3D model depth information. Our method needs more calculation on the GPU than previous works, but it has shown real-time performance and previously mentioned problems do not occur.
Keywords: Texture Mapping, Multi-view Images, Camera Calibration, GPU Shader.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19478508 Simulating the Dynamics of Distribution of Hazardous Substances Emitted by Motor Engines in a Residential Quarter
Authors: S. Grishin
Abstract:
This article is dedicated to development of mathematical models for determining the dynamics of concentration of hazardous substances in urban turbulent atmosphere. Development of the mathematical models implied taking into account the time-space variability of the fields of meteorological items and such turbulent atmosphere data as vortex nature, nonlinear nature, dissipativity and diffusivity. Knowing the turbulent airflow velocity is not assumed when developing the model. However, a simplified model implies that the turbulent and molecular diffusion ratio is a piecewise constant function that changes depending on vertical distance from the earth surface. Thereby an important assumption of vertical stratification of urban air due to atmospheric accumulation of hazardous substances emitted by motor vehicles is introduced into the mathematical model. The suggested simplified non-linear mathematical model of determining the sought exhaust concentration at a priori unknown turbulent flow velocity through non-degenerate transformation is reduced to the model which is subsequently solved analytically.Keywords: Urban ecology, time-dependent mathematical model, exhaust concentration, turbulent and molecular diffusion, airflow velocity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14118507 Wireless Sensor Network to Help Low Incomes Farmers to Face Drought Impacts
Authors: Fantazi Walid, Ezzedine Tahar, Bargaoui Zoubeida
Abstract:
This research presents the main ideas to implement an intelligent system composed by communicating wireless sensors measuring environmental data linked to drought indicators (such as air temperature, soil moisture , etc...). On the other hand, the setting up of a spatio temporal database communicating with a Web mapping application for a monitoring in real time in activity 24:00 /day, 7 days/week is proposed to allow the screening of the drought parameters time evolution and their extraction. Thus this system helps detecting surfaces touched by the phenomenon of drought. Spatio-temporal conceptual models seek to answer the users who need to manage soil water content for irrigating or fertilizing or other activities pursuing crop yield augmentation. Effectively, spatiotemporal conceptual models enable users to obtain a diagram of readable and easy data to apprehend. Based on socio-economic information, it helps identifying people impacted by the phenomena with the corresponding severity especially that this information is accessible by farmers and stakeholders themselves. The study will be applied in Siliana watershed Northern Tunisia.Keywords: WSN, database spatio-temporal, GIS, web-mapping, indicator of drought.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24458506 Imputing Missing Data in Electronic Health Records: A Comparison of Linear and Non-Linear Imputation Models
Authors: Alireza Vafaei Sadr, Vida Abedi, Jiang Li, Ramin Zand
Abstract:
Missing data is a common challenge in medical research and can lead to biased or incomplete results. When the data bias leaks into models, it further exacerbates health disparities; biased algorithms can lead to misclassification and reduced resource allocation and monitoring as part of prevention strategies for certain minorities and vulnerable segments of patient populations, which in turn further reduce data footprint from the same population – thus, a vicious cycle. This study compares the performance of six imputation techniques grouped into Linear and Non-Linear models, on two different real-world electronic health records (EHRs) datasets, representing 17864 patient records. The mean absolute percentage error (MAPE) and root mean squared error (RMSE) are used as performance metrics, and the results show that the Linear models outperformed the Non-Linear models in terms of both metrics. These results suggest that sometimes Linear models might be an optimal choice for imputation in laboratory variables in terms of imputation efficiency and uncertainty of predicted values.
Keywords: EHR, Machine Learning, imputation, laboratory variables, algorithmic bias.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738505 Stochastic Subspace Modelling of Turbulence
Authors: M. T. Sichani, B. J. Pedersen, S. R. K. Nielsen
Abstract:
Turbulence of the incoming wind field is of paramount importance to the dynamic response of civil engineering structures. Hence reliable stochastic models of the turbulence should be available from which time series can be generated for dynamic response and structural safety analysis. In the paper an empirical cross spectral density function for the along-wind turbulence component over the wind field area is taken as the starting point. The spectrum is spatially discretized in terms of a Hermitian cross-spectral density matrix for the turbulence state vector which turns out not to be positive definite. Since the succeeding state space and ARMA modelling of the turbulence rely on the positive definiteness of the cross-spectral density matrix, the problem with the non-positive definiteness of such matrices is at first addressed and suitable treatments regarding it are proposed. From the adjusted positive definite cross-spectral density matrix a frequency response matrix is constructed which determines the turbulence vector as a linear filtration of Gaussian white noise. Finally, an accurate state space modelling method is proposed which allows selection of an appropriate model order, and estimation of a state space model for the vector turbulence process incorporating its phase spectrum in one stage, and its results are compared with a conventional ARMA modelling method.Keywords: Turbulence, wind turbine, complex coherence, state space modelling, ARMA modelling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16468504 Seismic Analysis of a S-Curved Viaduct using Stick and Finite Element Models
Authors: Sourabh Agrawal, Ashok K. Jain
Abstract:
Stick models are widely used in studying the behaviour of straight as well as skew bridges and viaducts subjected to earthquakes while carrying out preliminary studies. The application of such models to highly curved bridges continues to pose challenging problems. A viaduct proposed in the foothills of the Himalayas in Northern India is chosen for the study. It is having 8 simply supported spans @ 30 m c/c. It is doubly curved in horizontal plane with 20 m radius. It is inclined in vertical plane as well. The superstructure consists of a box section. Three models have been used: a conventional stick model, an improved stick model and a 3D finite element model. The improved stick model is employed by making use of body constraints in order to study its capabilities. The first 8 frequencies are about 9.71% away in the latter two models. Later the difference increases to 80% in 50th mode. The viaduct was subjected to all three components of the El Centro earthquake of May 1940. The numerical integration was carried out using the Hilber- Hughes-Taylor method as implemented in SAP2000. Axial forces and moments in the bridge piers as well as lateral displacements at the bearing levels are compared for the three models. The maximum difference in the axial forces and bending moments and displacements vary by 25% between the improved and finite element model. Whereas, the maximum difference in the axial forces, moments, and displacements in various sections vary by 35% between the improved stick model and equivalent straight stick model. The difference for torsional moment was as high as 75%. It is concluded that the stick model with body constraints to model the bearings and expansion joints is not desirable in very sharp S curved viaducts even for preliminary analysis. This model can be used only to determine first 10 frequency and mode shapes but not for member forces. A 3D finite element analysis must be carried out for meaningful results.Keywords: Bearing, body constraint, box girder, curved viaduct, expansion joint, finite element, link element, seismic, stick model, time history analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23588503 Towards Model-Driven Communications
Authors: Antonio Natali, Ambra Molesini
Abstract:
In modern distributed software systems, the issue of communication among composing parts represents a critical point, but the idea of extending conventional programming languages with general purpose communication constructs seems difficult to realize. As a consequence, there is a (growing) gap between the abstraction level required by distributed applications and the concepts provided by platforms that enable communication. This work intends to discuss how the Model Driven Software Development approach can be considered as a mature technology to generate in automatic way the schematic part of applications related to communication, by providing at the same time high level specialized languages useful in all the phases of software production. To achieve the goal, a stack of languages (meta-meta¬models) has been introduced in order to describe – at different levels of abstraction – the collaborative behavior of generic entities in terms of communication actions related to a taxonomy of messages. Finally, the generation of platforms for communication is viewed as a form of specification of language semantics, that provides executable models of applications together with model-checking supports and effective runtime environments.
Keywords: Interactions, specific languages, meta-models, model driven development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18508502 Bioprocess Optimization Based On Relevance Vector Regression Models and Evolutionary Programming Technique
Authors: R. Simutis, V. Galvanauskas, D. Levisauskas, J. Repsyte
Abstract:
This paper proposes a bioprocess optimization procedure based on Relevance Vector Regression models and evolutionary programming technique. Relevance Vector Regression scheme allows developing a compact and stable data-based process model avoiding time-consuming modeling expenses. The model building and process optimization procedure could be done in a half-automated way and repeated after every new cultivation run. The proposed technique was tested in a simulated mammalian cell cultivation process. The obtained results are promising and could be attractive for optimization of industrial bioprocesses.
Keywords: Bioprocess optimization, Evolutionary programming, Relevance Vector Regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21948501 Transformation Method CIM to PIM: From Business Processes Models Defined in BPMN to Use Case and Class Models Defined in UML
Authors: Y. Rhazali, Y. Hadi, A. Mouloudi
Abstract:
This paper proposes a method to automatic transformation of CIM level to PIM level respecting the MDA approach. Our proposal is based on creating a good CIM level through well-defined rules allowing as achieving rich models that contain relevant information to facilitate the task of the transformation to the PIM level. We define, thereafter, an appropriate PIM level through various UML diagram. Next, we propose set rules to move from CIM to PIM. Our method follows the MDA approach by considering the business dimension in the CIM level through the use BPMN, standard modeling business of OMG, and the use of UML in PIM advocated by MDA in this level.
Keywords: Model transformation, MDA, CIM, PIM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36738500 Review of the Model-Based Supply Chain Management Research in the Construction Industry
Authors: Aspasia Koutsokosta, Stefanos Katsavounis
Abstract:
This paper reviews the model-based qualitative and quantitative Operations Management research in the context of Construction Supply Chain Management (CSCM). Construction industry has been traditionally blamed for low productivity, cost and time overruns, waste, high fragmentation and adversarial relationships. The construction industry has been slower than other industries to employ the Supply Chain Management (SCM) concept and develop models that support the decision-making and planning. However the last decade there is a distinct shift from a project-based to a supply-based approach of construction management. CSCM comes up as a new promising management tool of construction operations and improves the performance of construction projects in terms of cost, time and quality. Modeling the Construction Supply Chain (CSC) offers the means to reap the benefits of SCM, make informed decisions and gain competitive advantage. Different modeling approaches and methodologies have been applied in the multi-disciplinary and heterogeneous research field of CSCM. The literature review reveals that a considerable percentage of the CSC modeling research accommodates conceptual or process models which present general management frameworks and do not relate to acknowledged soft Operations Research methods. We particularly focus on the model-based quantitative research and categorize the CSCM models depending on their scope, objectives, modeling approach, solution methods and software used. Although over the last few years there has been clearly an increase of research papers on quantitative CSC models, we identify that the relevant literature is very fragmented with limited applications of simulation, mathematical programming and simulation-based optimization. Most applications are project-specific or study only parts of the supply system. Thus, some complex interdependencies within construction are neglected and the implementation of the integrated supply chain management is hindered. We conclude this paper by giving future research directions and emphasizing the need to develop optimization models for integrated CSCM. We stress that CSC modeling needs a multi-dimensional, system-wide and long-term perspective. Finally, prior applications of SCM to other industries have to be taken into account in order to model CSCs, but not without translating the generic concepts to the context of construction industry.Keywords: Construction supply chain management, modeling, operations research, optimization and simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2825