Search results for: polynomial segment model.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7733

Search results for: polynomial segment model.

7133 Dual-Network Memory Model for Temporal Sequences

Authors: Motonobu Hattori, Rina Suzuki

Abstract:

In neural networks, when new patters are learned by a network, they radically interfere with previously stored patterns. This drawback is called catastrophic forgetting. We have already proposed a biologically inspired dual-network memory model which can much reduce this forgetting for static patterns. In this model, information is first stored in the hippocampal network, and thereafter, it is transferred to the neocortical network using pseudopatterns. Because temporal sequence learning is more important than static pattern learning in the real world, in this study, we improve our conventional  dual-network memory model so that it can deal with temporal sequences without catastrophic forgetting. The computer simulation results show the effectiveness of the proposed dual-network memory model.  

Keywords: Catastrophic forgetting, dual-network, temporal sequences.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1424
7132 Level Set and Morphological Operation Techniques in Application of Dental Image Segmentation

Authors: Abdolvahab Ehsani Rad, Mohd Shafry Mohd Rahim, Alireza Norouzi

Abstract:

Medical image analysis is one of the great effects of computer image processing. There are several processes to analysis the medical images which the segmentation process is one of the challenging and most important step. In this paper the segmentation method proposed in order to segment the dental radiograph images. Thresholding method has been applied to simplify the images and to morphologically open binary image technique performed to eliminate the unnecessary regions on images. Furthermore, horizontal and vertical integral projection techniques used to extract the each individual tooth from radiograph images. Segmentation process has been done by applying the level set method on each extracted images. Nevertheless, the experiments results by 90% accuracy demonstrate that proposed method achieves high accuracy and promising result.

Keywords: Integral production, level set method, morphological operation, segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4232
7131 A Nonlinear ODE System for the Unsteady Hydrodynamic Force – A New Approach

Authors: Osama A. Marzouk

Abstract:

We propose a reduced-ordermodel for the instantaneous hydrodynamic force on a cylinder. The model consists of a system of two ordinary differential equations (ODEs), which can be integrated in time to yield very accurate histories of the resultant force and its direction. In contrast to several existing models, the proposed model considers the actual (total) hydrodynamic force rather than its perpendicular or parallel projection (the lift and drag), and captures the complete force rather than the oscillatory part only. We study and provide descriptions of the relationship between the model parameters, evaluated utilizing results from numerical simulations, and the Reynolds number so that the model can be used at any arbitrary value within the considered range of 100 to 500 to provide accurate representation of the force without the need to perform timeconsuming simulations and solving the partial differential equations (PDEs) governing the flow field.

Keywords: reduced-order model, wake oscillator, nonlinear, ODEsystem

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1566
7130 A Boundary Fitted Nested Grid Model for Modelling Tsunami Propagation of 2004 Indonesian Tsunami along Southern Thailand

Authors: Md. Fazlul Karim, Esa Al-Islam

Abstract:

This paper describes the development of a boundary fitted nested grid (BFNG) model to compute tsunami propagation of 2004 Indonesian tsunami in Southern Thailand coastal waters. We develop a numerical model employing the shallow water nested model and an orthogonal boundary fitted grid to investigate the tsunami impact on the Southern Thailand due to the Indonesian tsunami of 2004. Comparisons of water surface elevation obtained from numerical simulations and field measurements are made.

Keywords: Boundary-fitted nested grid model, finite difference method, Indonesian tsunami of 2004, Southern Thailand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1798
7129 A Comparative Analysis of the Performance of COSMO and WRF Models in Quantitative Rainfall Prediction

Authors: Isaac Mugume, Charles Basalirwa, Daniel Waiswa, Mary Nsabagwa, Triphonia Jacob Ngailo, Joachim Reuder, Sch¨attler Ulrich, Musa Semujju

Abstract:

The Numerical weather prediction (NWP) models are considered powerful tools for guiding quantitative rainfall prediction. A couple of NWP models exist and are used at many operational weather prediction centers. This study considers two models namely the Consortium for Small–scale Modeling (COSMO) model and the Weather Research and Forecasting (WRF) model. It compares the models’ ability to predict rainfall over Uganda for the period 21st April 2013 to 10th May 2013 using the root mean square (RMSE) and the mean error (ME). In comparing the performance of the models, this study assesses their ability to predict light rainfall events and extreme rainfall events. All the experiments used the default parameterization configurations and with same horizontal resolution (7 Km). The results show that COSMO model had a tendency of largely predicting no rain which explained its under–prediction. The COSMO model (RMSE: 14.16; ME: -5.91) presented a significantly (p = 0.014) higher magnitude of error compared to the WRF model (RMSE: 11.86; ME: -1.09). However the COSMO model (RMSE: 3.85; ME: 1.39) performed significantly (p = 0.003) better than the WRF model (RMSE: 8.14; ME: 5.30) in simulating light rainfall events. All the models under–predicted extreme rainfall events with the COSMO model (RMSE: 43.63; ME: -39.58) presenting significantly higher error magnitudes than the WRF model (RMSE: 35.14; ME: -26.95). This study recommends additional diagnosis of the models’ treatment of deep convection over the tropics.

Keywords: Comparative performance, the COSMO model, the WRF model, light rainfall events, extreme rainfall events.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1540
7128 Application of Build-up and Wash-off Models for an East-Australian Catchment

Authors: Iqbal Hossain, Monzur Alam Imteaz, Mohammed Iqbal Hossain

Abstract:

Estimation of stormwater pollutants is a pre-requisite for the protection and improvement of the aquatic environment and for appropriate management options. The usual practice for the stormwater quality prediction is performed through water quality modeling. However, the accuracy of the prediction by the models depends on the proper estimation of model parameters. This paper presents the estimation of model parameters for a catchment water quality model developed for the continuous simulation of stormwater pollutants from a catchment to the catchment outlet. The model is capable of simulating the accumulation and transportation of the stormwater pollutants; suspended solids (SS), total nitrogen (TN) and total phosphorus (TP) from a particular catchment. Rainfall and water quality data were collected for the Hotham Creek Catchment (HTCC), Gold Coast, Australia. Runoff calculations from the developed model were compared with the calculated discharges from the widely used hydrological models, WBNM and DRAINS. Based on the measured water quality data, model water quality parameters were calibrated for the above-mentioned catchment. The calibrated parameters are expected to be helpful for the best management practices (BMPs) of the region. Sensitivity analyses of the estimated parameters were performed to assess the impacts of the model parameters on overall model estimations of runoff water quality.

Keywords: Calibration, Model Parameters, Suspended Solids, TotalNitrogen, Total Phosphorus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2183
7127 Application of Company Financial Crisis Early Warning Model- Use of “Financial Reference Database“

Authors: Chiung-ying Lee, Chia-hua Chang

Abstract:

In July 1, 2007, Taiwan Stock Exchange (TWSE) on market observation post system (MOPS) adds a new "Financial reference database" for investors to do investment reference. This database as a warning to public offering companies listed on the public financial information and it original within eight targets. In this paper, this database provided by the indicators for the application of company financial crisis early warning model verify that the database provided by the indicator forecast for the financial crisis, whether or not companies have a high accuracy rate as opposed to domestic and foreign scholars have positive results. There is use of Logistic Regression Model application of the financial early warning model, in which no joined back-conditions is the first model, joined it in is the second model, has been taken occurred in the financial crisis of companies to research samples and then business took place before the financial crisis point with T-1 and T-2 sample data to do positive analysis. The results show that this database provided the debt ratio and net per share for the best forecast variables.

Keywords: Financial reference database, Financial early warning model, Logistic Regression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1427
7126 A Scheme of Model Verification of the Concurrent Discrete Wavelet Transform (DWT) for Image Compression

Authors: Kamrul Hasan Talukder, Koichi Harada

Abstract:

The scientific community has invested a great deal of effort in the fields of discrete wavelet transform in the last few decades. Discrete wavelet transform (DWT) associated with the vector quantization has been proved to be a very useful tool for the compression of image. However, the DWT is very computationally intensive process requiring innovative and computationally efficient method to obtain the image compression. The concurrent transformation of the image can be an important solution to this problem. This paper proposes a model of concurrent DWT for image compression. Additionally, the formal verification of the model has also been performed. Here the Symbolic Model Verifier (SMV) has been used as the formal verification tool. The system has been modeled in SMV and some properties have been verified formally.

Keywords: Computation Tree Logic, Discrete WaveletTransform, Formal Verification, Image Compression, Symbolic Model Verifier.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1749
7125 Spatial Time Series Models for Rice and Cassava Yields Based On Bayesian Linear Mixed Models

Authors: Panudet Saengseedam, Nanthachai Kantanantha

Abstract:

This paper proposes a linear mixed model (LMM) with spatial effects to forecast rice and cassava yields in Thailand at the same time. A multivariate conditional autoregressive (MCAR) model is assumed to present the spatial effects. A Bayesian method is used for parameter estimation via Gibbs sampling Markov Chain Monte Carlo (MCMC). The model is applied to the rice and cassava yields monthly data which have been extracted from the Office of Agricultural Economics, Ministry of Agriculture and Cooperatives of Thailand. The results show that the proposed model has better performance in most provinces in both fitting part and validation part compared to the simple exponential smoothing and conditional auto regressive models (CAR) from our previous study.

Keywords: Bayesian method, Linear mixed model, Multivariate conditional autoregressive model, Spatial time series.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2247
7124 A New Time Dependent, High Temperature Analytical Model for the Single-electron Box in Digital Applications

Authors: M.J. Sharifi

Abstract:

Several models have been introduced so far for single electron box, SEB, which all of them were restricted to DC response and or low temperature limit. In this paper we introduce a new time dependent, high temperature analytical model for SEB for the first time. DC behavior of the introduced model will be verified against SIMON software and its time behavior will be verified against a newly published paper regarding step response of SEB.

Keywords: Single electron box, SPICE, SIMON, Timedependent, Circuit model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1236
7123 Identification of Micromechanical Fracture Model for Predicting Fracture Performance of Steel Wires for Civil Engineering Applications

Authors: Kazeem K. Adewole, Julia M. Race, Steve J. Bull

Abstract:

The fracture performance of steel wires for civil engineering applications remains a major concern in civil engineering construction and maintenance of wire reinforced structures. The need to employ approaches that simulate micromechanical material processes which characterizes fracture in civil structures has been emphasized recently in the literature. However, choosing from the numerous micromechanics-based fracture models, and identifying their applicability and reliability remains an issue that still needs to be addressed in a greater depth. Laboratory tensile testing and finite element tensile testing simulations with the shear, ductile and Gurson-Tvergaard-Needleman’s micromechanics-based models conducted in this work reveal that the shear fracture model is an appropriate fracture model to predict the fracture performance of steel wires used for civil engineering applications. The need to consider the capability of the micromechanics-based fracture model to predict the “cup and cone” fracture exhibited by the wire in choosing the appropriate fracture model is demonstrated.

Keywords: Fracture performance, FE simulation, Shear fracture model, Ductile fracture model, Gurson-Tvergaard-Needleman fracture model, Wires.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2375
7122 A Combinatorial Model for ECG Interpretation

Authors: Costas S. Iliopoulos, Spiros Michalakopoulos

Abstract:

A new, combinatorial model for analyzing and inter- preting an electrocardiogram (ECG) is presented. An application of the model is QRS peak detection. This is demonstrated with an online algorithm, which is shown to be space as well as time efficient. Experimental results on the MIT-BIH Arrhythmia database show that this novel approach is promising. Further uses for this approach are discussed, such as taking advantage of its small memory requirements and interpreting large amounts of pre-recorded ECG data.

Keywords: Combinatorics, ECG analysis, MIT-BIH Arrhythmia Database, QRS Detection, String Algorithms

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1939
7121 Optimal Power Allocation for the Proposed Asymmetric Turbo Code for 3G Systems

Authors: K. Ramasamy, B. Balamuralithara, Mohammad Umar Siddiqi

Abstract:

We proposed a new class of asymmetric turbo encoder for 3G systems that performs well in both “water fall" and “error floor" regions in [7]. In this paper, a modified (optimal) power allocation scheme for the different bits of new class of asymmetric turbo encoder has been investigated to enhance the performance. The simulation results and performance bound for proposed asymmetric turbo code with modified Unequal Power Allocation (UPA) scheme for the frame length, N=400, code rate, r=1/3 with Log-MAP decoder over Additive White Gaussian Noise (AWGN) channel are obtained and compared with the system with typical UPA and without UPA. The performance tests are extended over AWGN channel for different frame size to verify the possibility of implementation of the modified UPA scheme for the proposed asymmetric turbo code. From the performance results, it is observed that the proposed asymmetric turbo code with modified UPA performs better than the system without UPA and with typical UPA and it provides a coding gain of 0.4 to 0.52dB.

Keywords: Asymmetric turbo code, Generator polynomial, Interleaver, UPA, WCDMA, cdma2000.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1888
7120 Color Image Segmentation using Adaptive Spatial Gaussian Mixture Model

Authors: M.Sujaritha, S. Annadurai

Abstract:

An adaptive spatial Gaussian mixture model is proposed for clustering based color image segmentation. A new clustering objective function which incorporates the spatial information is introduced in the Bayesian framework. The weighting parameter for controlling the importance of spatial information is made adaptive to the image content to augment the smoothness towards piecewisehomogeneous region and diminish the edge-blurring effect and hence the name adaptive spatial finite mixture model. The proposed approach is compared with the spatially variant finite mixture model for pixel labeling. The experimental results with synthetic and Berkeley dataset demonstrate that the proposed method is effective in improving the segmentation and it can be employed in different practical image content understanding applications.

Keywords: Adaptive; Spatial, Mixture model, Segmentation, Color.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2498
7119 An Efficient Iterative Updating Method for Damped Structural Systems

Authors: Jiashang Jiang

Abstract:

Model updating is an inverse eigenvalue problem which concerns the modification of an existing but inaccurate model with measured modal data. In this paper, an efficient gradient based iterative method for updating the mass, damping and stiffness matrices simultaneously using a few of complex measured modal data is developed. Convergence analysis indicates that the iterative solutions always converge to the unique minimum Frobenius norm symmetric solution of the model updating problem by choosing a special kind of initial matrices.

Keywords: Model updating, iterative algorithm, damped structural system, optimal approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2084
7118 Prediction on Housing Price Based on Deep Learning

Authors: Li Yu, Chenlu Jiao, Hongrun Xin, Yan Wang, Kaiyang Wang

Abstract:

In order to study the impact of various factors on the housing price, we propose to build different prediction models based on deep learning to determine the existing data of the real estate in order to more accurately predict the housing price or its changing trend in the future. Considering that the factors which affect the housing price vary widely, the proposed prediction models include two categories. The first one is based on multiple characteristic factors of the real estate. We built Convolution Neural Network (CNN) prediction model and Long Short-Term Memory (LSTM) neural network prediction model based on deep learning, and logical regression model was implemented to make a comparison between these three models. Another prediction model is time series model. Based on deep learning, we proposed an LSTM-1 model purely regard to time series, then implementing and comparing the LSTM model and the Auto-Regressive and Moving Average (ARMA) model. In this paper, comprehensive study of the second-hand housing price in Beijing has been conducted from three aspects: crawling and analyzing, housing price predicting, and the result comparing. Ultimately the best model program was produced, which is of great significance to evaluation and prediction of the housing price in the real estate industry.

Keywords: Deep learning, convolutional neural network, LSTM, housing prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4990
7117 Simulation Study on Vehicle Drag Reduction by Surface Dimples

Authors: S. F. Wong, S. S. Dol

Abstract:

Automotive designers have been trying to use dimples to reduce drag in vehicles. In this work, a car model has been applied with dimple surface with a parameter called dimple ratio DR, the ratio between the depths of the half dimple over the print diameter of the dimple, has been introduced and numerically simulated via k-ε turbulence model to study the aerodynamics performance with the increasing depth of the dimples The Ahmed body car model with 25 degree slant angle is simulated with the DR of 0.05, 0.2, 0.3 0.4 and 0.5 at Reynolds number of 176387 based on the frontal area of the car model. The geometry of dimple changes the kinematics and dynamics of flow. Complex interaction between the turbulent fluctuating flow and the mean flow escalates the turbulence quantities. The maximum level of turbulent kinetic energy occurs at DR = 0.4. It can be concluded that the dimples have generated extra turbulence energy at the surface and as a result, the application of dimples manages to reduce the drag coefficient of the car model compared to the model with smooth surface.

Keywords: Aerodynamics, Boundary Layer, Dimple, Drag, Kinetic Energy, Turbulence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2347
7116 Model Transformation with a Visual Control Flow Language

Authors: László Lengyel, Tihamér Levendovszky, Gergely Mezei, Hassan Charaf

Abstract:

Graph rewriting-based visual model processing is a widely used technique for model transformation. Visual model transformations often need to follow an algorithm that requires a strict control over the execution sequence of the transformation steps. Therefore, in Visual Model Processors (VMPs) the execution order of the transformation steps is crucial. This paper presents the visual control flow support of Visual Modeling and Transformation System (VMTS), which facilitates composing complex model transformations of simple transformation steps and executing them. The VMTS Visual Control Flow Language (VCFL) uses stereotyped activity diagrams to specify control flow structures and OCL constraints to choose between different control flow branches. This paper introduces VCFL, discusses its termination properties and provides an algorithm to support the termination analysis of VCFL transformations.

Keywords: Control Flow, Metamodel-Based Visual ModelTransformation, OCL, Termination Properties, UML.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695
7115 Automatic LV Segmentation with K-means Clustering and Graph Searching on Cardiac MRI

Authors: Hae-Yeoun Lee

Abstract:

Quantification of cardiac function is performed by calculating blood volume and ejection fraction in routine clinical practice. However, these works have been performed by manual contouring, which requires computational costs and varies on the observer. In this paper, an automatic left ventricle segmentation algorithm on cardiac magnetic resonance images (MRI) is presented. Using knowledge on cardiac MRI, a K-mean clustering technique is applied to segment blood region on a coil-sensitivity corrected image. Then, a graph searching technique is used to correct segmentation errors from coil distortion and noises. Finally, blood volume and ejection fraction are calculated. Using cardiac MRI from 15 subjects, the presented algorithm is tested and compared with manual contouring by experts to show outstanding performance.

Keywords: Cardiac MRI, Graph searching, Left ventricle segmentation, K-means clustering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2094
7114 Performance of Random Diagonal Codes for Spectral Amplitude Coding Optical CDMA Systems

Authors: Hilal A. Fadhil, Syed A. Aljunid, R. Badlishah Ahmed

Abstract:

In this paper we study the use of a new code called Random Diagonal (RD) code for Spectral Amplitude Coding (SAC) optical Code Division Multiple Access (CDMA) networks, using Fiber Bragg-Grating (FBG), FBG consists of a fiber segment whose index of reflection varies periodically along its length. RD code is constructed using code level and data level, one of the important properties of this code is that the cross correlation at data level is always zero, which means that Phase intensity Induced Phase (PIIN) is reduced. We find that the performance of the RD code will be better than Modified Frequency Hopping (MFH) and Hadamard code It has been observed through experimental and theoretical simulation that BER for RD code perform significantly better than other codes. Proof –of-principle simulations of encoding with 3 channels, and 10 Gbps data transmission have been successfully demonstrated together with FBG decoding scheme for canceling the code level from SAC-signal.

Keywords: FBG, MFH, OCDMA, PIIN, BER.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743
7113 Port Governance Model by International Freight Forwarders’ Point of View: A Study at Port of Santos - Brazil

Authors: Guilherme B. B. Vieira, Rafael M. da Silva, Eliana T. P. Senna, Luiz A. S. Senna, Francisco J. Kliemann Neto

Abstract:

Due to the importance of ports to trade and economic development of the regions in which they are inserted, in recent decades the number of studies devoted to this subject has increased. Part of these studies considers the ports as business agglomerations and focuses on port governance. This is an important approach since the port performance is the result of activities performed by actors belonging to the port-logistics chain, which need to be properly coordinated. This coordination takes place through a port governance model. Given this context, this study aims to analyze the governance model of the port of Santos from the perspective of port customers. To do this, a closed-ended questionnaire based on a conceptual model that considers the key dimensions associated with port governance was applied to the international freight forwarders that operate in the port. The results show the applicability of the considered model and highlight improvement opportunities to be implemented at the port of Santos.

Keywords: Port Governance, Model, Port of Santos, Customers’ Perception.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2168
7112 A Model for Estimation of Efforts in Development of Software Systems

Authors: Parvinder S. Sandhu, Manisha Prashar, Pourush Bassi, Atul Bisht

Abstract:

Software effort estimation is the process of predicting the most realistic use of effort required to develop or maintain software based on incomplete, uncertain and/or noisy input. Effort estimates may be used as input to project plans, iteration plans, budgets. There are various models like Halstead, Walston-Felix, Bailey-Basili, Doty and GA Based models which have already used to estimate the software effort for projects. In this study Statistical Models, Fuzzy-GA and Neuro-Fuzzy (NF) Inference Systems are experimented to estimate the software effort for projects. The performances of the developed models were tested on NASA software project datasets and results are compared with the Halstead, Walston-Felix, Bailey-Basili, Doty and Genetic Algorithm Based models mentioned in the literature. The result shows that the NF Model has the lowest MMRE and RMSE values. The NF Model shows the best results as compared with the Fuzzy-GA based hybrid Inference System and other existing Models that are being used for the Effort Prediction with lowest MMRE and RMSE values.

Keywords: Neuro-Fuzzy Model, Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model, GA Based Model, Genetic Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3227
7111 Elemental Graph Data Model: A Semantic and Topological Representation of Building Elements

Authors: Yasmeen A. S. Essawy, Khaled Nassar

Abstract:

With the rapid increase of complexity in the building industry, professionals in the A/E/C industry were forced to adopt Building Information Modeling (BIM) in order to enhance the communication between the different project stakeholders throughout the project life cycle and create a semantic object-oriented building model that can support geometric-topological analysis of building elements during design and construction. This paper presents a model that extracts topological relationships and geometrical properties of building elements from an existing fully designed BIM, and maps this information into a directed acyclic Elemental Graph Data Model (EGDM). The model incorporates BIM-based search algorithms for automatic deduction of geometrical data and topological relationships for each building element type. Using graph search algorithms, such as Depth First Search (DFS) and topological sortings, all possible construction sequences can be generated and compared against production and construction rules to generate an optimized construction sequence and its associated schedule. The model is implemented in a C# platform.

Keywords: Building information modeling, elemental graph data model, geometric and topological data models, and graph theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1203
7110 Analysis and Prototyping of Biological Systems: the Abstract Biological Process Model

Authors: Antonio Di Leva, Roberto Berchi, Gianpiero Pescarmona, Michele Sonnessa

Abstract:

The aim of a biological model is to understand the integrated structure and behavior of complex biological systems as a function of the underlying molecular networks to achieve simulation and forecast of their operation. Although several approaches have been introduced to take into account structural and environment related features, relatively little attention has been given to represent the behavior of biological systems. The Abstract Biological Process (ABP) model illustrated in this paper is an object-oriented model based on UML (the standard object-oriented language). Its main objective is to bring into focus the functional aspects of the biological system under analysis.

Keywords: Biological processes, system dynamics, systemmodeling, UML.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1640
7109 An Economical Operation Analysis Optimization Model for Heavy Equipment Selection

Authors: A. Jrade, N. Markiz, N. Albelwi

Abstract:

Optimizing equipment selection in heavy earthwork operations is a critical key in the success of any construction project. The objective of this research incentive was geared towards developing a computer model to assist contractors and construction managers in estimating the cost of heavy earthwork operations. Economical operation analysis was conducted for an equipment fleet taking into consideration the owning and operating costs involved in earthwork operations. The model is being developed in a Microsoft environment and is capable of being integrated with other estimating and optimization models. In this study, Caterpillar® Performance Handbook [5] was the main resource used to obtain specifications of selected equipment. The implementation of the model shall give optimum selection of equipment fleet not only based on cost effectiveness but also in terms of versatility. To validate the model, a case study of an actual dam construction project was selected to quantify its degree of accuracy.

Keywords: Operation analysis, optimization model, equipment economics, equipment selection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4258
7108 Definition of Foot Size Model using Kohonen Network

Authors: Khawla Ben Abderrahim

Abstract:

In order to define a new model of Tunisian foot sizes and for building the most comfortable shoes, Tunisian industrialists must be able to offer for their customers products able to put on and adjust the majority of the target population concerned. Moreover, the use of models of shoes, mainly from others country, causes a mismatch between the foot and comfort of the Tunisian shoes. But every foot is unique; these models become uncomfortable for the Tunisian foot. We have a set of measures produced from a 3D scan of the feet of a diverse population (women, men ...) and we try to analyze this data to define a model of foot specific to the Tunisian footwear design. In this paper we propose tow new approaches to modeling a new foot sizes model. We used, indeed, the neural networks, and specially the Kohonen network. Next, we combine neural networks with the concept of half-foot size to improve the models already found. Finally, it was necessary to compare the results obtained by applying each approach and we decide what-s the best approach that give us the most model of foot improving more comfortable shoes.

Keywords: Morphology of the foot, foot size, half foot size, neural network, Kohonen network, model of foot size.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1555
7107 A Model for the Characterization and Selection of Beeswaxes for use as base Substitute Tissue in Photon Teletherapy

Authors: R.M.V. Silva, D.N. Souza

Abstract:

This paper presents a model for the characterization and selection of beeswaxes for use as base substitute tissue for the manufacture of objects suitable for external radiotherapy using megavoltage photon beams. The model of characterization was divided into three distinct stages: 1) verification of aspects related to the origin of the beeswax, the bee species, the flora in the vicinity of the beehives and procedures to detect adulterations; 2) evaluation of physical and chemical properties; and 3) evaluation of beam attenuation capacity. The chemical composition of the beeswax evaluated in this study was similar to other simulators commonly used in radiotherapy. The behavior of the mass attenuation coefficient in the radiotherapy energy range was comparable to other simulators. The proposed model is efficient and enables convenient assessment of the use of any particular beeswax as a base substitute tissue for radiotherapy.

Keywords: Beeswaxes, characterization, model, radiotherapy

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1524
7106 A Numerical Model for Simulation of Blood Flow in Vascular Networks

Authors: Houman Tamaddon, Mehrdad Behnia, Masud Behnia

Abstract:

An accurate study of blood flow is associated with an accurate vascular pattern and geometrical properties of the organ of interest. Due to the complexity of vascular networks and poor accessibility in vivo, it is challenging to reconstruct the entire vasculature of any organ experimentally. The objective of this study is to introduce an innovative approach for the reconstruction of a full vascular tree from available morphometric data. Our method consists of implementing morphometric data on those parts of the vascular tree that are smaller than the resolution of medical imaging methods. This technique reconstructs the entire arterial tree down to the capillaries. Vessels greater than 2 mm are obtained from direct volume and surface analysis using contrast enhanced computed tomography (CT). Vessels smaller than 2mm are reconstructed from available morphometric and distensibility data and rearranged by applying Murray’s Laws. Implementation of morphometric data to reconstruct the branching pattern and applying Murray’s Laws to every vessel bifurcation simultaneously, lead to an accurate vascular tree reconstruction. The reconstruction algorithm generates full arterial tree topography down to the first capillary bifurcation. Geometry of each order of the vascular tree is generated separately to minimize the construction and simulation time. The node-to-node connectivity along with the diameter and length of every vessel segment is established and order numbers, according to the diameter-defined Strahler system, are assigned. During the simulation, we used the averaged flow rate for each order to predict the pressure drop and once the pressure drop is predicted, the flow rate is corrected to match the computed pressure drop for each vessel. The final results for 3 cardiac cycles is presented and compared to the clinical data.

Keywords: Blood flow, Morphometric data, Vascular tree, Strahler ordering system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2101
7105 Connected Vertex Cover in 2-Connected Planar Graph with Maximum Degree 4 is NP-complete

Authors: Priyadarsini P. L. K, Hemalatha T.

Abstract:

This paper proves that the problem of finding connected vertex cover in a 2-connected planar graph ( CVC-2 ) with maximum degree 4 is NP-complete. The motivation for proving this result is to give a shorter and simpler proof of NP-Completeness of TRA-MLC (the Top Right Access point Minimum-Length Corridor) problem [1], by finding the reduction from CVC-2. TRA-MLC has many applications in laying optical fibre cables for data communication and electrical wiring in floor plans.The problem of finding connected vertex cover in any planar graph ( CVC ) with maximum degree 4 is NP-complete [2]. We first show that CVC-2 belongs to NP and then we find a polynomial reduction from CVC to CVC-2. Let a graph G0 and an integer K form an instance of CVC, where G0 is a planar graph and K is an upper bound on the size of the connected vertex cover in G0. We construct a 2-connected planar graph, say G, by identifying the blocks and cut vertices of G0, and then finding the planar representation of all the blocks of G0, leading to a plane graph G1. We replace the cut vertices with cycles in such a way that the resultant graph G is a 2-connected planar graph with maximum degree 4. We consider L = K -2t+3 t i=1 di where t is the number of cut vertices in G1 and di is the number of blocks for which ith cut vertex is common. We prove that G will have a connected vertex cover with size less than or equal to L if and only if G0 has a connected vertex cover of size less than or equal to K.

Keywords: NP-complete, 2-Connected planar graph, block, cut vertex

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2003
7104 Improving the Quantification Model of Internal Control Impact on Banking Risks

Authors: M. Ndaw, G. Mendy, S. Ouya

Abstract:

Risk management in banking sector is a key issue linked to financial system stability and its importance has been elevated by technological developments and emergence of new financial instruments. In this paper, we improve the model previously defined for quantifying internal control impact on banking risks by automatizing the residual criticality estimation step of FMECA. For this, we defined three equations and a maturity coefficient to obtain a mathematical model which is tested on all banking processes and type of risks. The new model allows an optimal assessment of residual criticality and improves the correlation rate that has become 98%.

Keywords: Risk, Control, Banking, FMECA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1524