Search results for: Ridge estimate
489 Modeling and Simulation of Position Estimation of Switched Reluctance Motor with Artificial Neural Networks
Authors: Oguz Ustun, Erdal Bekiroglu
Abstract:
In the present study, position estimation of switched reluctance motor (SRM) has been achieved on the basis of the artificial neural networks (ANNs). The ANNs can estimate the rotor position without using an extra rotor position sensor by measuring the phase flux linkages and phase currents. Flux linkage-phase current-rotor position data set and supervised backpropagation learning algorithm are used in training of the ANN based position estimator. A 4-phase SRM have been used to verify the accuracy and feasibility of the proposed position estimator. Simulation results show that the proposed position estimator gives precise and accurate position estimations for both under the low and high level reference speeds of the SRM
Keywords: Artificial neural networks, modeling andsimulation, position observer, switched reluctance motor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2062488 An Improved Scheduling Strategy in Cloud Using Trust Based Mechanism
Authors: D. Sumathi, P. Poongodi
Abstract:
Cloud Computing refers to applications delivered as services over the internet, and the datacenters that provide those services with hardware and systems software. These were earlier referred to as Software as a Service (SaaS). Scheduling is justified by job components (called tasks), lack of information. In fact, in a large fraction of jobs from machine learning, bio-computing, and image processing domains, it is possible to estimate the maximum time required for a task in the job. This study focuses on Trust based scheduling to improve cloud security by modifying Heterogeneous Earliest Finish Time (HEFT) algorithm. It also proposes TR-HEFT (Trust Reputation HEFT) which is then compared to Dynamic Load Scheduling.Keywords: Software as a Service (SaaS), Trust, Heterogeneous Earliest Finish Time (HEFT) algorithm, Dynamic Load Scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2196487 Image Modeling Using Gibbs-Markov Random Field and Support Vector Machines Algorithm
Authors: Refaat M Mohamed, Ayman El-Baz, Aly A. Farag
Abstract:
This paper introduces a novel approach to estimate the clique potentials of Gibbs Markov random field (GMRF) models using the Support Vector Machines (SVM) algorithm and the Mean Field (MF) theory. The proposed approach is based on modeling the potential function associated with each clique shape of the GMRF model as a Gaussian-shaped kernel. In turn, the energy function of the GMRF will be in the form of a weighted sum of Gaussian kernels. This formulation of the GMRF model urges the use of the SVM with the Mean Field theory applied for its learning for estimating the energy function. The approach has been tested on synthetic texture images and is shown to provide satisfactory results in retrieving the synthesizing parameters.Keywords: Image Modeling, MRF, Parameters Estimation, SVM Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1636486 Bail-in Capital: The New Box
Authors: Manu Krishnan, Phil Jacoby
Abstract:
In this paper, we discuss the paradigm shift in bank capital from the “gone concern" to the “going concern" mindset. We then propose a methodology for pricing a product of this shift called Contingent Capital Notes (“CoCos"). The Merton Model can determine a price for credit risk by using the firm-s equity value as a call option on those assets. Our pricing methodology for CoCos also uses the credit spread implied by the Merton Model in a subsequent derivative form created by John Hull et al . Here, a market implied asset volatility is calculated by using observed market CDS spreads. This implied asset volatility is then used to estimate the probability of triggering a predetermined “contingency event" given the distanceto- trigger (DTT). The paper then investigates the effect of varying DTTs and recovery assumptions on the CoCo yield. We conclude with an investment rationale.Keywords: CoCo, Contingent capital, Bank Capital, Tier1 Capital
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544485 Evaluating Spectral Relationships between Signals by Removing the Contribution of a Common, Periodic Source A Partial Coherence-based Approach
Authors: Antonio Mauricio F. L. Miranda de Sá
Abstract:
Partial coherence between two signals removing the contribution of a periodic, deterministic signal is proposed for evaluating the interrelationship in multivariate systems. The estimator expression was derived and shown to be independent of such periodic signal. Simulations were used for obtaining its critical value, which were found to be the same as those for Gaussian signals, as well as for evaluating the technique. An Illustration with eletroencephalografic (EEG) signals during photic stimulation is also provided. The application of the proposed technique in both simulation and real EEG data indicate that it seems to be very specific in removing the contribution of periodic sources. The estimate independence of the periodic signal may widen partial coherence application to signal analysis, since it could be used together with simple coherence to test for contamination in signals by a common, periodic noise source.
Keywords: Partial coherence, periodic input, spectral analysis, statistical signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1464484 Parameter Estimation for Viewing Rank Distribution of Video-on-Demand
Authors: Hyoup-Sang Yoon
Abstract:
Video-on-demand (VOD) is designed by using content delivery networks (CDN) to minimize the overall operational cost and to maximize scalability. Estimation of the viewing pattern (i.e., the relationship between the number of viewings and the ranking of VOD contents) plays an important role in minimizing the total operational cost and maximizing the performance of the VOD systems. In this paper, we have analyzed a large body of commercial VOD viewing data and found that the viewing rank distribution fits well with the parabolic fractal distribution. The weighted linear model fitting function is used to estimate the parameters (coefficients) of the parabolic fractal distribution. This paper presents an analytical basis for designing an optimal hierarchical VOD contents distribution system in terms of its cost and performance.
Keywords: VOD, CDN, parabolic fractal distribution, viewing rank, weighted linear model fitting
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1790483 Validation of the WAsP Model for a Terrain Surrounded by Mountainous Region
Authors: Mohammadamin Zanganeh, Vahid Khalajzadeh
Abstract:
The problems associated with wind predictions of WAsP model in complex terrain are already the target of several studies in the last decade. In this paper, the influence of surrounding orography on accuracy of wind data analysis of a train is investigated. For the case study, a site with complex surrounding orography is considered. This site is located in Manjil, one of the windiest cities of Iran. For having precise evaluation of wind regime in the site, one-year wind data measurements from two metrological masts are used. To validate the obtained results from WAsP, the cross prediction between each mast is performed. The analysis reveals that WAsP model can estimate the wind speed behavior accurately. In addition, results show that this software can be used for predicting the wind regime in flat sites with complex surrounding orography.Keywords: Complex terrain, Meteorological mast, WAsPmodel, Wind prediction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1790482 A Survey on Quasi-Likelihood Estimation Approaches for Longitudinal Set-ups
Authors: Naushad Mamode Khan
Abstract:
The Com-Poisson (CMP) model is one of the most popular discrete generalized linear models (GLMS) that handles both equi-, over- and under-dispersed data. In longitudinal context, an integer-valued autoregressive (INAR(1)) process that incorporates covariate specification has been developed to model longitudinal CMP counts. However, the joint likelihood CMP function is difficult to specify and thus restricts the likelihood-based estimating methodology. The joint generalized quasi-likelihood approach (GQL-I) was instead considered but is rather computationally intensive and may not even estimate the regression effects due to a complex and frequently ill-conditioned covariance structure. This paper proposes a new GQL approach for estimating the regression parameters (GQL-III) that is based on a single score vector representation. The performance of GQL-III is compared with GQL-I and separate marginal GQLs (GQL-II) through some simulation experiments and is proved to yield equally efficient estimates as GQL-I and is far more computationally stable.
Keywords: Longitudinal, Com-Poisson, Ill-conditioned, INAR(1), GLMS, GQL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1776481 Comparison of the Parameter using ECG with Bisepctrum Parameter using EEG during General Anesthesia
Authors: Seong-wan Baik, Soo-young Ye, Byeong-cheol Choi, Gye-rok Jeon
Abstract:
The measurement of anesthetic depth is necessary in anesthesiology. NN10 is very simple method among the RR intervals analysis methods. NN10 parameter means the numbers of above the 10 ms intervals of the normal to normal RR intervals. Bispectrum analysis is defined as 2D FFT. EEG signal reflected the non-linear peristalsis phenomena according to the change brain function. After analyzing the bispectrum of the 2 dimension, the most significant power spectrum density peaks appeared abundantly at the specific area in awakening and anesthesia state. These points are utilized to create the new index since many peaks appeared at the specific area in the frequency coordinate. The measured range of an index was 0-100. An index is 20-50 at an anesthesia, while the index is 90-60 at the awake. In this paper, the relation between NN10 parameter using ECG and bisepctrum index using EEG is observed to estimate the depth of anesthesia during anesthesia and then we estimated the utility of the anesthetic.Keywords: Anesthesia, Bispectrum index, ECG, EEG
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573480 Aliasing Free and Additive Error in Spectra for Alpha Stable Signals
Authors: R. Sabre
Abstract:
This work focuses on the symmetric alpha stable process with continuous time frequently used in modeling the signal with indefinitely growing variance, often observed with an unknown additive error. The objective of this paper is to estimate this error from discrete observations of the signal. For that, we propose a method based on the smoothing of the observations via Jackson polynomial kernel and taking into account the width of the interval where the spectral density is non-zero. This technique allows avoiding the “Aliasing phenomenon” encountered when the estimation is made from the discrete observations of a process with continuous time. We have studied the convergence rate of the estimator and have shown that the convergence rate improves in the case where the spectral density is zero at the origin. Thus, we set up an estimator of the additive error that can be subtracted for approaching the original signal without error.
Keywords: Spectral density, stable processes, aliasing, p-adic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 585479 Criticality Assessment of Failures in Multipoint Communication Networks
Authors: Myriam Noureddine, Rachid Noureddine
Abstract:
Following the current economic challenges and competition, all systems, whatever their field, must be efficient and operational during their activity. In this context, it is imperative to anticipate, identify, eliminate and estimate the failures of systems, which may lead to an interruption of their function. This need requires the management of possible risks, through an assessment of the failures criticality following a dependability approach. On the other hand, at the time of new information technologies and considering the networks field evolution, the data transmission has evolved towards a multipoint communication, which can simultaneously transmit information from a sender to multiple receivers. This article proposes the failures criticality assessment of a multipoint communication network, integrates a database of network failures and their quantifications. The proposed approach is validated on a case study and the final result allows having the criticality matrix associated with failures on the considered network, giving the identification of acceptable risks.
Keywords: Dependability, failure, multipoint network, criticality matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1604478 Probabilistic Center Voting Method for Subsequent Object Tracking and Segmentation
Authors: Suryanto, Hyo-Kak Kim, Sang-Hee Park, Dae-Hwan Kim, Sung-Jea Ko
Abstract:
In this paper, we introduce a novel algorithm for object tracking in video sequence. In order to represent the object to be tracked, we propose a spatial color histogram model which encodes both the color distribution and spatial information. The object tracking from frame to frame is accomplished via center voting and back projection method. The center voting method has every pixel in the new frame to cast a vote on whereabouts the object center is. The back projection method segments the object from the background. The segmented foreground provides information on object size and orientation, omitting the need to estimate them separately. We do not put any assumption on camera motion; the proposed algorithm works equally well for object tracking in both static and moving camera videos.
Keywords: center voting, back projection, object tracking, size adaptation, non-stationary camera tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1667477 Real Time Video Based Smoke Detection Using Double Optical Flow Estimation
Authors: Anton Stadler, Thorsten Ike
Abstract:
In this paper, we present a video based smoke detection algorithm based on TVL1 optical flow estimation. The main part of the algorithm is an accumulating system for motion angles and upward motion speed of the flow field. We optimized the usage of TVL1 flow estimation for the detection of smoke with very low smoke density. Therefore, we use adapted flow parameters and estimate the flow field on difference images. We show in theory and in evaluation that this improves the performance of smoke detection significantly. We evaluate the smoke algorithm using videos with different smoke densities and different backgrounds. We show that smoke detection is very reliable in varying scenarios. Further we verify that our algorithm is very robust towards crowded scenes disturbance videos.Keywords: Low density, optical flow, upward smoke motion, video based smoke detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1419476 Performance Analysis of Cellular Wireless Network by Queuing Priority Handoff calls
Authors: Raj Kumar Samanta, Partha Bhattacharjee Gautam Sanyal
Abstract:
In this paper, a mathematical model is proposed to estimate the dropping probabilities of cellular wireless networks by queuing handoff instead of reserving guard channels. Usually, prioritized handling of handoff calls is done with the help of guard channel reservation. To evaluate the proposed model, gamma inter-arrival and general service time distributions have been considered. Prevention of some of the attempted calls from reaching to the switching center due to electromagnetic propagation failure or whimsical user behaviour (missed call, prepaid balance etc.), make the inter-arrival time of the input traffic to follow gamma distribution. The performance is evaluated and compared with that of guard channel scheme.Keywords: Cellular wireless networks, non-classical traffic, mathematicalmodel, guard channel, queuing, handoff.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2393475 Numerical Simulation of Lightning Strike Direct Effects on Aircraft Skin Composite Laminate
Authors: Muhammad Khalil, Nader Abuelfoutouh, Gasser Abdelal, Adrian Murphy
Abstract:
Nowadays, the direct effects of lightning to aircrafts are of great importance because of the massive use of composite materials. In comparison with metallic materials, composites present several weaknesses for lightning strike direct effects. Especially, their low electrical and thermal conductivities lead to severe lightning strike damage. The lightning strike direct effects are burning, heating, magnetic force, sparking and arcing. As the problem is complex, we investigated it gradually. A magnetohydrodynamics (MHD) model is developed to simulate the lightning strikes in order to estimate the damages on the composite materials. Then, a coupled thermal-electrical finite element analysis is used to study the interaction between the lightning arc and the composite laminate and to investigate the material degradation.
Keywords: Composite structures, lightning multiphysics, magnetohydrodynamics, coupled thermal-electrical analysis, thermal plasmas.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2623474 On Bayesian Analysis of Failure Rate under Topp Leone Distribution using Complete and Censored Samples
Abstract:
The article is concerned with analysis of failure rate (shape parameter) under the Topp Leone distribution using a Bayesian framework. Different loss functions and a couple of noninformative priors have been assumed for posterior estimation. The posterior predictive distributions have also been derived. A simulation study has been carried to compare the performance of different estimators. A real life example has been used to illustrate the applicability of the results obtained. The findings of the study suggest that the precautionary loss function based on Jeffreys prior and singly type II censored samples can effectively be employed to obtain the Bayes estimate of the failure rate under Topp Leone distribution.
Keywords: loss functions, type II censoring, posterior distribution, Bayes estimators.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2560473 A Dynamic Hybrid Option Pricing Model by Genetic Algorithm and Black- Scholes Model
Authors: Yi-Chang Chen, Shan-Lin Chang, Chia-Chun Wu
Abstract:
Unlike this study focused extensively on trading behavior of option market, those researches were just taken their attention to model-driven option pricing. For example, Black-Scholes (B-S) model is one of the most famous option pricing models. However, the arguments of B-S model are previously mentioned by some pricing models reviewing. This paper following suggests the importance of the dynamic character for option pricing, which is also the reason why using the genetic algorithm (GA). Because of its natural selection and species evolution, this study proposed a hybrid model, the Genetic-BS model which combining GA and B-S to estimate the price more accurate. As for the final experiments, the result shows that the output estimated price with lower MAE value than the calculated price by either B-S model or its enhanced one, Gram-Charlier garch (G-C garch) model. Finally, this work would conclude that the Genetic-BS pricing model is exactly practical.Keywords: genetic algorithm, Genetic-BS, option pricing model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2245472 The Use of Degradation Measures to Design Reliability Test Plans
Authors: Stephen V. Crowder, Jonathan W. Lane
Abstract:
With short production development times, there is an increased need to demonstrate product reliability relatively quickly with minimal testing. In such cases there may be few if any observed failures. Thus it may be difficult to assess reliability using the traditional reliability test plans that measure only time (or cycles) to failure. For many components, degradation measures will contain important information about performance and reliability. These measures can be used to design a minimal test plan, in terms of number of units placed on test and duration of the test, necessary to demonstrate a reliability goal. In this work we present a case study involving an electronic component subject to degradation. The data, consisting of 42 degradation paths of cycles to failure, are first used to estimate a reliability function. Bootstrapping techniques are then used to perform power studies and develop a minimal reliability test plan for future production of this component.
Keywords: Degradation Measure, Time to Failure Distribution, Bootstrap.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1878471 Model-Based Small Area Estimation with Application to Unemployment Estimates
Authors: Hichem Omrani, Philippe Gerber, Patrick Bousch
Abstract:
The problem of Small Area Estimation (SAE) is complex because of various information sources and insufficient data. In this paper, an approach for SAE is presented for decision-making at national, regional and local level. We propose an Empirical Best Linear Unbiased Predictor (EBLUP) as an estimator in order to combine several information sources to evaluate various indicators. First, we present the urban audit project and its environmental, social and economic indicators. Secondly, we propose an approach for decision making in order to estimate indicators. An application is used to validate the theoretical proposal. Finally, a decision support system is presented based on open-source environment.
Keywords: Small area estimation, statistical method, sampling, empirical best linear unbiased predictor (EBLUP), decision-making.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1712470 Nonlinear Modeling of the PEMFC Based On NNARX Approach
Authors: Shan-Jen Cheng, Te-Jen Chang, Kuang-Hsiung Tan, Shou-Ling Kuo
Abstract:
Polymer Electrolyte Membrane Fuel Cell (PEMFC) is such a time-vary nonlinear dynamic system. The traditional linear modeling approach is hard to estimate structure correctly of PEMFC system. From this reason, this paper presents a nonlinear modeling of the PEMFC using Neural Network Auto-regressive model with eXogenous inputs (NNARX) approach. The multilayer perception (MLP) network is applied to evaluate the structure of the NNARX model of PEMFC. The validity and accuracy of NNARX model are tested by one step ahead relating output voltage to input current from measured experimental of PEMFC. The results show that the obtained nonlinear NNARX model can efficiently approximate the dynamic mode of the PEMFC and model output and system measured output consistently.Keywords: PEMFC, neural network, nonlinear identification, NNARX.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2200469 Analyzing the Factors Effecting the Passenger Car Breakdowns using Com-Poisson GLM
Authors: N. Mamode Khan, V. Jowaheer
Abstract:
Number of breakdowns experienced by a machinery is a highly under-dispersed count random variable and its value can be attributed to the factors related to the mechanical input and output of that machinery. Analyzing such under-dispersed count observations as a function of the explanatory factors has been a challenging problem. In this paper, we aim at estimating the effects of various factors on the number of breakdowns experienced by a passenger car based on a study performed in Mauritius over a year. We remark that the number of passenger car breakdowns is highly under-dispersed. These data are therefore modelled and analyzed using Com-Poisson regression model. We use quasi-likelihood estimation approach to estimate the parameters of the model. Under-dispersion parameter is estimated to be 2.14 justifying the appropriateness of Com-Poisson distribution in modelling under-dispersed count responses recorded in this study.
Keywords: Breakdowns, under-dispersion, com-poisson, generalized linear model, quasi-likelihood estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544468 Deterministic Method to Assess Kalman Filter Passive Ranging Solution Reliability
Authors: Ronald M. Yannone
Abstract:
For decades, the defense business has been plagued by not having a reliable, deterministic method to know when the Kalman filter solution for passive ranging application is reliable for use by the fighter pilot. This has made it hard to accurately assess when the ranging solution can be used for situation awareness and weapons use. To date, we have used ad hoc rules-of-thumb to assess when we think the estimate of the Kalman filter standard deviation on range is reliable. A reliable algorithm has been developed at BAE Systems Electronics & Integrated Solutions that monitors the Kalman gain matrix elements – and a patent is pending. The “settling" of the gain matrix elements relates directly to when we can assess the time when the passive ranging solution is within the 10 percent-of-truth value. The focus of the paper is on surface-based passive ranging – but the method is applicable to airborne targets as well.Keywords: Electronic warfare, extended Kalman filter (EKF), fighter aircraft, passive ranging, track convergence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2064467 The Role of Heat Pumps for the Decarbonization of European Regions
Authors: D. M. Mongelli, M. De Carli, L. Carnieletto, F. Busato
Abstract:
This research aims to provide a contribution to the reduction of fossil fuels and the consequent reduction of CO2eq emissions for each European region. Simulations have been carried out to replace fossil fuel fired heating boilers with air-to-water heat pumps, when allowed by favorable environmental conditions (outdoor temperature, water temperature in emission systems, etc.). To estimate the potential coverage of high-temperature heat pumps in European regions, the energy profiles of buildings were considered together with the potential coefficient of performance (COP) of heat pumps operating with flow temperature with variable climatic regulation. The electrification potential for heating buildings was estimated by dividing the 38 European countries examined into 179 territorial units. The yields have been calculated in terms of energy savings and CO2eq reduction.
Keywords: Decarbonization, Space heating, Heat pumps, Energy policies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 212466 Inter-frame Collusion Attack in SS-N Video Watermarking System
Authors: Yaser Mohammad Taheri, Alireza Zolghadr–asli, Mehran Yazdi
Abstract:
Video watermarking is usually considered as watermarking of a set of still images. In frame-by-frame watermarking approach, each video frame is seen as a single watermarked image, so collusion attack is more critical in video watermarking. If the same or redundant watermark is used for embedding in every frame of video, the watermark can be estimated and then removed by watermark estimate remodolulation (WER) attack. Also if uncorrelated watermarks are used for every frame, these watermarks can be washed out with frame temporal filtering (FTF). Switching watermark system or so-called SS-N system has better performance against WER and FTF attacks. In this system, for each frame, the watermark is randomly picked up from a finite pool of watermark patterns. At first SS-N system will be surveyed and then a new collusion attack for SS-N system will be proposed using a new algorithm for separating video frame based on watermark pattern. So N sets will be built in which every set contains frames carrying the same watermark. After that, using WER attack in every set, N different watermark patterns will be estimated and removed later.
Keywords: Watermark estimation remodulation (WER), Frame Temporal Averaging (FTF), switching watermark system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1497465 Carbon Storage in Above-Ground Biomass of Tropical Deciduous Forest in Ratchaburi Province, Thailand
Authors: Ubonwan Chaiyo, Savitri Garivait, Kobsak Wanthongchai
Abstract:
The study site was located in Ratchaburi Province, Thailand. Four experimental plots in dry dipterocarp forest (DDF) and four plots in mixed deciduous forest (MDF) were set up to estimate the above-ground biomass of tree, sapling and bamboo. The allometry equations were used to investigate above-ground biomass of these vegetation. Seedling and other understory were determined using direct harvesting method. Carbon storage in above-ground biomass was calculated based on IPCC 2006. The results showed that the above-ground biomass of DDF at 20-40% slope, <20% slope and MDF at <20% slope were 91.96, 30.95 and 59.44 ton/ha, respectively. Bamboo covers about half of total aboveground biomass in MDF, which is a specific characteristic of this area. The carbon sequestration potential in above-ground biomass of plot slope range 20-40% DDF, <20% DDF and <20% MDF are 43.22, 14.55 and 27.94 ton C/ha, respectively.Keywords: Carbon storage, aboveground biomass, tropical deciduous forest, dry dipterocarp forest, mixed deciduous forest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2967464 Comparing Spontaneous Hydrolysis Rates of Activated Models of DNA and RNA
Authors: Mohamed S. Sasi, Adel M. Mlitan, Abdulfattah M. Alkherraz
Abstract:
This research project aims to investigate difference in relative rates concerning phosphoryl transfer relevant to biological catalysis of DNA and RNA in the pH-independent reactions. Activated Models of DNA and RNA for alkyl-aryl phosphate diesters (with 4-nitrophenyl as a good leaving group) have successfully been prepared to gather kinetic parameters. Eyring plots for the pH– independent hydrolysis of 1 and 2 were established at different temperatures in the range 100–160 °C. These measurements have been used to provide a better estimate for the difference in relative rates between the reactivity of DNA and RNA cleavage. Eyring plot gave an extrapolated rate of kH2O = 1 × 10-10 s -1 for 1 (RNA model) and 2 (DNA model) at 25°C. Comparing the reactivity of RNA model and DNA model shows that the difference in relative rates in the pH-independent reactions is surprisingly very similar at 25°. This allows us to obtain chemical insights into how biological catalysts such as enzymes may have evolved to perform their current functions.
Keywords: DNA & RNA Models, Relative Rates, Reactivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2396463 A Strategic Evaluation Approach for Defining the Maturity of Manufacturing Technologies
Authors: G. Reinhart, S. Schindler
Abstract:
Due to dynamic evolution, the ability of a manufacturing technology to produce a special product is changing. Therefore, it is essential to monitor the established techniques and processes to detect whether a company-s production will fit future circumstances. Concerning the manufacturing technology planning process, companies must decide when to change to a new technology for maintaining and increasing competitive advantages. In this context, the maturity assessment of the focused technologies is crucial. This article presents an approach for defining the maturity of a manufacturing technology from a strategic point of view. The concept is based on the approach of technology readiness level (TRL) according to NASA (National Aeronautics and Space Administration), but also includes dynamic changes. Therefore, the model takes into account the concept of the technology life cycle. Furthermore, it enables a company to estimate the ideal date for implementation of a new manufacturing technology.Keywords: Maturity Assessment, Manufacturing Technology Planning, Technology Life Cycle, Technology Readiness Level.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2492462 Static Response of Homogeneous Clay Stratum to Imposed Structural Loads
Authors: Aaron Aboshio
Abstract:
Numerical study of the static response of homogeneous clay stratum considering a wide range of cohesion and subject to foundation loads is presented. The linear elastic–perfectly plastic constitutive relation with the von Mises yield criterion were utilised to develop a numerically cost effective finite element model for the soil while imposing a rigid body constrain to the foundation footing. From the analyses carried out, estimate of the bearing capacity factor, Nc as well as the ultimate load-carrying capacities of these soils, effect of cohesion on foundation settlements, stress fields and failure propagation were obtained. These are consistent with other findings in the literature and hence can be a useful guide in design of safe foundations in clay soils for buildings and other structure.Keywords: Bearing capacity factors, finite element method, safe bearing pressure, structure-soil interaction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1999461 An Adaptive Fuzzy Clustering Approach for the Network Management
Authors: Amal Elmzabi, Mostafa Bellafkih, Mohammed Ramdani
Abstract:
The Chiu-s method which generates a Takagi-Sugeno Fuzzy Inference System (FIS) is a method of fuzzy rules extraction. The rules output is a linear function of inputs. In addition, these rules are not explicit for the expert. In this paper, we develop a method which generates Mamdani FIS, where the rules output is fuzzy. The method proceeds in two steps: first, it uses the subtractive clustering principle to estimate both the number of clusters and the initial locations of a cluster centers. Each obtained cluster corresponds to a Mamdani fuzzy rule. Then, it optimizes the fuzzy model parameters by applying a genetic algorithm. This method is illustrated on a traffic network management application. We suggest also a Mamdani fuzzy rules generation method, where the expert wants to classify the output variables in some fuzzy predefined classes.
Keywords: Fuzzy entropy, fuzzy inference systems, genetic algorithms, network management, subtractive clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1883460 Operating Room Capacity Planning Decisions
Authors: Abdulrahim Shamayleh, John Fowler, Muhong Zhang
Abstract:
Operating rooms are important assets for hospitals as they generate the largest revenue and, at the same time, produce the largest cost for hospitals. The model presented in this paper helps make capacity planning decisions on the combination of open operating rooms (ORs) and estimated overtime to satisfy the allocated OR time to each specialty. The model combines both decisions on determining the amount of OR time to open and to allocate to different surgical specialties. The decisions made are based on OR costs, overutilization and underutilization costs, and contribution margins from allocating OR time. The results show the importance of having a good estimate of specialty usage of OR time to determine the amount of needed capacity and highlighted the tradeoff that the OR manager faces between opening more ORs versus extending the working time of the ORs already in use.Keywords: capacity planning, contribution margins, operating room, overutilization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2723