Search results for: Conditional probability
276 Towards a Systematic, Cost-Effective Approach for ERP Selection
Authors: Hassan Haghighi, Omid Mafi
Abstract:
Existing experiences indicate that one of the most prominent reasons that some ERP implementations fail is related to selecting an improper ERP package. Among those important factors resulting in inappropriate ERP selections, one is to ignore preliminary activities that should be done before the evaluation of ERP packages. Another factor yielding these unsuitable selections is that usually organizations employ prolonged and costly selection processes in such extent that sometimes the process would never be finalized or sometimes the evaluation team might perform many key final activities in an incomplete or inaccurate way due to exhaustion, lack of interest or out-of-date data. In this paper, a systematic approach that recommends some activities to be done before and after the main selection phase is introduced for choosing an ERP package. On the other hand, the proposed approach has utilized some ideas that accelerates the selection process at the same time that reduces the probability of an erroneous final selection.Keywords: enterprise resource planning, evaluation and selectionof ERP packages, organizational readiness for employing ERP, evaluationlists.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1550275 Novel Schemes of Pilot-Aided Integer Frequency Offset Estimation for OFDM-Based DVB-T Systems
Authors: Youngyoon Lee, Dahae Chong, Myungsoo Lee, Seongho Chae, Seokho Yoon
Abstract:
This paper proposes two novel schemes for pilot-aided integer frequency offset (IFO) estimation in orthogonal frequency division multiplexing (OFDM)-based digital video broadcastingterrestrial (DVB-T) systems. The conventional scheme proposed for estimating the IFO uses only partial information of combinations that pilots can provide, which stems from a rigorous assumption that the channel responses of pilots used for estimating the IFO change very rapidly. Thus, in this paper, we propose the novel IFO estimation schemes exploiting all information of combinations that pilots can provide to improve the performance of IFO estimation. The simulation results show that the proposed schemes are highly accurate in terms of the IFO detection probability.Keywords: OFDM, DVB-T, pilot, IFO, estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793274 An Elin Load Tap Changer Diagnosis by DGA
Authors: Hoda Molavi, Alireza Zahiri, Katayoon Anvarizadeh
Abstract:
Dissolved gas analysis has been accepted as a sensitive, informative and reliable technique for incipient faults detection in power transformers and is widely used. In the last few years this method, which has been recommended by IEEE Power & Energy society, has been applied for fault detection in load tap changers. Regarding the critical role of load tap changers in electrical network and essential of catastrophic failures prevention, it is necessary to choose "condition based preventative maintenance strategy" which leads to reduction in costs, the number of unnecessary visits as well as the probability of interruptions and also increment in equipment reliability. In current work, considering the condition based preventative maintenance strategy, condition assessment of an Elin tap changer was carried out using dissolved gas analysis.
Keywords: Condition Assessment, Dissolved Gas Analysis, Load Tap Changer
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3716273 Secondary Organic Contribution to Particles Formed on the Ice Melted Arctic Ocean
Authors: Petri Vaattovaara, Zoran D. Ristovski, Martin Graus, Marcus Müller, EijaAsmi, Luca Di Liberto, StaffanSjögren, Douglas Orsini, Caroline Leck, Ari Laaksonen
Abstract:
Due to climate warming and consequently due to ice and snow melting of the Arctic Ocean, the highly biologically active ocean surface area has been expanding quickly making possible longer marine biota growth seasons during polar summers. That increase the probability of the remote marine environment secondary contribution, especially secondary organic contribution, to the particle production and particle growth events and particle properties, consequently effecting on the open ocean, pack ice and ground based regions radiation budget and thus on the feedbacks between arctic biota, particles, clouds, and climate.
Keywords: Arctic Ocean, ice melting, nucleation, secondary organics, clouds, climate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492272 Development of Risk Assessment and Occupational Safety Management Model for Building Construction Projects
Authors: Preeda Sansakorn, Min An
Abstract:
In order to be capable of dealing with uncertainties, subjectivities, including vagueness arising in building construction projects, the application of fuzzy reasoning technique based on fuzzy set theory is proposed. This study contributes significantly to the development of a fuzzy reasoning safety risk assessment model for building construction projects that could be employed to assess the risk magnitude of each hazardous event identified during construction, and a third parameter of probability of consequence is incorporated in the model. By using the proposed safety risk analysis methodology, more reliable and less ambiguities, which provide the safety risk management project team for decision-making purposes.
Keywords: Safety risks assessment, building construction safety, fuzzy reasoning, construction risk assessment model, building construction projects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2343271 Seismic Fragility for Sliding Failure of Weir Structure Considering the Process of Concrete Aging
Authors: HoYoung Son, Ki Young Kim, Woo Young Jung
Abstract:
This study investigated the change of weir structure performances when durability of concrete, which is the main material of weir structure, decreased due to their aging by mean of seismic fragility analysis. In the analysis, it was assumed that the elastic modulus of concrete was reduced by 10% in order to account for their aged deterioration. Additionally, the analysis of seismic fragility was based on Monte Carlo Simulation method combined with a 2D nonlinear finite element in ABAQUS platform with the consideration of deterioration of concrete. Finally, the comparison of seismic fragility of model pre- and post-deterioration was made to study the performance of weir. Results show that the probability of failure in moderate damage for deteriorated model was found to be larger than pre-deterioration model when peak ground acceleration (PGA) passed 0.4 g.Keywords: Weir, FEM, concrete, fragility, aging.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 919270 The Temperature Range in the Simulation of Residual Stress and Hot Tearing During Investment Casting
Authors: Saeid Norouzi, Ali Shams, Hassan Farhangi, Alireza Darvish
Abstract:
Hot tear cracking and residual stress are two different consequences of thermal stress both of which can be considered as casting problem. The purpose of the present study is simulation of the effect of casting shape characteristic on hot tearing and residual stress. This study shows that the temperature range for simulation of hot tearing and residual stress are different. In this study, in order to study the development of thermal stress and to predict the hot tearing and residual stress of shaped casting, MAGMASOFT simulation program was used. The strategy of this research was the prediction of hot tear location using pinpointing hot spot and thermal stress concentration zones. The results shows that existing of stress concentration zone increases the hot tearing probability and consequently reduces the amount of remaining residual stress in casting parts.
Keywords: Hot tearing, residual stress, simulation, investment casting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2724269 Hopfield Network as Associative Memory with Multiple Reference Points
Authors: Domingo López-Rodríguez, Enrique Mérida-Casermeiro, Juan M. Ortiz-de-Lazcano-Lobato
Abstract:
Hopfield model of associative memory is studied in this work. In particular, two main problems that it possesses: the apparition of spurious patterns in the learning phase, implying the well-known effect of storing the opposite pattern, and the problem of its reduced capacity, meaning that it is not possible to store a great amount of patterns without increasing the error probability in the retrieving phase. In this paper, a method to avoid spurious patterns is presented and studied, and an explanation of the previously mentioned effect is given. Another technique to increase the capacity of a network is proposed here, based on the idea of using several reference points when storing patterns. It is studied in depth, and an explicit formula for the capacity of the network with this technique is provided.
Keywords: Associative memory, Hopfield network, network capacity, spurious patterns.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1108268 Text Summarization for Oil and Gas Drilling Topic
Authors: Y. Y. Chen, O. M. Foong, S. P. Yong, Kurniawan Iwan
Abstract:
Information sharing and gathering are important in the rapid advancement era of technology. The existence of WWW has caused rapid growth of information explosion. Readers are overloaded with too many lengthy text documents in which they are more interested in shorter versions. Oil and gas industry could not escape from this predicament. In this paper, we develop an Automated Text Summarization System known as AutoTextSumm to extract the salient points of oil and gas drilling articles by incorporating statistical approach, keywords identification, synonym words and sentence-s position. In this study, we have conducted interviews with Petroleum Engineering experts and English Language experts to identify the list of most commonly used keywords in the oil and gas drilling domain. The system performance of AutoTextSumm is evaluated using the formulae of precision, recall and F-score. Based on the experimental results, AutoTextSumm has produced satisfactory performance with F-score of 0.81.
Keywords: Keyword's probability, synonym sets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1731267 Estimating Regression Parameters in Linear Regression Model with a Censored Response Variable
Authors: Jesus Orbe, Vicente Nunez-Anton
Abstract:
In this work we study the effect of several covariates X on a censored response variable T with unknown probability distribution. In this context, most of the studies in the literature can be located in two possible general classes of regression models: models that study the effect the covariates have on the hazard function; and models that study the effect the covariates have on the censored response variable. Proposals in this paper are in the second class of models and, more specifically, on least squares based model approach. Thus, using the bootstrap estimate of the bias, we try to improve the estimation of the regression parameters by reducing their bias, for small sample sizes. Simulation results presented in the paper show that, for reasonable sample sizes and censoring levels, the bias is always smaller for the new proposals.
Keywords: Censored response variable, regression, bias.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1475266 Systemic Approach to Risk Measurement of Drainage Systems in Urban Areas
Authors: Jadwiga Królikowska, Andrzej Królikowski, Jarosław Bajer
Abstract:
The work delineates the threats of maladjustment of the capacity of rain canals, designed and built in the early 20th century, in connection to heavy rainfall, especially in summer. This is the cause of the so called 'urban floods.' It directly relates to fierce raise of paving in the cities. Resolving this problem requires a change in philosophy of draining the rainfall by wider use of retention, infiltration and usage of rainwater.
In systemic approach to managing the safety of urban drainage systems the risk, which is directly connected to safety failures, has been accepted as a measure. The risk level defines the probability of occurrence of losses greater than the ones forecast for a given time frame. The procedure of risk modelling, enabling its numeric analysis by using appropriate weights, is a significant issue in this paper.
Keywords: Drainage system, urban areas, risk measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1694265 A Method for Modeling Multiple Antenna Channels
Authors: S. Rajabi, M. ArdebiliPoor, M. Shahabadi
Abstract:
In this paper we propose a method for modeling the correlation between the received signals by two or more antennas operating in a multipath environment. Considering the maximum excess delay in the channel being modeled, an elliptical region surrounding both transmitter and receiver antennas is produced. A number of scatterers are randomly distributed in this region and scatter the incoming waves. The amplitude and phase of incoming waves are computed and used to obtain statistical properties of the received signals. This model has the distinguishable advantage of being applicable for any configuration of antennas. Furthermore the common PDF (Probability Distribution Function) of received wave amplitudes for any pair of antennas can be calculated and used to produce statistical parameters of received signals.Keywords: MIMO (Multiple Input Multiple Output), SIMO (Single Input Multiple Output), GBSBEM (Geometrically Based Single Bounce Elliptical Model).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1422264 Human Body Configuration using Bayesian Model
Authors: Rui. Zhang, Yiming. Pi
Abstract:
In this paper we present a novel approach for human Body configuration based on the Silhouette. We propose to address this problem under the Bayesian framework. We use an effective Model based MCMC (Markov Chain Monte Carlo) method to solve the configuration problem, in which the best configuration could be defined as MAP (maximize a posteriori probability) in Bayesian model. This model based MCMC utilizes the human body model to drive the MCMC sampling from the solution space. It converses the original high dimension space into a restricted sub-space constructed by the human model and uses a hybrid sampling algorithm. We choose an explicit human model and carefully select the likelihood functions to represent the best configuration solution. The experiments show that this method could get an accurate configuration and timesaving for different human from multi-views.Keywords: Bayesian framework, MCMC, model based, human body configuration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1316263 Performance Analysis of M-Ary Pulse Position Modulation in Multihop Multiple Input Multiple Output-Free Space Optical System over Uncorrelated Gamma-Gamma Atmospheric Turbulence Channels
Authors: Hechmi Saidi, Noureddine Hamdi
Abstract:
The performance of Decode and Forward (DF) multihop Free Space Optical ( FSO) scheme deploying Multiple Input Multiple Output (MIMO) configuration under Gamma-Gamma (GG) statistical distribution, that adopts M-ary Pulse Position Modulation (MPPM) coding, is investigated. We have extracted exact and estimated values of Symbol-Error Rates (SERs) respectively. A closed form formula related to the Probability Density Function (PDF) is expressed for our designed system. Thanks to the use of DF multihop MIMO FSO configuration and MPPM signaling, atmospheric turbulence is combatted; hence the transmitted signal quality is improved.Keywords: FSO, MIMO, MIMO, multihop, DF, SER, GG channel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 645262 Parametric and Nonparametric Analysis of Breast Cancer Treatments
Authors: Chunling Cong, Chris.P.Tsokos
Abstract:
The objective of the present research manuscript is to perform parametric, nonparametric, and decision tree analysis to evaluate two treatments that are being used for breast cancer patients. Our study is based on utilizing real data which was initially used in “Tamoxifen with or without breast irradiation in women of 50 years of age or older with early breast cancer" [1], and the data is supplied to us by N.A. Ibrahim “Decision tree for competing risks survival probability in breast cancer study" [2]. We agree upon certain aspects of our findings with the published results. However, in this manuscript, we focus on relapse time of breast cancer patients instead of survival time and parametric analysis instead of semi-parametric decision tree analysis is applied to provide more precise recommendations of effectiveness of the two treatments with respect to reoccurrence of breast cancer.Keywords: decision tree, breast cancer treatments, parametricanalysis, non-parametric analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2052261 On Identity Disclosure Risk Measurement for Shared Microdata
Authors: M. N. Huda, S. Yamada, N. Sonehara
Abstract:
Probability-based identity disclosure risk measurement may give the same overall risk for different anonymization strategy of the same dataset. Some entities in the anonymous dataset may have higher identification risks than the others. Individuals are more concerned about higher risks than the average and are more interested to know if they have a possibility of being under higher risk. A notation of overall risk in the above measurement method doesn-t indicate whether some of the involved entities have higher identity disclosure risk than the others. In this paper, we have introduced an identity disclosure risk measurement method that not only implies overall risk, but also indicates whether some of the members have higher risk than the others. The proposed method quantifies the overall risk based on the individual risk values, the percentage of the records that have a risk value higher than the average and how larger the higher risk values are compared to the average. We have analyzed the disclosure risks for different disclosure control techniques applied to original microdata and present the results.Keywords: Anonymization, microdata, disclosure risk, privacy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1365260 Experimenting the Influence of Input Modality on Involvement Load Hypothesis
Authors: Mohammad Hassanzadeh
Abstract:
As far as incidental vocabulary learning is concerned, the basic contention of the Involvement Load Hypothesis (ILH) is that retention of unfamiliar words is, generally, conditional upon the degree of involvement in processing them. This study examined input modality and incidental vocabulary uptake in a task-induced setting whereby three variously loaded task types (marginal glosses, fill-in-task, and sentence-writing) were alternately assigned to one group of students at Allameh Tabataba’i University (n=2l) during six classroom sessions. While one round of exposure was comprised of the audiovisual medium (TV talk shows), the second round consisted of textual materials with approximately similar subject matter (reading texts). In both conditions, however, the tasks were equivalent to one another. Taken together, the study pursued the dual objectives of establishing a litmus test for the ILH and its proposed values of ‘need’, ‘search’ and ‘evaluation’ in the first place. Secondly, it sought to bring to light the superiority issue of exposure to audiovisual input versus the written input as far as the incorporation of tasks is concerned. At the end of each treatment session, a vocabulary active recall test was administered to measure their incidental gains. Running a one-way analysis of variance revealed that the audiovisual intervention yielded higher gains than the written version even when differing tasks were included. Meanwhile, task 'three' (sentence-writing) turned out the most efficient in tapping learners' active recall of the target vocabulary items. In addition to shedding light on the superiority of audiovisual input over the written input when circumstances are relatively held constant, this study for the most part, did support the underlying tenets of ILH.
Keywords: Evaluation, incidental vocabulary learning, input mode, involvement load hypothesis, need, search.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1153259 Optimal Location of the I/O Point in the Parking System
Authors: Jing Zhang, Jie Chen
Abstract:
In this paper, we deal with the optimal I/O point location in an automated parking system. In this system, the S/R machine (storage and retrieve machine) travels independently in vertical and horizontal directions. Based on the characteristics of the parking system and the basic principle of AS/RS system (Automated Storage and Retrieval System), we obtain the continuous model in units of time. For the single command cycle using the randomized storage policy, we calculate the probability density function for the system travel time and thus we develop the travel time model. And we confirm that the travel time model shows a good performance by comparing with discrete case. Finally in this part, we establish the optimal model by minimizing the expected travel time model and it is shown that the optimal location of the I/O point is located at the middle of the left-hand above corner.
Keywords: Parking system, optimal location, response time, S/R machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 677258 Bail-in Capital: The New Box
Authors: Manu Krishnan, Phil Jacoby
Abstract:
In this paper, we discuss the paradigm shift in bank capital from the “gone concern" to the “going concern" mindset. We then propose a methodology for pricing a product of this shift called Contingent Capital Notes (“CoCos"). The Merton Model can determine a price for credit risk by using the firm-s equity value as a call option on those assets. Our pricing methodology for CoCos also uses the credit spread implied by the Merton Model in a subsequent derivative form created by John Hull et al . Here, a market implied asset volatility is calculated by using observed market CDS spreads. This implied asset volatility is then used to estimate the probability of triggering a predetermined “contingency event" given the distanceto- trigger (DTT). The paper then investigates the effect of varying DTTs and recovery assumptions on the CoCo yield. We conclude with an investment rationale.Keywords: CoCo, Contingent capital, Bank Capital, Tier1 Capital
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544257 The Announcer Trainee Satisfaction by National Broadcasting and Telecommunications Commission of Thailand
Authors: Nareenad Panbun
Abstract:
The objective is to study the knowledge utilization from the participants of the announcer training program by National Broadcasting and Telecommunications Commission (NBTC). This study is a quantitative research based on surveys and self-answering questionnaires. The population of this study is 100 participants randomly chosen by non-probability sampling method. The results have shown that most of the participants were satisfied with the topics of general knowledge about the broadcasting and television business for 37 people representing 37%, followed by the topics of broadcasting techniques. The legal issues, consumer rights, television business ethics, and credibility of the media are, in addition to the media's role and responsibilities in society, the use of language for successful communication. Therefore, the communication language skills are the most important for all of the trainees and will also build up the image of the broadcasting center.
Keywords: Announcer training program, participant, requirements announced, theory of utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 751256 A Cumulative Learning Approach to Data Mining Employing Censored Production Rules (CPRs)
Authors: Rekha Kandwal, Kamal K.Bharadwaj
Abstract:
Knowledge is indispensable but voluminous knowledge becomes a bottleneck for efficient processing. A great challenge for data mining activity is the generation of large number of potential rules as a result of mining process. In fact sometimes result size is comparable to the original data. Traditional data mining pruning activities such as support do not sufficiently reduce the huge rule space. Moreover, many practical applications are characterized by continual change of data and knowledge, thereby making knowledge voluminous with each change. The most predominant representation of the discovered knowledge is the standard Production Rules (PRs) in the form If P Then D. Michalski & Winston proposed Censored Production Rules (CPRs), as an extension of production rules, that exhibit variable precision and supports an efficient mechanism for handling exceptions. A CPR is an augmented production rule of the form: If P Then D Unless C, where C (Censor) is an exception to the rule. Such rules are employed in situations in which the conditional statement 'If P Then D' holds frequently and the assertion C holds rarely. By using a rule of this type we are free to ignore the exception conditions, when the resources needed to establish its presence, are tight or there is simply no information available as to whether it holds or not. Thus the 'If P Then D' part of the CPR expresses important information while the Unless C part acts only as a switch changes the polarity of D to ~D. In this paper a scheme based on Dempster-Shafer Theory (DST) interpretation of a CPR is suggested for discovering CPRs from the discovered flat PRs. The discovery of CPRs from flat rules would result in considerable reduction of the already discovered rules. The proposed scheme incrementally incorporates new knowledge and also reduces the size of knowledge base considerably with each episode. Examples are given to demonstrate the behaviour of the proposed scheme. The suggested cumulative learning scheme would be useful in mining data streams.
Keywords: Censored production rules, cumulative learning, data mining, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1485255 A Novel Fuzzy Technique for Image Noise Reduction
Authors: Hamed Vahdat Nejad, Hameed Reza Pourreza, Hasan Ebrahimi
Abstract:
A new fuzzy filter is presented for noise reduction of images corrupted with additive noise. The filter consists of two stages. In the first stage, all the pixels of image are processed for determining noisy pixels. For this, a fuzzy rule based system associates a degree to each pixel. The degree of a pixel is a real number in the range [0,1], which denotes a probability that the pixel is not considered as a noisy pixel. In the second stage, another fuzzy rule based system is employed. It uses the output of the previous fuzzy system to perform fuzzy smoothing by weighting the contributions of neighboring pixel values. Experimental results are obtained to show the feasibility of the proposed filter. These results are also compared to other filters by numerical measure and visual inspection.Keywords: Additive noise, Fuzzy logic, Image processing, Noise reduction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2111254 Design of Bayesian MDS Sampling Plan Based on the Process Capability Index
Authors: Davood Shishebori, Mohammad Saber Fallah Nezhad, Sina Seifi
Abstract:
In this paper, a variable multiple dependent state (MDS) sampling plan is developed based on the process capability index using Bayesian approach. The optimal parameters of the developed sampling plan with respect to constraints related to the risk of consumer and producer are presented. Two comparison studies have been done. First, the methods of double sampling model, sampling plan for resubmitted lots and repetitive group sampling (RGS) plan are elaborated and average sample numbers of the developed MDS plan and other classical methods are compared. A comparison study between the developed MDS plan based on Bayesian approach and the exact probability distribution is carried out.
Keywords: MDS sampling plan, RGS plan, sampling plan for resubmitted lots, process capability index, average sample number, Bayesian approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1009253 Segmentation of Images through Clustering to Extract Color Features: An Application forImage Retrieval
Authors: M. V. Sudhamani, C. R. Venugopal
Abstract:
This paper deals with the application for contentbased image retrieval to extract color feature from natural images stored in the image database by segmenting the image through clustering. We employ a class of nonparametric techniques in which the data points are regarded as samples from an unknown probability density. Explicit computation of the density is avoided by using the mean shift procedure, a robust clustering technique, which does not require prior knowledge of the number of clusters, and does not constrain the shape of the clusters. A non-parametric technique for the recovery of significant image features is presented and segmentation module is developed using the mean shift algorithm to segment each image. In these algorithms, the only user set parameter is the resolution of the analysis and either gray level or color images are accepted as inputs. Extensive experimental results illustrate excellent performance.Keywords: Segmentation, Clustering, Image Retrieval, Features.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1460252 Mathematical Modeling for Dengue Transmission with the Effect of Season
Authors: R. Kongnuy., P. Pongsumpun
Abstract:
Mathematical models can be used to describe the transmission of disease. Dengue disease is the most significant mosquito-borne viral disease of human. It now a leading cause of childhood deaths and hospitalizations in many countries. Variations in environmental conditions, especially seasonal climatic parameters, effect to the transmission of dengue viruses the dengue viruses and their principal mosquito vector, Aedes aegypti. A transmission model for dengue disease is discussed in this paper. We assume that the human and vector populations are constant. We showed that the local stability is completely determined by the threshold parameter, 0 B . If 0 B is less than one, the disease free equilibrium state is stable. If 0 B is more than one, a unique endemic equilibrium state exists and is stable. The numerical results are shown for the different values of the transmission probability from vector to human populations.Keywords: Dengue disease, mathematical model, season, threshold parameters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2213251 Evolutionary Program Based Approach for Manipulator Grasping Color Objects
Authors: Y. Harold Robinson, M. Rajaram, Honey Raju
Abstract:
Image segmentation and color identification is an important process used in various emerging fields like intelligent robotics. A method is proposed for the manipulator to grasp and place the color object into correct location. The existing methods such as PSO, has problems like accelerating the convergence speed and converging to a local minimum leading to sub optimal performance. To improve the performance, we are using watershed algorithm and for color identification, we are using EPSO. EPSO method is used to reduce the probability of being stuck in the local minimum. The proposed method offers the particles a more powerful global exploration capability. EPSO methods can determine the particles stuck in the local minimum and can also enhance learning speed as the particle movement will be faster.Keywords: Color information, EPSO, hue, saturation, value (HSV), image segmentation, particle swarm optimization (PSO). Active Contour, GMM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581250 Reliability Based Optimal Design of Laterally Loaded Pile with Limited Residual Strain Energy Capacity
Authors: M. Movahedi Rad
Abstract:
In this study, a general approach to the reliability based limit analysis of laterally loaded piles is presented. In engineering practice the uncertainties play a very important role. The aim of this study is to evaluate the lateral load capacity of free-head and fixed-head long pile when plastic limit analysis is considered. In addition to the plastic limit analysis to control the plastic behaviour of the structure, uncertain bound on the complementary strain energy of the residual forces is also applied. This bound has significant effect for the load parameter. The solution to reliability-based problems is obtained by a computer program which is governed by the reliability index calculation.Keywords: Reliability, laterally loaded pile, residual strain energy, probability, limit analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1902249 3D Object Model Reconstruction Based on Polywogs Wavelet Network Parametrization
Authors: Mohamed Othmani, Yassine Khlifi
Abstract:
This paper presents a technique for compact three dimensional (3D) object model reconstruction using wavelet networks. It consists to transform an input surface vertices into signals,and uses wavelet network parameters for signal approximations. To prove this, we use a wavelet network architecture founded on several mother wavelet families. POLYnomials WindOwed with Gaussians (POLYWOG) wavelet families are used to maximize the probability to select the best wavelets which ensure the good generalization of the network. To achieve a better reconstruction, the network is trained several iterations to optimize the wavelet network parameters until the error criterion is small enough. Experimental results will shown that our proposed technique can effectively reconstruct an irregular 3D object models when using the optimized wavelet network parameters. We will prove that an accurateness reconstruction depends on the best choice of the mother wavelets.Keywords: 3D object, optimization, parametrization, Polywog wavelets, reconstruction, wavelet networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1501248 Active Contours with Prior Corner Detection
Authors: U.A.A. Niroshika, Ravinda G.N. Meegama
Abstract:
Deformable active contours are widely used in computer vision and image processing applications for image segmentation, especially in biomedical image analysis. The active contour or “snake" deforms towards a target object by controlling the internal, image and constraint forces. However, if the contour initialized with a lesser number of control points, there is a high probability of surpassing the sharp corners of the object during deformation of the contour. In this paper, a new technique is proposed to construct the initial contour by incorporating prior knowledge of significant corners of the object detected using the Harris operator. This new reconstructed contour begins to deform, by attracting the snake towards the targeted object, without missing the corners. Experimental results with several synthetic images show the ability of the new technique to deal with sharp corners with a high accuracy than traditional methods.Keywords: Active Contours, Image Segmentation, Harris Operator, Snakes
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2281247 Integral Operators Related to Problems of Interface Dynamics
Authors: Pa Pa Lin
Abstract:
This research work is concerned with the eigenvalue problem for the integral operators which are obtained by linearization of a nonlocal evolution equation. The purpose of section II.A is to describe the nature of the problem and the objective of the project. The problem is related to the “stable solution" of the evolution equation which is the so-called “instanton" that describe the interface between two stable phases. The analysis of the instanton and its asymptotic behavior are described in section II.C by imposing the Green function and making use of a probability kernel. As a result , a classical Theorem which is important for an instanton is proved. Section III devoted to a study of the integral operators related to interface dynamics which concern the analysis of the Cauchy problem for the evolution equation with initial data close to different phases and different regions of space.
Keywords: Evolution, Green function, instanton, integral operators.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1231