Search results for: comparison research method.
12579 On Algebraic Structure of Improved Gauss-Seidel Iteration
Authors: O. M. Bamigbola, A. A. Ibrahim
Abstract:
Analysis of real life problems often results in linear systems of equations for which solutions are sought. The method to employ depends, to some extent, on the properties of the coefficient matrix. It is not always feasible to solve linear systems of equations by direct methods, as such the need to use an iterative method becomes imperative. Before an iterative method can be employed to solve a linear system of equations there must be a guaranty that the process of solution will converge. This guaranty, which must be determined apriori, involve the use of some criterion expressible in terms of the entries of the coefficient matrix. It is, therefore, logical that the convergence criterion should depend implicitly on the algebraic structure of such a method. However, in deference to this view is the practice of conducting convergence analysis for Gauss- Seidel iteration on a criterion formulated based on the algebraic structure of Jacobi iteration. To remedy this anomaly, the Gauss- Seidel iteration was studied for its algebraic structure and contrary to the usual assumption, it was discovered that some property of the iteration matrix of Gauss-Seidel method is only diagonally dominant in its first row while the other rows do not satisfy diagonal dominance. With the aid of this structure we herein fashion out an improved version of Gauss-Seidel iteration with the prospect of enhancing convergence and robustness of the method. A numerical section is included to demonstrate the validity of the theoretical results obtained for the improved Gauss-Seidel method.
Keywords: Linear system of equations, Gauss-Seidel iteration, algebraic structure, convergence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 293412578 An Improved Sub-Nyquist Sampling Jamming Method for Deceiving Inverse Synthetic Aperture Radar
Authors: Yanli Qi, Ning Lv, Jing Li
Abstract:
Sub-Nyquist sampling jamming method (SNSJ) is a well known deception jamming method for inverse synthetic aperture radar (ISAR). However, the anti-decoy of the SNSJ method performs easier since the amplitude of the false-target images are weaker than the real-target image; the false-target images always lag behind the real-target image, and all targets are located in the same cross-range. In order to overcome the drawbacks mentioned above, a simple modulation based on SNSJ (M-SNSJ) is presented in this paper. The method first uses amplitude modulation factor to make the amplitude of the false-target images consistent with the real-target image, then uses the down-range modulation factor and cross-range modulation factor to make the false-target images move freely in down-range and cross-range, respectively, thus the capacity of deception is improved. Finally, the simulation results on the six available combinations of three modulation factors are given to illustrate our conclusion.
Keywords: Inverse synthetic aperture radar, ISAR, deceptive jamming, Sub-Nyquist sampling jamming method, SNSJ, modulation based on Sub-Nyquist sampling jamming method, M-SNSJ.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128512577 Optimized Data Fusion in an Intelligent Integrated GPS/INS System Using Genetic Algorithm
Authors: Ali Asadian, Behzad Moshiri, Ali Khaki Sedigh, Caro Lucas
Abstract:
Most integrated inertial navigation systems (INS) and global positioning systems (GPS) have been implemented using the Kalman filtering technique with its drawbacks related to the need for predefined INS error model and observability of at least four satellites. Most recently, a method using a hybrid-adaptive network based fuzzy inference system (ANFIS) has been proposed which is trained during the availability of GPS signal to map the error between the GPS and the INS. Then it will be used to predict the error of the INS position components during GPS signal blockage. This paper introduces a genetic optimization algorithm that is used to update the ANFIS parameters with respect to the INS/GPS error function used as the objective function to be minimized. The results demonstrate the advantages of the genetically optimized ANFIS for INS/GPS integration in comparison with conventional ANFIS specially in the cases of satellites- outages. Coping with this problem plays an important role in assessment of the fusion approach in land navigation.Keywords: Adaptive Network based Fuzzy Inference System (ANFIS), Genetic optimization, Global Positioning System (GPS), Inertial Navigation System (INS).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 191312576 Semi-Lagrangian Method for Advection Equation on GPU in Unstructured R3 Mesh for Fluid Dynamics Application
Authors: Irakli V. Gugushvili, Nickolay M. Evstigneev
Abstract:
Numerical integration of initial boundary problem for advection equation in 3 ℜ is considered. The method used is conditionally stable semi-Lagrangian advection scheme with high order interpolation on unstructured mesh. In order to increase time step integration the BFECC method with limiter TVD correction is used. The method is adopted on parallel graphic processor unit environment using NVIDIA CUDA and applied in Navier-Stokes solver. It is shown that the calculation on NVIDIA GeForce 8800 GPU is 184 times faster than on one processor AMDX2 4800+ CPU. The method is extended to the incompressible fluid dynamics solver. Flow over a Cylinder for 3D case is compared to the experimental data.
Keywords: Advection equations, CUDA technology, Flow overthe 3D Cylinder, Incompressible Pressure Projection Solver, Parallel computation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 285312575 Minimal Residual Method for Adaptive Filtering with Finite Termination
Authors: Noor Atinah Ahmad, Shazia Javed
Abstract:
We present a discussion of three adaptive filtering algorithms well known for their one-step termination property, in terms of their relationship with the minimal residual method. These algorithms are the normalized least mean square (NLMS), Affine Projection algorithm (APA) and the recursive least squares algorithm (RLS). The NLMS is shown to be a result of the orthogonality condition imposed on the instantaneous approximation of the Wiener equation, while APA and RLS algorithm result from orthogonality condition in multi-dimensional minimal residual formulation. Further analysis of the minimal residual formulation for the RLS leads to a triangular system which also possesses the one-step termination property (in exact arithmetic)Keywords: Adaptive filtering, minimal residual method, projection method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155812574 Recovering the Boundary Data in the Two Dimensional Inverse Heat Conduction Problem Using the Ritz-Galerkin Method
Authors: Saeed Sarabadan, Kamal Rashedi
Abstract:
This article presents a numerical method to find the heat flux in an inhomogeneous inverse heat conduction problem with linear boundary conditions and an extra specification at the terminal. The method is based upon applying the satisfier function along with the Ritz-Galerkin technique to reduce the approximate solution of the inverse problem to the solution of a system of algebraic equations. The instability of the problem is resolved by taking advantage of the Landweber’s iterations as an admissible regularization strategy. In computations, we find the stable and low-cost results which demonstrate the efficiency of the technique.Keywords: Inverse problem, parabolic equations, heat equation, Ritz-Galerkin method, Landweber iterations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 119712573 A Protocol for Applied Consumer Behavior Research in Academia
Abstract:
A Montana university has used applied consumer research in experiential learning with non-profit clients for over a decade. Through trial and error, a successful protocol has been established from problem statement through formative research to integrated marketing campaign execution. In this paper, we describe the protocol and its applications. Analysis was completed to determine the effectiveness of the campaigns and the results of how pre- and post-consumer research mark societal change because of media.
Keywords: Marketing, experiential learning, consumer behavior, community partner.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19512572 Online Collaboration Learning: A Way to Enhance Students' Achievement at Kingdom of Bahrain
Authors: Jaflah H. Al-Ammary
Abstract:
The increasing recognition of the need for education to be closely aligned with team playing, project based learning and problem solving approaches has increase the interest in collaborative learning among university and college instructors. Using online collaboration learning in learning can enhance the outcome and achievement of students as well as improve their communication, critical thinking and personnel skills. The current research aims at examining the effect of OCL on the student's achievement at Kingdom of Bahrain. Numbers of objectives were set to achieve the aim of the research include: investigating the current situation regarding the collaborative learning and OCL at the Kingdom of Bahrain by identifying the advantages and effectiveness of OCL as a learning tool over traditional learning, examining the factors that affect OCL as well as examining the impact of OCL on the student's achievement. To achieve these objectives, quantitative method was adopted. Two hundred and thirty one questionnaires were distributed to students in different local and private universities at Kingdom of Bahrain. The findings of the research show that most of the students prefer to use FTFCL in learning and that OCL is already adopted in some universities especially in University of Bahrain. Moreover, the most factors affecting the adopted OCL are perceived readiness, and guidance and support.
Keywords: Collaborative learning, perceived readiness, student achievement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 230212571 Implicit Eulerian Fluid-Structure Interaction Method for the Modeling of Highly Deformable Elastic Membranes
Authors: Aymen Laadhari, Gábor Székely
Abstract:
This paper is concerned with the development of a fully implicit and purely Eulerian fluid-structure interaction method tailored for the modeling of the large deformations of elastic membranes in a surrounding Newtonian fluid. We consider a simplified model for the mechanical properties of the membrane, in which the surface strain energy depends on the membrane stretching. The fully Eulerian description is based on the advection of a modified surface tension tensor, and the deformations of the membrane are tracked using a level set strategy. The resulting nonlinear problem is solved by a Newton-Raphson method, featuring a quadratic convergence behavior. A monolithic solver is implemented, and we report several numerical experiments aimed at model validation and illustrating the accuracy of the presented method. We show that stability is maintained for significantly larger time steps.Keywords: Fluid-membrane interaction, stretching, Eulerian, finite element method, Newton, implicit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128912570 Comparison of Indoor and Outdoor Air Quality in Children Homes at Prenatal Period and One Year Old
Authors: S. Lakestani, B. Karakas, S. Acar Vaizoglu, B. Guciz Dogan, C. Guler, B. Sekerel, A. Taner, G. Gullu
Abstract:
Abstract–Indoor air (VOCs) samples were collected simultaneously from variety of indoors (e.g. living rooms, baby-s rooms) and outdoor environments which were voluntarily selected from the houses in which pregnant residents live throughout Ankara. This is the first comprehensive study done in Turkey starting from prenatal period and continued till the babies had one year old. VOCs levels were measured over 76 homes. Air samples were collected in Tenax TA sorbent filled tubes with active sampling method and analyzed with Thermal Desorber and Gas Chromatography/Mass spectrometry (TD-GC/MS). At the first sampling period in the baby-s rooms maximum concentration of toluene was measured about 240.77μg.m-3 and in the living rooms maximum concentration of naphthalene was 180.24μg.m-3. At the second sampling period in the baby-s rooms maximum concentration of toluene was measured about 144.97μg.m-3 and in the living rooms maximum concentration of naphthalene was 247.89μg.m-3. Concentration of TVOCs in the first period was generally higher than the second period.Keywords: Indoor Air, Volatile Organic Compounds (VOCs), Gas Chromatography
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 200312569 Improvements in Edge Detection Based on Mathematical Morphology and Wavelet Transform using Fuzzy Rules
Authors: Masrour Dowlatabadi, Jalil Shirazi
Abstract:
In this paper, an improved edge detection algorithm based on fuzzy combination of mathematical morphology and wavelet transform is proposed. The combined method is proposed to overcome the limitation of wavelet based edge detection and mathematical morphology based edge detection in noisy images. Experimental results show superiority of the proposed method, as compared to the traditional Prewitt, wavelet based and morphology based edge detection methods. The proposed method is an effective edge detection method for noisy image and keeps clear and continuous edges.Keywords: Edge detection, Wavelet transform, Mathematical morphology, Fuzzy logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 240612568 A Kernel Based Rejection Method for Supervised Classification
Authors: Abdenour Bounsiar, Edith Grall, Pierre Beauseroy
Abstract:
In this paper we are interested in classification problems with a performance constraint on error probability. In such problems if the constraint cannot be satisfied, then a rejection option is introduced. For binary labelled classification, a number of SVM based methods with rejection option have been proposed over the past few years. All of these methods use two thresholds on the SVM output. However, in previous works, we have shown on synthetic data that using thresholds on the output of the optimal SVM may lead to poor results for classification tasks with performance constraint. In this paper a new method for supervised classification with rejection option is proposed. It consists in two different classifiers jointly optimized to minimize the rejection probability subject to a given constraint on error rate. This method uses a new kernel based linear learning machine that we have recently presented. This learning machine is characterized by its simplicity and high training speed which makes the simultaneous optimization of the two classifiers computationally reasonable. The proposed classification method with rejection option is compared to a SVM based rejection method proposed in recent literature. Experiments show the superiority of the proposed method.Keywords: rejection, Chow's rule, error-reject tradeoff, SupportVector Machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 144912567 Design for Manufacturability and Concurrent Engineering for Product Development
Authors: Alemu Moges Belay
Abstract:
In the 1980s, companies began to feel the effect of three major influences on their product development: newer and innovative technologies, increasing product complexity and larger organizations. And therefore companies were forced to look for new product development methods. This paper tries to focus on the two of new product development methods (DFM and CE). The aim of this paper is to see and analyze different product development methods specifically on Design for Manufacturability and Concurrent Engineering. Companies can achieve and be benefited by minimizing product life cycle, cost and meeting delivery schedule. This paper also presents simplified models that can be modified and used by different companies based on the companies- objective and requirements. Methodologies that are followed to do this research are case studies. Two companies were taken and analysed on the product development process. Historical data, interview were conducted on these companies in addition to that, Survey of literatures and previous research works on similar topics has been done during this research. This paper also tries to show the implementation cost benefit analysis and tries to calculate the implementation time. From this research, it has been found that the two companies did not achieve the delivery time to the customer. Some of most frequently coming products are analyzed and 50% to 80 % of their products are not delivered on time to the customers. The companies are following the traditional way of product development that is sequentially design and production method, which highly affect time to market. In the case study it is found that by implementing these new methods and by forming multi disciplinary team in designing and quality inspection; the company can reduce the workflow steps from 40 to 30.
Keywords: Design for manufacturability, Concurrent Engineering, Time-to-Market, Product development
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 558812566 An Analysis of Dynamic Economic Dispatch Using Search Space Reduction Based Gravitational Search Algorithm
Authors: K. C. Meher, R. K. Swain, C. K. Chanda
Abstract:
This paper presents the performance analysis of dynamic search space reduction (DSR) based gravitational search algorithm (GSA) to solve dynamic economic dispatch of thermal generating units with valve point effects. Dynamic economic dispatch basically dictates the best setting of generator units with anticipated load demand over a definite period of time. In this paper, the presented technique is considered that deals an inequality constraints treatment mechanism known as DSR strategy to accelerate the optimization process. The presented method is demonstrated through five-unit test systems to verify its effectiveness and robustness. The simulation results are compared with other existing evolutionary methods reported in the literature. It is intuited from the comparison that the fuel cost and other performances of the presented approach yield fruitful results with marginal value of simulation time.Keywords: Dynamic economic dispatch, dynamic search space reduction strategy, gravitational search algorithm, ramp rate limits, valve-point effects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 149812565 Choosing R-tree or Quadtree Spatial DataIndexing in One Oracle Spatial Database System to Make Faster Showing Geographical Map in Mobile Geographical Information System Technology
Authors: Maruto Masserie Sardadi, Mohd Shafry bin Mohd Rahim, Zahabidin Jupri, Daut bin Daman
Abstract:
The latest Geographic Information System (GIS) technology makes it possible to administer the spatial components of daily “business object," in the corporate database, and apply suitable geographic analysis efficiently in a desktop-focused application. We can use wireless internet technology for transfer process in spatial data from server to client or vice versa. However, the problem in wireless Internet is system bottlenecks that can make the process of transferring data not efficient. The reason is large amount of spatial data. Optimization in the process of transferring and retrieving data, however, is an essential issue that must be considered. Appropriate decision to choose between R-tree and Quadtree spatial data indexing method can optimize the process. With the rapid proliferation of these databases in the past decade, extensive research has been conducted on the design of efficient data structures to enable fast spatial searching. Commercial database vendors like Oracle have also started implementing these spatial indexing to cater to the large and diverse GIS. This paper focuses on the decisions to choose R-tree and quadtree spatial indexing using Oracle spatial database in mobile GIS application. From our research condition, the result of using Quadtree and R-tree spatial data indexing method in one single spatial database can save the time until 42.5%.Keywords: Indexing, Mobile GIS, MapViewer, Oracle SpatialDatabase.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 403912564 Chemical and Sensory Properties of Chardonnay Wines Produced in Different Oak Barrels
Authors: Valentina Obradović, Josip Mesić, Maja Ergović Ravančić, Kamila Mijowska, Brankica Svitlica
Abstract:
French oak and American oak barrels are most famous all over the world, but barrels of different origin can also be used for obtaining high quality wines. The aim of this research was to compare the influence of different Slovenian (Croatian) and French oak barrels on the quality of Chardonnay wine. Grapes were grown in the Croatian wine growing region of Kutjevo in 2015. Chardonnay wines were tested for basic oenological parameters (alcohol, extract, reducing sugar, SO2, acidity), total polyphenols content (Folin-Ciocalteu method), antioxidant activity (ABTS and DPPH method) and colour density. Sensory evaluation was performed by students of viticulture/oenology. Samples produced by classical fermentation and ageing in French oak barrels had better results for polyphenols and sensory evaluation (especially low toasting level) than samples in Slovenian barrels. All tested samples were scored as a “quality” or “premium quality” wines. Sur lie method of fermentation and ageing in Slovenian oak barrel had very good extraction of polyphenols and high antioxidant activity with the usage of authentic yeasts, while commercial yeast strain resulted in worse chemical and sensory parameters.
Keywords: Chardonnay, French oak, Slovenian oak, sur lie.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 95012563 Financial Literacy Testing: Results of Conducted Research and Introduction of a Project
Authors: J. Nesleha, H. Florianova
Abstract:
The goal of the study is to provide results of a conducted study devoted to financial literacy in the Czech Republic and to introduce a project related to financial education in the Czech Republic. Financial education has become an important part of education in the country, yet it is still neglected on the lowest level of formal education–primary schools. The project is based on investigation of financial literacy on primary schools in the Czech Republic. Consequently, the authors aim to formulate possible amendments related to this type of education. The gained dataset is intended to be used for analysis concerning financial education in the Czech Republic. With regard to used methods, the most important one is regression analysis for disclosure of predictors causing different levels of financial literacy. Furthermore, comparison of different groups is planned, for which t-tests are intended to be used. The study also employs descriptive statistics to introduce basic relationship in the data file.Keywords: Czech Republic, financial education, financial literacy, primary school, regression analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 85712562 Increasing Value Added of Recycling Business Management: A Case of Thailand
Authors: Yananda Siraphatthada
Abstract:
This policy participation action research explores the roles of Thai government units during its 2010 fiscal year on how to create value added to recycling business in the central part of Thailand. The research aims to a) study how the government plays a role to support the business, and its problems and obstacles on supporting the business, b) to design a strategic action – short, medium, and long term plans -- to create value added to the recycling business, particularly in local full-loop companies/organizations licensed by Wongpanit Waste Separation Plant as well as those licensed by the Department of Provincial Administration. Mixed method research design, i.e., a combination of quantitative and qualitative methods is utilized in the present study in both data collection and analysis procedures. Quantitative data was analyzed by frequency, percent value, mean scores, and standard deviation, and aimed to note trend and generalizations. Qualitative data was collected via semi-structured interviews/focus group interviews to explore in-depth views of the operators. The sampling included 1,079 operators in eight provinces in the central part of Thailand.Keywords: Management, Recycling Business, Value Added.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 276012561 Absorption of CO2 in EAF Reducing Slag from Stainless Steel Making Process by Wet Grinding
Authors: B.M.N. Nik Hisyamudin, S. Yokoyama, M. Umemoto
Abstract:
In the current study, we have conducted an experimental investigation on the utilization of electronic arc furnace (EAF) reducing slag for the absorption of CO2 via wet grinding method. It was carried out by various grinding conditions. The slag was ground in the vibrating ball mill in the presence of CO2 and pure water under ambient temperature. The reaction behavior was monitored with constant pressure method, and the changes of experimental systems volume as a function of grinding time were measured. It was found that the CO2 absorption occurred as soon as the grinding started. The CO2 absorption was significantly increased in the case of wet grinding compare to the dry grinding. Generally, the amount of CO2 absorption increased as the amount of water, weight of slag and initial pressure increased. However, it was decreased when the amount of water exceeds 200ml and when smaller balls were used. The absorption of CO2 occurred simultaneously with the start of the grinding and it stopped when the grinding was stopped. According to this research, the CO2 reacted with the CaO inside the slag, forming CaCO3.Keywords: CO2 absorption, EAF reducing slag, vibration ball mill, wet grinding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 178812560 Robust Face Recognition using AAM and Gabor Features
Authors: Sanghoon Kim, Sun-Tae Chung, Souhwan Jung, Seoungseon Jeon, Jaemin Kim, Seongwon Cho
Abstract:
In this paper, we propose a face recognition algorithm using AAM and Gabor features. Gabor feature vectors which are well known to be robust with respect to small variations of shape, scaling, rotation, distortion, illumination and poses in images are popularly employed for feature vectors for many object detection and recognition algorithms. EBGM, which is prominent among face recognition algorithms employing Gabor feature vectors, requires localization of facial feature points where Gabor feature vectors are extracted. However, localization method employed in EBGM is based on Gabor jet similarity and is sensitive to initial values. Wrong localization of facial feature points affects face recognition rate. AAM is known to be successfully applied to localization of facial feature points. In this paper, we devise a facial feature point localization method which first roughly estimate facial feature points using AAM and refine facial feature points using Gabor jet similarity-based facial feature localization method with initial points set by the rough facial feature points obtained from AAM, and propose a face recognition algorithm using the devised localization method for facial feature localization and Gabor feature vectors. It is observed through experiments that such a cascaded localization method based on both AAM and Gabor jet similarity is more robust than the localization method based on only Gabor jet similarity. Also, it is shown that the proposed face recognition algorithm using this devised localization method and Gabor feature vectors performs better than the conventional face recognition algorithm using Gabor jet similarity-based localization method and Gabor feature vectors like EBGM.Keywords: Face Recognition, AAM, Gabor features, EBGM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 221012559 Edge Detection Using Multi-Agent System: Evaluation on Synthetic and Medical MR Images
Authors: A. Nachour, L. Ouzizi, Y. Aoura
Abstract:
Recent developments on multi-agent system have brought a new research field on image processing. Several algorithms are used simultaneously and improved in deferent applications while new methods are investigated. This paper presents a new automatic method for edge detection using several agents and many different actions. The proposed multi-agent system is based on parallel agents that locally perceive their environment, that is to say, pixels and additional environmental information. This environment is built using Vector Field Convolution that attract free agent to the edges. Problems of partial, hidden or edges linking are solved with the cooperation between agents. The presented method was implemented and evaluated using several examples on different synthetic and medical images. The obtained experimental results suggest that this approach confirm the efficiency and accuracy of detected edge.
Keywords: Edge detection, medical MR images, multi-agent systems, vector field convolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 190812558 Parametric Approach for Reserve Liability Estimate in Mortgage Insurance
Authors: Rajinder Singh, Ram Valluru
Abstract:
Chain Ladder (CL) method, Expected Loss Ratio (ELR) method and Bornhuetter-Ferguson (BF) method, in addition to more complex transition-rate modeling, are commonly used actuarial reserving methods in general insurance. There is limited published research about their relative performance in the context of Mortgage Insurance (MI). In our experience, these traditional techniques pose unique challenges and do not provide stable claim estimates for medium to longer term liabilities. The relative strengths and weaknesses among various alternative approaches revolve around: stability in the recent loss development pattern, sufficiency and reliability of loss development data, and agreement/disagreement between reported losses to date and ultimate loss estimate. CL method results in volatile reserve estimates, especially for accident periods with little development experience. The ELR method breaks down especially when ultimate loss ratios are not stable and predictable. While the BF method provides a good tradeoff between the loss development approach (CL) and ELR, the approach generates claim development and ultimate reserves that are disconnected from the ever-to-date (ETD) development experience for some accident years that have more development experience. Further, BF is based on subjective a priori assumption. The fundamental shortcoming of these methods is their inability to model exogenous factors, like the economy, which impact various cohorts at the same chronological time but at staggered points along their life-time development. This paper proposes an alternative approach of parametrizing the loss development curve and using logistic regression to generate the ultimate loss estimate for each homogeneous group (accident year or delinquency period). The methodology was tested on an actual MI claim development dataset where various cohorts followed a sigmoidal trend, but levels varied substantially depending upon the economic and operational conditions during the development period spanning over many years. The proposed approach provides the ability to indirectly incorporate such exogenous factors and produce more stable loss forecasts for reserving purposes as compared to the traditional CL and BF methods.
Keywords: Actuarial loss reserving techniques, logistic regression, parametric function, volatility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41912557 Neuro-fuzzy Model and Regression Model a Comparison Study of MRR in Electrical Discharge Machining of D2 Tool Steel
Authors: M. K. Pradhan, C. K. Biswas,
Abstract:
In the current research, neuro-fuzzy model and regression model was developed to predict Material Removal Rate in Electrical Discharge Machining process for AISI D2 tool steel with copper electrode. Extensive experiments were conducted with various levels of discharge current, pulse duration and duty cycle. The experimental data are split into two sets, one for training and the other for validation of the model. The training data were used to develop the above models and the test data, which was not used earlier to develop these models were used for validation the models. Subsequently, the models are compared. It was found that the predicted and experimental results were in good agreement and the coefficients of correlation were found to be 0.999 and 0.974 for neuro fuzzy and regression model respectively
Keywords: Electrical discharge machining, material removal rate, neuro-fuzzy model, regression model, mountain clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 139112556 The Harada Method – A Method for Employee Development during Production Ramp Up
Authors: M. Goerke, J. Gehrmann
Abstract:
Caused by shorter product life cycles and higher product variety the importance of production ramp ups is increasing. Even though companies are aware of that fact, up to 40% of the ramp up projects still miss technical and economical requirements. The success of a ramp up depends on the planning of human factors, organizational aspects and technological solutions. Since only partly considered in scientific literature, this paper lays its focus on the human factor during production ramp up. There are only incoherent methods which address the problems in this area. A systematic and holistic method to improve the capabilities of the employees during ramp up is missing. The Harada Method is a relatively young approach for developing highly-skilled workers. It consists of different worksheets which help employees to set guidelines and reach overall objectives. This approach is going to be transferred into a tool for ramp up management.
Keywords: Employee Development, Harada, Production Ramp Up.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 228012555 Walking Hexapod Robot in Disaster Recovery: Developing Algorithm for Terrain Negotiation and Navigation
Authors: Md. Masum Billah, Mohiuddin Ahmed, Soheli Farhana
Abstract:
In modern day disaster recovery mission has become one of the top priorities in any natural disaster management regime. Smart autonomous robots may play a significant role in such missions, including search for life under earth quake hit rubbles, Tsunami hit islands, de-mining in war affected areas and many other such situations. In this paper current state of many walking robots are compared and advantages of hexapod systems against wheeled robots are described. In our research we have selected a hexapod spider robot; we are developing focusing mainly on efficient navigation method in different terrain using apposite gait of locomotion, which will make it faster and at the same time energy efficient to navigate and negotiate difficult terrain. This paper describes the method of terrain negotiation navigation in a hazardous field.Keywords: Walking robots, locomotion, hexapod robot, gait, hazardous field.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 443912554 Comparison of Domain and Hydrophobicity Features for the Prediction of Protein-Protein Interactions using Support Vector Machines
Authors: Hany Alashwal, Safaai Deris, Razib M. Othman
Abstract:
The protein domain structure has been widely used as the most informative sequence feature to computationally predict protein-protein interactions. However, in a recent study, a research group has reported a very high accuracy of 94% using hydrophobicity feature. Therefore, in this study we compare and verify the usefulness of protein domain structure and hydrophobicity properties as the sequence features. Using the Support Vector Machines (SVM) as the learning system, our results indicate that both features achieved accuracy of nearly 80%. Furthermore, domains structure had receiver operating characteristic (ROC) score of 0.8480 with running time of 34 seconds, while hydrophobicity had ROC score of 0.8159 with running time of 20,571 seconds (5.7 hours). These results indicate that protein-protein interaction can be predicted from domain structure with reliable accuracy and acceptable running time.
Keywords: Bioinformatics, protein-protein interactions, support vector machines, protein features.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 192112553 The Study on Evaluation System and Method of Legacy System
Authors: Chao Qi, Fuyang Peng, Bo Deng, Xiaoyan Su
Abstract:
In the upgrade process of enterprise information systems, how to deal with and utilize those legacy systems affects the efficiency of construction and development of the new system. We propose an evaluation system, which comprehensively describes the capacity of legacy information systems in five aspects. Then we propose a practical legacy systems evaluation method. Base on the evaluation result, we can determine the current state of legacy system which was evaluated.Keywords: Legacy Information Systems, Evaluation IndexSystem, Evaluation Method, Evaluation Level
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162212552 A Systematic Approach for Identifying Turning Center Capabilities with Vertical Machining Center in Milling Operation
Abstract:
Conventional machining is a form of subtractive manufacturing, in which a collection of material-working processes utilizing power-driven machine tools are used to remove undesired material to achieve a desired geometry. This paper presents an approach for comparison between turning center and vertical machining center by optimization of cutting parameters at cylindrical workpieces leading to minimum surface roughness by using taguchi methodology. Aluminum alloy was taken to conduct experiments due to its unique high strength-weight ratio that is maintained at elevated temperatures and their exceptional corrosion resistance. During testing, the effects of the cutting parameters on the surface roughness were investigated. Additionally, by using taguchi methodology for each of the cutting parameters (spindle speed, depth of cut, insert diameter, and feed rate) minimum surface roughness for the process of turn-milling was determined according to the cutting parameters. A confirmation experiment demonstrates the effectiveness of taguchi method.
Keywords: Surface roughness, taguchi parameter design, turning center, turn-milling operations, vertical machining center.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 252612551 Spectral Entropy Employment in Speech Enhancement based on Wavelet Packet
Authors: Talbi Mourad, Salhi Lotfi, Chérif Adnen
Abstract:
In this work, we are interested in developing a speech denoising tool by using a discrete wavelet packet transform (DWPT). This speech denoising tool will be employed for applications of recognition, coding and synthesis. For noise reduction, instead of applying the classical thresholding technique, some wavelet packet nodes are set to zero and the others are thresholded. To estimate the non stationary noise level, we employ the spectral entropy. A comparison of our proposed technique to classical denoising methods based on thresholding and spectral subtraction is made in order to evaluate our approach. The experimental implementation uses speech signals corrupted by two sorts of noise, white and Volvo noises. The obtained results from listening tests show that our proposed technique is better than spectral subtraction. The obtained results from SNR computation show the superiority of our technique when compared to the classical thresholding method using the modified hard thresholding function based on u-law algorithm.
Keywords: Enhancement, spectral subtraction, SNR, discrete wavelet packet transform, spectral entropy Histogram
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 199512550 Study of the Electromagnetic Resonances of a Cavity with an Aperture Using Numerical Method and Equivalent Circuit Method
Authors: Ming-Chu Yin, Ping-An Du
Abstract:
The shielding ability of a shielding cavity with an aperture will be greatly degraded at resonance frequencies, and the resonance modes and frequencies are affected by aperture resonances and aperture-cavity coupling, which are closely related with aperture sizes. The equivalent circuit method and numerical method of Transmission Line Matrix (TLM) are used to analyze the effects of aperture resonances and aperture-cavity coupling on the electromagnetic resonances of a cavity with an aperture in this paper. Both analytical and numerical results show that the resonance modes of a shielding cavity with an aperture consist of cavity resonance modes and aperture resonance modes, and the resonance frequencies will shift with the change of the aperture sizes because of the aperture resonances and aperture-cavity coupling. Variation rules of electromagnetic resonances with aperture sizes for a cavity with an aperture are given, which will be useful for design of shielding cavities.Keywords: Aperture-cavity coupling, equivalent circuit method, resonances, shielding equipment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1601