Search results for: regularized loss minimization.
1045 Minimizing Fish-feed Loss due to Sea Currents: An Economic Methodology
Authors: V. Vassiliou, M. Charalambides, M. Menicou
Abstract:
Fish-feed is a major cost component of operating expenses for any aquaculture farm. Due to soaring prices of fish-feed ingredients, the need for better feeding schedule management has become imperative. On such factor that influences the utilization rate of fish-feed are sea currents. Up to now, practical monitoring of fishfeed loss due to sea currents is not exercised. This paper gives a description of an economic methodology that aims at quantifying the amount of fish-feed lost due to sea currents and draws on data from a Mediterranean aquaculture farm to formulate the associated model.
Keywords: Aquaculture, economic model, fish-feed loss, sea currents.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18221044 LMI Approach to Regularization and Stabilization of Linear Singular Systems: The Discrete-time Case
Authors: Salim Ibrir
Abstract:
Sufficient linear matrix inequalities (LMI) conditions for regularization of discrete-time singular systems are given. Then a new class of regularizing stabilizing controllers is discussed. The proposed controllers are the sum of predictive and memoryless state feedbacks. The predictive controller aims to regularizing the singular system while the memoryless state feedback is designed to stabilize the resulting regularized system. A systematic procedure is given to calculate the controller gains through linear matrix inequalities.
Keywords: Singular systems, Discrete-time systems, Regularization, LMIs
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15911043 Multi-Objective Multi-Mode Resource-Constrained Project Scheduling Problem by Preemptive Fuzzy Goal Programming
Authors: Phruksaphanrat B.
Abstract:
This research proposes a preemptive fuzzy goal programming model for multi-objective multi-mode resource constrained project scheduling problem. The objectives of the problem are minimization of the total time and the total cost of the project. Objective in a multi-mode resource-constrained project scheduling problem is often a minimization of makespan. However, both time and cost should be considered at the same time with different level of important priorities. Moreover, all elements of cost functions in a project are not included in the conventional cost objective function. Incomplete total project cost causes an error in finding the project scheduling time. In this research, preemptive fuzzy goal programming is presented to solve the multi-objective multi-mode resource constrained project scheduling problem. It can find the compromise solution of the problem. Moreover, it is also flexible in adjusting to find a variety of alternative solutions.
Keywords: Multi-mode resource constrained project scheduling problem, Fuzzy set, Goal programming, Preemptive fuzzy goal programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27551042 Beyond Taguchi’s Concept of the Quality Loss Function
Authors: Atul Dev, Pankaj Jha
Abstract:
Dr. Genichi Taguchi looked at quality in a broader term and gave an excellent definition of quality in terms of loss to society. However the scope of this definition is limited to the losses imparted by a poor quality product to the customer only and are considered during the useful life of the product and further in a certain situation this loss can even be zero. In this paper, it has been proposed that the scope of quality of a product shall be further enhanced by considering the losses imparted by a poor quality product to society at large, due to associated environmental and safety related factors, over the complete life cycle of the product. Moreover, though these losses can be further minimized with the use of techno-safety interventions, the net losses to society however can never be made zero. This paper proposes an entirely new approach towards defining product quality and is based on Taguchi’s definition of quality.
Keywords: Existing concept, goal post philosophy, life cycle, proposed concept, quality loss function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20031041 Operational Risk – Scenario Analysis
Authors: Milan Rippel, Petr Teply
Abstract:
This paper focuses on operational risk measurement techniques and on economic capital estimation methods. A data sample of operational losses provided by an anonymous Central European bank is analyzed using several approaches. Loss Distribution Approach and scenario analysis method are considered. Custom plausible loss events defined in a particular scenario are merged with the original data sample and their impact on capital estimates and on the financial institution is evaluated. Two main questions are assessed – What is the most appropriate statistical method to measure and model operational loss data distribution? and What is the impact of hypothetical plausible events on the financial institution? The g&h distribution was evaluated to be the most suitable one for operational risk modeling. The method based on the combination of historical loss events modeling and scenario analysis provides reasonable capital estimates and allows for the measurement of the impact of extreme events on banking operations.Keywords: operational risk, scenario analysis, economic capital, loss distribution approach, extreme value theory, stress testing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24231040 Adaptive Few-Shot Deep Metric Learning
Authors: Wentian Shi, Daming Shi, Maysam Orouskhani, Feng Tian
Abstract:
Currently the most prevalent deep learning methods require a large amount of data for training, whereas few-shot learning tries to learn a model from limited data without extensive retraining. In this paper, we present a loss function based on triplet loss for solving few-shot problem using metric based learning. Instead of setting the margin distance in triplet loss as a constant number empirically, we propose an adaptive margin distance strategy to obtain the appropriate margin distance automatically. We implement the strategy in the deep siamese network for deep metric embedding, by utilizing an optimization approach by penalizing the worst case and rewarding the best. Our experiments on image recognition and co-segmentation model demonstrate that using our proposed triplet loss with adaptive margin distance can significantly improve the performance.
Keywords: Few-shot learning, triplet network, adaptive margin, deep learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9041039 Assessment of Path Loss Prediction Models for Wireless Propagation Channels at L-Band Frequency over Different Micro-Cellular Environments of Ekiti State, Southwestern Nigeria
Authors: C. I. Abiodun, S. O. Azi, J. S. Ojo, P. Akinyemi
Abstract:
The design of accurate and reliable mobile communication systems depends majorly on the suitability of path loss prediction methods and the adaptability of the methods to various environments of interest. In this research, the results of the adaptability of radio channel behavior are presented based on practical measurements carried out in the 1800 MHz frequency band. The measurements are carried out in typical urban, suburban and rural environments in Ekiti State, Southwestern part of Nigeria. A total number of seven base stations of MTN GSM service located in the studied environments were monitored. Path loss and break point distances were deduced from the measured received signal strength (RSS) and a practical path loss model is proposed based on the deduced break point distances. The proposed two slope model, regression line and four existing path loss models were compared with the measured path loss values. The standard deviations of each model with respect to the measured path loss were estimated for each base station. The proposed model and regression line exhibited lowest standard deviations followed by the Cost231-Hata model when compared with the Erceg Ericsson and SUI models. Generally, the proposed two-slope model shows closest agreement with the measured values with a mean error values of 2 to 6 dB. These results show that, either the proposed two slope model or Cost 231-Hata model may be used to predict path loss values in mobile micro cell coverage in the well-considered environments. Information from this work will be useful for link design of microwave band wireless access systems in the region.
Keywords: Break-point distances, path loss models, path loss exponent, received signal strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8151038 Low Nonlinear Effects Index-Guiding Nanostructured Photonic Crystal Fiber
Authors: S. Olyaee, M. Seifouri, A. Nikoosohbat, M. Shams Esfand Abadi
Abstract:
Photonic Crystal Fibers (PCFs) can be used in optical communications as transmission lines. For this reason, the PCFs with low confinement loss, low chromatic dispersion, and low nonlinear effects are highly suitable transmission media. In this paper, we introduce a new design of index-guiding nanostructured photonic crystal fiber (IG-NPCF) with ultra-low chromatic dispersion, low nonlinearity effects, and low confinement loss. Relatively low dispersion is achieved in the wavelength range of 1200 to 1600nm using the proposed design. According to the new structure of nanostructured PCF presented in this study, the chromatic dispersion slope is -30(ps/km.nm) and the confinement loss reaches below 10-7 dB/km. While in the wavelength range mentioned above at the same time an effective area of more than 50.2μm2 is obtained.
Keywords: Optical communication systems, nanostructured, index-guiding, dispersion, confinement loss, photonic crystal fiber.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29521037 Highly Accurate Target Motion Compensation Using Entropy Function Minimization
Authors: Amin Aghatabar Roodbary, Mohammad Hassan Bastani
Abstract:
One of the defects of stepped frequency radar systems is their sensitivity to target motion. In such systems, target motion causes range cell shift, false peaks, Signal to Noise Ratio (SNR) reduction and range profile spreading because of power spectrum interference of each range cell in adjacent range cells which induces distortion in High Resolution Range Profile (HRRP) and disrupt target recognition process. Thus Target Motion Parameters (TMPs) effects compensation should be employed. In this paper, such a method for estimating TMPs (velocity and acceleration) and consequently eliminating or suppressing the unwanted effects on HRRP based on entropy minimization has been proposed. This method is carried out in two major steps: in the first step, a discrete search method has been utilized over the whole acceleration-velocity lattice network, in a specific interval seeking to find a less-accurate minimum point of the entropy function. Then in the second step, a 1-D search over velocity is done in locus of the minimum for several constant acceleration lines, in order to enhance the accuracy of the minimum point found in the first step. The provided simulation results demonstrate the effectiveness of the proposed method.
Keywords: ATR, HRRP, motion compensation, SFW, TMP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6521036 Optimization of a Triangular Fin with Variable Fin Base Thickness
Authors: Hyung Suk Kang
Abstract:
A triangular fin with variable fin base thickness is analyzed and optimized using a two-dimensional analytical method. The influence of fin base height and fin base thickness on the temperature in the fin is listed. For the fixed fin volumes, the maximum heat loss, the corresponding optimum fin effectiveness, fin base height and fin tip length as a function of the fin base thickness, convection characteristic number and dimensionless fin volume are represented. One of the results shows that the optimum heat loss increases whereas the corresponding optimum fin effectiveness decreases with the increase of fin volume.Keywords: A triangular fin, Convection characteristic number, Heat loss, Fin base thickness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41211035 Application of Computational Methods Mm2 and Gussian for Studing Unimolecular Decomposition of Vinil Ethers based on the Mechanism of Hydrogen Bonding
Authors: Behnaz Shahrokh, Garnik N. Sargsyan, Arkadi B. Harutyunyan
Abstract:
Investigations of the unimolecular decomposition of vinyl ethyl ether (VEE), vinyl propyl ether (VPE) and vinyl butyl ether (VBE) have shown that activation of the molecule of a ether results in formation of a cyclic construction - the transition state (TS), which may lead to the displacement of the thermodynamic equilibrium towards the reaction products. The TS is obtained by applying energy minimization relative to the ground state of an ether under the program MM2 when taking into account the hydrogen bond formation between a hydrogen atom of alkyl residue and the extreme atom of carbon of the vinyl group. The dissociation of TS up to the products is studied by energy minimization procedure using the mathematical program Gaussian. The obtained calculation data for VEE testify that the decomposition of this ether may be conditioned by hydrogen bond formation for two possible versions: when α- or β- hydrogen atoms of the ethyl group are bound to carbon atom of the vinyl group. Applying the same calculation methods to other ethers (VPE and VBE) it is shown that only in the case of hydrogen bonding between α-hydrogen atom of the alkyl residue and the extreme atom of carbon of the vinyl group (αH---C) results in decay of theses ethers.Keywords: Gaussian, MM2, ethers, TS, decomposition
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12161034 Microstrip Slot Antenna for Triple Band Application in Wireless Communication
Authors: Biplab Bag
Abstract:
In this paper, the design of a coaxial feed single layer rectangular microstrip patch antenna for three different wireless communication band applications is presented. The proposed antenna is designed by using substrate Roger RT/duroid 5880 having permittivity of about 2.2 and tangent loss of 0.0009. The characteristics of the substrate are designed and to evaluate the performance of modeled antenna using HFSS v.11 EM simulator, from Ansoft. The proposed antenna has small in size and operates at 2.25GHz, 3.76GHz and 5.23GHz suitable for mobile satellite service (MSS) network, WiMAX and WLAN applications. The dimension of the patch and slots are optimized to obtain these desired functional frequency ranges. The simulation results with frequency response, radiation pattern and return loss, VSWR, Input Impedance are presented with appropriate table and graph.
Keywords: Microstrip, Tangent Loss, MSS, WiMAX, WLAN, Radiation Pattern, Return Loss, VSWR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31111033 Estimation of the Spent Fuel Pool Water Temperature at a Loss-of-Pool-Cooling Accident
Authors: Chan Hee Park, Arim Lee, Jung Min Lee, Joo Hyun Moon
Abstract:
Accident in spent fuel pool (SFP) of Fukushima Daiichi Unit 4 showed the importance of continuous monitoring of the key environmental parameters such as water temperature, water level, and radiation level in the SFP at accident conditions. Because the SFP water temperature is one of the key parameters indicating SFP conditions, its behavior at accident conditions shall be understood to prepare appropriate measures. This study estimated temporal change in the SFP water temperature at Kori Unit 1 with 587 MWe for 1 hour after initiation of a loss-of-pool-cooling accident. For the estimation, ANSYS CFX 13.0 code was used. The estimation showed that the increasing rate of the water temperature was 3.90C per hour and the SFP water temperature could reach 1000C in 25.6 hours after the initiation of loss-of-pool-cooling accident.
Keywords: Spent fuel pool, water temperature, Kori Unit 1, a loss-of-pool-cooling accident.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26861032 Chinese Event Detection Technique Based on Dependency Parsing and Rule Matching
Authors: Weitao Lin
Abstract:
To quickly extract adequate information from large-scale unstructured text data, this paper studies the representation of events in Chinese scenarios and performs the regularized abstraction. It proposes a Chinese event detection technique based on dependency parsing and rule matching. The method first performs dependency parsing on the original utterance, then performs pattern matching at the word or phrase granularity based on the results of dependent syntactic analysis, filters out the utterances with prominent non-event characteristics, and obtains the final results. The experimental results show the effectiveness of the method.
Keywords: Natural Language Processing, Chinese event detection, rules matching, dependency parsing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1681031 An Hybrid Approach for Loss Reduction in Distribution Systems using Harmony Search Algorithm
Authors: R. Srinivasa Rao
Abstract:
Individually Network reconfiguration or Capacitor control perform well in minimizing power loss and improving voltage profile of the distribution system. But for heavy reactive power loads network reconfiguration and for heavy active power loads capacitor placement can not effectively reduce power loss and enhance voltage profiles in the system. In this paper, an hybrid approach that combine network reconfiguration and capacitor placement using Harmony Search Algorithm (HSA) is proposed to minimize power loss reduction and improve voltage profile. The proposed approach is tested on standard IEEE 33 and 16 bus systems. Computational results show that the proposed hybrid approach can minimize losses more efficiently than Network reconfiguration or Capacitor control. The results of proposed method are also compared with results obtained by Simulated Annealing (SA). The proposed method has outperformed in terms of the quality of solution compared to SA.Keywords: Capacitor Control, Network Reconfiguration, HarmonySearch Algorithm, Loss Reduction, Voltage Profile.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21641030 A Study on Behaviour of Normal Strength Concrete and High Strength Concrete Subjected to Elevated Temperatures
Authors: C. B. K.Rao, Rooban Kumar
Abstract:
Cement concrete is a complex mixture of different materials. Behaviour of concrete depends on its mix proportions and constituents when it is subjected to elevated temperatures. Principal effects due to elevated temperatures are loss in compressive strength, loss in weight or mass, change in colour and spall of concrete. The experimental results of normal concrete and high strength concrete subjected elevated temperatures at 200°C, 400°C, 600°C, and 800°C and different cooling regimes viz. air cooling, water quenching on different grade of concrete are reported in this paper.
Keywords: High strength concrete, Normal strength concrete, Elevated Temperature, Loss of mass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37741029 Probabilistic Robustness Assessment of Structures under Sudden Column-Loss Scenario
Authors: Ali Y Al-Attraqchi, P. Rajeev, M. Javad Hashemi, Riadh Al-Mahaidi
Abstract:
This paper presents a probabilistic incremental dynamic analysis (IDA) of a full reinforced concrete building subjected to column loss scenario for the assessment of progressive collapse. The IDA is chosen to explicitly account for uncertainties in loads and system capacity. Fragility curves are developed to predict the probability of progressive collapse given the loss of one or more columns. At a broader scale, it will also provide critical information needed to support the development of a new generation of design codes that attempt to explicitly quantify structural robustness.
Keywords: Incremental dynamic analysis, progressive collapse, structural engineering, pushdown analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10561028 Effect of Modeling of Hydraulic Form Loss Coefficient to Break on Emergency Core Coolant Bypass
Authors: Young S. Bang, Dong H. Yoon, Seung H. Yoo
Abstract:
Emergency Core Coolant Bypass (ECC Bypass) has been regarded as an important phenomenon to peak cladding temperature of large-break loss-of-coolant-accidents (LBLOCA) in nuclear power plants (NPP). A modeling scheme to address the ECC Bypass phenomena and the calculation of LBLOCA using that scheme are discussed in the present paper. A hydraulic form loss coefficient (HFLC) from the reactor vessel downcomer to the broken cold leg is predicted by the computational fluid dynamics (CFD) code with a variation of the void fraction incoming from the downcomer. The maximum, mean, and minimum values of FLC are derived from the CFD results and are incorporated into the LBLOCA calculation using a system thermal-hydraulic code, MARS-KS. As a relevant parameter addressing the ECC Bypass phenomena, the FLC to the break and its range are proposed.
Keywords: CFD analysis, ECC Bypass, hydraulic form loss coefficient, system thermal-hydraulic code.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8171027 Characterization of Three Photodetector Types for Computed Tomography Dosimetry
Authors: C. M. M. Paschoal, D. do N. Souza, L. A. P. Santos
Abstract:
In this study three commercial semiconductor devices were characterized in the laboratory for computed tomography dosimetry: one photodiode and two phototransistors. It was evaluated four responses to the irradiation: dose linearity, energy dependence, angular dependence and loss of sensitivity after X ray exposure. The results showed that the three devices have proportional response with the air kerma; the energy dependence displayed for each device suggests that some calibration factors would be applied for each one; the angular dependence showed a similar pattern among the three electronic components. In respect to the fourth parameter analyzed, one phototransistor has the highest sensitivity however it also showed the greatest loss of sensitivity with the accumulated dose. The photodiode was the device with the smaller sensitivity to radiation, on the other hand, the loss of sensitivity after irradiation is negligible. Since high accuracy is a desired feature for a dosimeter, the photodiode can be the most suitable of the three devices for dosimetry in tomography. The phototransistors can also be used for CT dosimetry, however it would be necessary a correction factor due to loss of sensitivity with accumulated dose.Keywords: Dosimetry, computed tomography, phototransistor, photodiode
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22731026 Determination of Soil Loss by Erosion in Different Land Covers Categories and Slope Classes in Bovilla Watershed, Tirana, Albania
Authors: Valmir Baloshi, Fran Gjoka, Nehat Çollaku, Elvin Toromani
Abstract:
As a sediment production mechanism, soil erosion is the main environmental threat to the Bovilla watershed, including the decline of water quality of the Bovilla reservoir that provides drinking water to Tirana city (the capital of Albania). Therefore, an experiment with 25 erosion plots for soil erosion monitoring has been set up since June 2017. The aim was to determine the soil loss on plot and watershed scale in Bovilla watershed (Tirana region) for implementation of soil and water protection measures or payments for ecosystem services (PES) programs. The results of erosion monitoring for the period June 2017 - May 2018 showed that the highest values of surface runoff were noted in bare land of 38829.91 liters on slope of 74% and the lowest values in forest land of 12840.6 liters on slope of 64% while the highest values of soil loss were found in bare land of 595.15 t/ha on slope of 62% and lowest values in forest land of 18.99 t/ha on slope of 64%. These values are much higher than the average rate of soil loss in the European Union (2.46 ton/ha/year). In the same sloping class, the soil loss was reduced from orchard or bare land to the forest land, and in the same category of land use, the soil loss increased with increasing land slope. It is necessary to conduct chemical analyses of sediments to determine the amount of chemical elements leached out of the soil and end up in the reservoir of Bovilla. It is concluded that PES programs should be implemented for rehabilitation of sub-watersheds Ranxe, Vilez and Zall-Bastar of the Bovilla watershed with valuable conservation practices.
Keywords: ANOVA, Bovilla, land cover, slope, soil loss, watershed management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8831025 Simulation of Loss-of-Flow Transient in a Radiant Steam Boiler with Relap5/Mod3.2
Authors: A.L.Deghal.Cheridi, A.Chaker, A.Loubar
Abstract:
loss of feedwater accident is one of the frequently sever accidents in steam boiler facilities. It threatens the system structural integrity and generates serious hazards and economic loses. The safety analysis of the thermal installations, based extensively on the numeric simulation. The simulation analysis using realistic computer codes like Relap5/Mod3.2 will help understand steam boiler thermal-hydraulic behavior during normal and abnormal conditions. In this study, we are interested on the evaluation of the radiant steam boiler assessment and response to loss-of-feedwater accident. Pressure, temperature and flow rate profiles are presented in various steam boiler system components. The obtained results demonstrate the importance and capability of the Relap5/Mod3.2 code in the thermal-hydraulic analysis of the steam boiler facilities.
Keywords: Radiant steam boiler, Relap5/Mod3.2 code system, Steady-state simulation, Transient simulation, Loss of feedwateraccident
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22201024 Normalization Discriminant Independent Component Analysis
Authors: Liew Yee Ping, Pang Ying Han, Lau Siong Hoe, Ooi Shih Yin, Housam Khalifa Bashier Babiker
Abstract:
In face recognition, feature extraction techniques attempts to search for appropriate representation of the data. However, when the feature dimension is larger than the samples size, it brings performance degradation. Hence, we propose a method called Normalization Discriminant Independent Component Analysis (NDICA). The input data will be regularized to obtain the most reliable features from the data and processed using Independent Component Analysis (ICA). The proposed method is evaluated on three face databases, Olivetti Research Ltd (ORL), Face Recognition Technology (FERET) and Face Recognition Grand Challenge (FRGC). NDICA showed it effectiveness compared with other unsupervised and supervised techniques.
Keywords: Face recognition, small sample size, regularization, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19511023 Characterization and Modeling of Packet Loss of a VoIP Communication
Authors: L. Estrada, D. Torres, H. Toral
Abstract:
In this work, a characterization and modeling of packet loss of a Voice over Internet Protocol (VoIP) communication is developed. The distributions of the number of consecutive received and lost packets (namely gap and burst) are modeled from the transition probabilities of two-state and four-state model. Measurements show that both models describe adequately the burst distribution, but the decay of gap distribution for non-homogeneous losses is better fit by the four-state model. The respective probabilities of transition between states for each model were estimated with a proposed algorithm from a set of monitored VoIP calls in order to obtain representative minimum, maximum and average values for both models.Keywords: Packet loss, gap and burst distribution, Markovchain, VoIP measurements.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18651022 Parametric Approach for Reserve Liability Estimate in Mortgage Insurance
Authors: Rajinder Singh, Ram Valluru
Abstract:
Chain Ladder (CL) method, Expected Loss Ratio (ELR) method and Bornhuetter-Ferguson (BF) method, in addition to more complex transition-rate modeling, are commonly used actuarial reserving methods in general insurance. There is limited published research about their relative performance in the context of Mortgage Insurance (MI). In our experience, these traditional techniques pose unique challenges and do not provide stable claim estimates for medium to longer term liabilities. The relative strengths and weaknesses among various alternative approaches revolve around: stability in the recent loss development pattern, sufficiency and reliability of loss development data, and agreement/disagreement between reported losses to date and ultimate loss estimate. CL method results in volatile reserve estimates, especially for accident periods with little development experience. The ELR method breaks down especially when ultimate loss ratios are not stable and predictable. While the BF method provides a good tradeoff between the loss development approach (CL) and ELR, the approach generates claim development and ultimate reserves that are disconnected from the ever-to-date (ETD) development experience for some accident years that have more development experience. Further, BF is based on subjective a priori assumption. The fundamental shortcoming of these methods is their inability to model exogenous factors, like the economy, which impact various cohorts at the same chronological time but at staggered points along their life-time development. This paper proposes an alternative approach of parametrizing the loss development curve and using logistic regression to generate the ultimate loss estimate for each homogeneous group (accident year or delinquency period). The methodology was tested on an actual MI claim development dataset where various cohorts followed a sigmoidal trend, but levels varied substantially depending upon the economic and operational conditions during the development period spanning over many years. The proposed approach provides the ability to indirectly incorporate such exogenous factors and produce more stable loss forecasts for reserving purposes as compared to the traditional CL and BF methods.
Keywords: Actuarial loss reserving techniques, logistic regression, parametric function, volatility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4141021 Estimation of the Upper Tail Dependence Coefficient for Insurance Loss Data Using an Empirical Copula-Based Approach
Authors: Adrian O’Hagan, Robert McLoughlin
Abstract:
Considerable focus in the world of insurance risk quantification is placed on modeling loss values from lines of business (LOBs) that possess upper tail dependence. Copulas such as the Joe, Gumbel and Student-t copula may be used for this purpose. The copula structure imparts a desired level of tail dependence on the joint distribution of claims from the different LOBs. Alternatively, practitioners may possess historical or simulated data that already exhibit upper tail dependence, through the impact of catastrophe events such as hurricanes or earthquakes. In these circumstances, it is not desirable to induce additional upper tail dependence when modeling the joint distribution of the loss values from the individual LOBs. Instead, it is of interest to accurately assess the degree of tail dependence already present in the data. The empirical copula and its associated upper tail dependence coefficient are presented in this paper as robust, efficient means of achieving this goal.
Keywords: Empirical copula, extreme events, insurance loss reserving, upper tail dependence coefficient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 48431020 The Kinetic of Biodegradation Lignin in Water Hyacinth (Eichhornia Crassipes) by Phanerochaete Chrysosporium using Solid State Fermentation (SSF) Method for Bioethanol Production, Indonesia
Authors: Eka Sari, Siti Syamsiah, Hary Sulistyo, Muslikhin
Abstract:
Lignocellulosic materials are considered the most abundant renewable resource available for the Bioethanol Production. Water Hyacinth is one of potential raw material of the world-s worst aquatic plant as a feedstock to produce Bioethanol. The purposed this research is obtain reduced of matter for biodegradation lignin in Biological pretreatment with White Rot Fungi eg. Phanerochaete Chrysosporium using Solid state Fermentation methods. Phanerochaete Chrysosporium is known to have the best ability to degraded lignin, but simultaneously it can also degraded cellulose and hemicelulose. During 8 weeks incubation, water hyacinth occurred loss of weight reached 34,67%, while loss of lignin reached 67,21%, loss of cellulose reached 11,01% and loss of hemicellulose reached 36,56%. The kinetic of losses lignin using regression linear plot, the results is obtained constant rate (k) of reduction lignin is -0.1053 and the equation of reduction of lignin is y = wo - 0, 1.53 xKeywords: Biodegradation, lignin, PhanerochaeteChrysosporium, SSF, Water Hyacinth, Bioethanol
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25771019 Image Compression with Back-Propagation Neural Network using Cumulative Distribution Function
Authors: S. Anna Durai, E. Anna Saro
Abstract:
Image Compression using Artificial Neural Networks is a topic where research is being carried out in various directions towards achieving a generalized and economical network. Feedforward Networks using Back propagation Algorithm adopting the method of steepest descent for error minimization is popular and widely adopted and is directly applied to image compression. Various research works are directed towards achieving quick convergence of the network without loss of quality of the restored image. In general the images used for compression are of different types like dark image, high intensity image etc. When these images are compressed using Back-propagation Network, it takes longer time to converge. The reason for this is, the given image may contain a number of distinct gray levels with narrow difference with their neighborhood pixels. If the gray levels of the pixels in an image and their neighbors are mapped in such a way that the difference in the gray levels of the neighbors with the pixel is minimum, then compression ratio as well as the convergence of the network can be improved. To achieve this, a Cumulative distribution function is estimated for the image and it is used to map the image pixels. When the mapped image pixels are used, the Back-propagation Neural Network yields high compression ratio as well as it converges quickly.Keywords: Back-propagation Neural Network, Cumulative Distribution Function, Correlation, Convergence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25491018 Indigenous Dayak People’s Perceptions of Wildlife Loss and Gain Related to Oil Palm Development
Authors: A. Sunkar, A. Saraswati, Y. Santosa
Abstract:
Controversies surrounding the impacts of oil palm plantations have resulted in some heated debates, especially concerning biodiversity loss and indigenous people well-being. The indigenous people of Dayak generally used wildlife to fulfill their daily needs thus were assumed to have experienced negative impacts due to oil palm developments within and surrounding their settlement areas. This study was conducted to identify the characteristics of the Dayak community settled around an oil palm plantation, to determine their perceptions of wildlife loss or gain as the results of the development of oil palm plantations, and to identify the determinant characteristic of the perceptions. The research was conducted on March 2018 in Nanga Tayap and Tajok Kayong Villages, which were located around the oil palm plantation of NTYE of Ketapang, West Kalimantan-Indonesia. Data were collected through in depth-structured interview, using closed and semi-open questionnaires and three-scale Likert statements. Interviews were conducted with 74 respondents using accidental sampling, and categorized into respondents who were dependent on oil palm for their livelihoods and those who were not. Data were analyzed using quantitative statistics method, Likert Scale, Chi-Square Test, Spearman Test, and Mann-Whitney Test. The research found that the indigenous Dayak people were aware of wildlife species loss and gain since the establishment of the plantation. Nevertheless, wildlife loss did not affect their social, economic, and cultural needs since they could find substitutions. It was found that prior to the plantation’s development, the local Dayak communities were already slowly experiencing some livelihood transitions through local village development. The only determinant characteristic of the community that influenced their perceptions of wildlife loss/gain was level of education.
Keywords: Wildlife, oil palm plantations, indigenous Dayak, biodiversity loss and gain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13581017 Soil Mass Loss Reduction during Rainfalls by Reinforcing the Slopes with the Surficial Confinement
Authors: Ramli Nazir, Hossein Moayedi
Abstract:
Soil confinement systems serve as effective solutions to any erosion control project. Various confinements systems, namely triangular, circular and rectangular with the size of 50, 100, and 150 mm, and with a depth of 10 mm, were embedded in soil samples at slope angle of 60°. The observed soil mass losses for the confined soil systems were much smaller than those from unconfined system. As a result, the size of confinement and rainfall intensity have a direct effect on the soil mass loss. The triangular and rectangular confinement systems showed the lowest and highest soil loss masses, respectively. The slopes also failed much faster in the unconfined system than in the confined slope.
Keywords: Erosion control, Soil confinement, Soil erosion, Slope stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18601016 A Shape Optimization Method in Viscous Flow Using Acoustic Velocity and Four-step Explicit Scheme
Authors: Yoichi Hikino, Mutsuto Kawahara
Abstract:
The purpose of this study is to derive optimal shapes of a body located in viscous flows by the finite element method using the acoustic velocity and the four-step explicit scheme. The formulation is based on an optimal control theory in which a performance function of the fluid force is introduced. The performance function should be minimized satisfying the state equation. This problem can be transformed into the minimization problem without constraint conditions by using the adjoint equation with adjoint variables corresponding to the state equation. The performance function is defined by the drag and lift forces acting on the body. The weighted gradient method is applied as a minimization technique, the Galerkin finite element method is used as a spatial discretization and the four-step explicit scheme is used as a temporal discretization to solve the state equation and the adjoint equation. As the interpolation, the orthogonal basis bubble function for velocity and the linear function for pressure are employed. In case that the orthogonal basis bubble function is used, the mass matrix can be diagonalized without any artificial centralization. The shape optimization is performed by the presented method.Keywords: Shape Optimization, Optimal Control Theory, Finite Element Method, Weighted Gradient Method, Fluid Force, Orthogonal Basis Bubble Function, Four-step Explicit Scheme, Acoustic Velocity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1462