Search results for: epidemiological methods
14701 Net Fee and Commission Income Determinants of European Cooperative Banks
Authors: Karolína Vozková, Matěj Kuc
Abstract:
Net fee and commission income is one of the key elements of a bank’s core income. In the current low-interest rate environment, this type of income is gaining importance relative to net interest income. This paper analyses the effects of bank and country specific determinants of net fee and commission income on a set of cooperative banks from European countries in the 2007-2014 period. In order to do that, dynamic panel data methods (system Generalized Methods of Moments) were employed. Subsequently, alternative panel data methods were run as robustness checks of the analysis. Strong positive impact of bank concentration on the share of net fee and commission income was found, which proves that cooperative banks tend to display a higher share of fee income in less competitive markets. This is probably connected with the fact that they stick with their traditional deposit-taking and loan-providing model and fees on these services are driven down by the competitors. Moreover, compared to commercial banks, cooperatives do not expand heavily into non-traditional fee bearing services under competition and their overall fee income share is therefore decreasing with the increased competitiveness of the sector.Keywords: cooperative banking, dynamic panel data models, net fee and commission income, system GMM
Procedia PDF Downloads 32914700 Parallel Asynchronous Multi-Splitting Methods for Differential Algebraic Systems
Authors: Malika Elkyal
Abstract:
We consider an iterative parallel multi-splitting method for differential algebraic equations. The main feature of the proposed idea is to use the asynchronous form. We prove that the multi-splitting technique can effectively accelerate the convergent performance of the iterative process. The main characteristic of an asynchronous mode is that the local algorithm does not have to wait at predetermined messages to become available. We allow some processors to communicate more frequently than others, and we allow the communication delays to be substantial and unpredictable. Accordingly, we note that synchronous algorithms in the computer science sense are particular cases of our formulation of asynchronous one.Keywords: parallel methods, asynchronous mode, multisplitting, differential algebraic equations
Procedia PDF Downloads 55614699 Analysis of Diabetes Patients Using Pearson, Cost Optimization, Control Chart Methods
Authors: Devatha Kalyan Kumar, R. Poovarasan
Abstract:
In this paper, we have taken certain important factors and health parameters of diabetes patients especially among children by birth (pediatric congenital) where using the above three metrics methods we are going to assess the importance of each attributes in the dataset and thereby determining the most highly responsible and co-related attribute causing diabetics among young patients. We use cost optimization, control chart and Spearmen methodologies for the real-time application of finding the data efficiency in this diabetes dataset. The Spearmen methodology is the correlation methodologies used in software development process to identify the complexity between the various modules of the software. Identifying the complexity is important because if the complexity is higher, then there is a higher chance of occurrence of the risk in the software. With the use of control; chart mean, variance and standard deviation of data are calculated. With the use of Cost optimization model, we find to optimize the variables. Hence we choose the Spearmen, control chart and cost optimization methods to assess the data efficiency in diabetes datasets.Keywords: correlation, congenital diabetics, linear relationship, monotonic function, ranking samples, pediatric
Procedia PDF Downloads 25414698 Modern Trends in Foreign Direct Investments in Georgia
Authors: Rusudan Kinkladze, Guguli Kurashvili, Ketevan Chitaladze
Abstract:
Foreign direct investment is a driving force in the development of the interdependent national economies, and the study and analysis of investments is an urgent problem. It is particularly important for transitional economies, such as Georgia, and the study and analysis of investments is an urgent problem. Consequently, the goal of the research is the study and analysis of direct foreign investments in Georgia, and identification and forecasting of modern trends, and covers the period of 2006-2015. The study uses the methods of statistical observation, grouping and analysis, the methods of analytical indicators of time series, trend identification and the predicted values are calculated, as well as various literary and Internet sources relevant to the research. The findings showed that modern investment policy In Georgia is favorable for domestic as well as foreign investors. Georgia is still a net importer of investments. In 2015, the top 10 investing countries was led by Azerbaijan, United Kingdom and Netherlands, and the largest share of FDIs were allocated in the transport and communication sector; the financial sector was the second, followed by the health and social work sector, and the same trend will continue in the future.Keywords: foreign direct investments, methods, statistics, analysis
Procedia PDF Downloads 32914697 Data Mining in Medicine Domain Using Decision Trees and Vector Support Machine
Authors: Djamila Benhaddouche, Abdelkader Benyettou
Abstract:
In this paper, we used data mining to extract biomedical knowledge. In general, complex biomedical data collected in studies of populations are treated by statistical methods, although they are robust, they are not sufficient in themselves to harness the potential wealth of data. For that you used in step two learning algorithms: the Decision Trees and Support Vector Machine (SVM). These supervised classification methods are used to make the diagnosis of thyroid disease. In this context, we propose to promote the study and use of symbolic data mining techniques.Keywords: biomedical data, learning, classifier, algorithms decision tree, knowledge extraction
Procedia PDF Downloads 55614696 Isolate-Specific Variations among Clinical Isolates of Brucella Identified by Whole-Genome Sequencing, Bioinformatics and Comparative Genomics
Authors: Abu S. Mustafa, Mohammad W. Khan, Faraz Shaheed Khan, Nazima Habibi
Abstract:
Brucellosis is a zoonotic disease of worldwide prevalence. There are at least four species and several strains of Brucella that cause human disease. Brucella genomes have very limited variation across strains, which hinder strain identification using classical molecular techniques, including PCR and 16 S rDNA sequencing. The aim of this study was to perform whole genome sequencing of clinical isolates of Brucella and perform bioinformatics and comparative genomics analyses to determine the existence of genetic differences across the isolates of a single Brucella species and strain. The draft sequence data were generated from 15 clinical isolates of Brucella melitensis (biovar 2 strain 63/9) using MiSeq next generation sequencing platform. The generated reads were used for further assembly and analysis. All the analysis was performed using Bioinformatics work station (8 core i7 processor, 8GB RAM with Bio-Linux operating system). FastQC was used to determine the quality of reads and low quality reads were trimmed or eliminated using Fastx_trimmer. Assembly was done by using Velvet and ABySS softwares. The ordering of assembled contigs was performed by Mauve. An online server RAST was employed to annotate the contigs assembly. Annotated genomes were compared using Mauve and ACT tools. The QC score for DNA sequence data, generated by MiSeq, was higher than 30 for 80% of reads with more than 100x coverage, which suggested that data could be utilized for further analysis. However when analyzed by FastQC, quality of four reads was not good enough for creating a complete genome draft so remaining 11 samples were used for further analysis. The comparative genome analyses showed that despite sharing same gene sets, single nucleotide polymorphisms and insertions/deletions existed across different genomes, which provided a variable extent of diversity to these bacteria. In conclusion, the next generation sequencing, bioinformatics, and comparative genome analysis can be utilized to find variations (point mutations, insertions and deletions) across different genomes of Brucella within a single strain. This information could be useful in surveillance and epidemiological studies supported by Kuwait University Research Sector grants MI04/15 and SRUL02/13.Keywords: brucella, bioinformatics, comparative genomics, whole genome sequencing
Procedia PDF Downloads 38014695 Predication Model for Leukemia Diseases Based on Data Mining Classification Algorithms with Best Accuracy
Authors: Fahd Sabry Esmail, M. Badr Senousy, Mohamed Ragaie
Abstract:
In recent years, there has been an explosion in the rate of using technology that help discovering the diseases. For example, DNA microarrays allow us for the first time to obtain a "global" view of the cell. It has great potential to provide accurate medical diagnosis, to help in finding the right treatment and cure for many diseases. Various classification algorithms can be applied on such micro-array datasets to devise methods that can predict the occurrence of Leukemia disease. In this study, we compared the classification accuracy and response time among eleven decision tree methods and six rule classifier methods using five performance criteria. The experiment results show that the performance of Random Tree is producing better result. Also it takes lowest time to build model in tree classifier. The classification rules algorithms such as nearest- neighbor-like algorithm (NNge) is the best algorithm due to the high accuracy and it takes lowest time to build model in classification.Keywords: data mining, classification techniques, decision tree, classification rule, leukemia diseases, microarray data
Procedia PDF Downloads 31914694 Towards an Intelligent Ontology Construction Cost Estimation System: Using BIM and New Rules of Measurement Techniques
Authors: F. H. Abanda, B. Kamsu-Foguem, J. H. M. Tah
Abstract:
Construction cost estimation is one of the most important aspects of construction project design. For generations, the process of cost estimating has been manual, time-consuming and error-prone. This has partly led to most cost estimates to be unclear and riddled with inaccuracies that at times lead to over- or under-estimation of construction cost. The development of standard set of measurement rules that are understandable by all those involved in a construction project, have not totally solved the challenges. Emerging Building Information Modelling (BIM) technologies can exploit standard measurement methods to automate cost estimation process and improves accuracies. This requires standard measurement methods to be structured in ontologically and machine readable format; so that BIM software packages can easily read them. Most standard measurement methods are still text-based in textbooks and require manual editing into tables or Spreadsheet during cost estimation. The aim of this study is to explore the development of an ontology based on New Rules of Measurement (NRM) commonly used in the UK for cost estimation. The methodology adopted is Methontology, one of the most widely used ontology engineering methodologies. The challenges in this exploratory study are also reported and recommendations for future studies proposed.Keywords: BIM, construction projects, cost estimation, NRM, ontology
Procedia PDF Downloads 54914693 Stress Analysis of Vertebra Using Photoelastic and Finite Element Methods
Authors: Jamal A. Hassan, Ali Q. Abdulrazzaq, Sadiq J. Abass
Abstract:
In this study, both the photoelastic, as well as the finite element methods, are used to study the stress distribution within human vertebra (L4) under forces similar to those that occur during normal life. Two & three dimensional models of vertebra were created by the software AutoCAD. The coordinates obtained were fed into a computer numerical control (CNC) tensile machine to fabricate the models from photoelastic sheets. Completed models were placed in a transmission polariscope and loaded with static force (up to 1500N). Stresses can be quantified and localized by counting the number of fringes. In both methods the Principle stresses were calculated at different regions. The results noticed that the maximum von-mises stress on the area of the extreme superior vertebral body surface and the facet surface with high normal stress (σ) and shear stress (τ). The facets and other posterior elements have a load-bearing function to help support the weight of the upper body and anything that it carries, and are also acted upon by spinal muscle forces. The numerical FE results have been compared with the experimental method using photoelasticity which shows good agreement between experimental and simulation results.Keywords: photoelasticity, stress, load, finite element
Procedia PDF Downloads 28514692 Curvature Based-Methods for Automatic Coarse and Fine Registration in Dimensional Metrology
Authors: Rindra Rantoson, Hichem Nouira, Nabil Anwer, Charyar Mehdi-Souzani
Abstract:
Multiple measurements by means of various data acquisition systems are generally required to measure the shape of freeform workpieces for accuracy, reliability and holisticity. The obtained data are aligned and fused into a common coordinate system within a registration technique involving coarse and fine registrations. Standardized iterative methods have been established for fine registration such as Iterative Closest Points (ICP) and its variants. For coarse registration, no conventional method has been adopted yet despite a significant number of techniques which have been developed in the literature to supply an automatic rough matching between data sets. Two main issues are addressed in this paper: the coarse registration and the fine registration. For coarse registration, two novel automated methods based on the exploitation of discrete curvatures are presented: an enhanced Hough Transformation (HT) and an improved Ransac Transformation. The use of curvature features in both methods aims to reduce computational cost. For fine registration, a new variant of ICP method is proposed in order to reduce registration error using curvature parameters. A specific distance considering the curvature similarity has been combined with Euclidean distance to define the distance criterion used for correspondences searching. Additionally, the objective function has been improved by combining the point-to-point (P-P) minimization and the point-to-plane (P-Pl) minimization with automatic weights. These ones are determined from the preliminary calculated curvature features at each point of the workpiece surface. The algorithms are applied on simulated and real data performed by a computer tomography (CT) system. The obtained results reveal the benefit of the proposed novel curvature-based registration methods.Keywords: discrete curvature, RANSAC transformation, hough transformation, coarse registration, ICP variant, point-to-point and point-to-plane minimization combination, computer tomography
Procedia PDF Downloads 42314691 KCBA, A Method for Feature Extraction of Colonoscopy Images
Authors: Vahid Bayrami Rad
Abstract:
In recent years, the use of artificial intelligence techniques, tools, and methods in processing medical images and health-related applications has been highlighted and a lot of research has been done in this regard. For example, colonoscopy and diagnosis of colon lesions are some cases in which the process of diagnosis of lesions can be improved by using image processing and artificial intelligence algorithms, which help doctors a lot. Due to the lack of accurate measurements and the variety of injuries in colonoscopy images, the process of diagnosing the type of lesions is a little difficult even for expert doctors. Therefore, by using different software and image processing, doctors can be helped to increase the accuracy of their observations and ultimately improve their diagnosis. Also, by using automatic methods, the process of diagnosing the type of disease can be improved. Therefore, in this paper, a deep learning framework called KCBA is proposed to classify colonoscopy lesions which are composed of several methods such as K-means clustering, a bag of features and deep auto-encoder. Finally, according to the experimental results, the proposed method's performance in classifying colonoscopy images is depicted considering the accuracy criterion.Keywords: colorectal cancer, colonoscopy, region of interest, narrow band imaging, texture analysis, bag of feature
Procedia PDF Downloads 5214690 Domain Adaptive Dense Retrieval with Query Generation
Authors: Rui Yin, Haojie Wang, Xun Li
Abstract:
Recently, mainstream dense retrieval methods have obtained state-of-the-art results on some datasets and tasks. However, they require large amounts of training data, which is not available in most domains. The severe performance degradation of dense retrievers on new data domains has limited the use of dense retrieval methods to only a few domains with large training datasets. In this paper, we propose an unsupervised domain-adaptive approach based on query generation. First, a generative model is used to generate relevant queries for each passage in the target corpus, and then, the generated queries are used for mining negative passages. Finally, the query-passage pairs are labeled with a cross-encoder and used to train a domain-adapted dense retriever. We also explore contrastive learning as a method for training domain-adapted dense retrievers and show that it leads to strong performance in various retrieval settings. Experiments show that our approach is more robust than previous methods in target domains that require less unlabeled data.Keywords: dense retrieval, query generation, contrastive learning, unsupervised training
Procedia PDF Downloads 10114689 Mathematical Model of Corporate Bond Portfolio and Effective Border Preview
Authors: Sergey Podluzhnyy
Abstract:
One of the most important tasks of investment and pension fund management is building decision support system which helps to make right decision on corporate bond portfolio formation. Today there are several basic methods of bond portfolio management. They are duration management, immunization and convexity management. Identified methods have serious disadvantage: they do not take into account credit risk or insolvency risk of issuer. So, identified methods can be applied only for management and evaluation of high-quality sovereign bonds. Applying article proposes mathematical model for building an optimal in case of risk and yield corporate bond portfolio. Proposed model takes into account the default probability in formula of assessment of bonds which results to more correct evaluation of bonds prices. Moreover, applied model provides tools for visualization of the efficient frontier of corporate bonds portfolio taking into account the exposure to credit risk, which will increase the quality of the investment decisions of portfolio managers.Keywords: corporate bond portfolio, default probability, effective boundary, portfolio optimization task
Procedia PDF Downloads 31714688 Evolution of Performance Measurement Methods in Conditions of Uncertainty: The Implementation of Fuzzy Sets in Performance Measurement
Authors: E. A. Tkachenko, E. M. Rogova, V. V. Klimov
Abstract:
One of the basic issues of development management is connected with performance measurement as a prerequisite for identifying the achievement of development objectives. The aim of our research is to develop an improved model of assessing a company’s development results. The model should take into account the cyclical nature of development and the high degree of uncertainty in dealing with numerous management tasks. Our hypotheses may be formulated as follows: Hypothesis 1. The cycle of a company’s development may be studied from the standpoint of a project cycle. To do that, methods and tools of project analysis are to be used. Hypothesis 2. The problem of the uncertainty when justifying managerial decisions within the framework of a company’s development cycle can be solved through the use of the mathematical apparatus of fuzzy logic. The reasoned justification of the validity of the hypotheses made is given in the suggested article. The fuzzy logic toolkit applies to the case of technology shift within an enterprise. It is proven that some restrictions in performance measurement that are incurred to conventional methods could be eliminated by implementation of the fuzzy logic apparatus in performance measurement models.Keywords: logic, fuzzy sets, performance measurement, project analysis
Procedia PDF Downloads 38114687 The Effect of Exercise Therapy and Electroacupuncture on Some Clinical Outcomes in People with Post Total Hip Arthroplasty
Authors: Marzieh Yassin, Masoud Rashed, Soheil Mansour Sohani, Reza Salehi
Abstract:
Background: Hip arthroplasty is one of the surgical methods to improve symptoms in patients with hip osteoarthritis. The use of electroacupuncture and TENS reduces pain, increases range of motion and improves performance. Methods: In this clinical trial study, 30 patients after hip arthroplasty were randomly divided into two groups: electroacupuncture (n=16) with exercise therapy and TENS with exercise therapy (n=14). Severity of pain, quality of life, range of motion, edema and function were evaluated in two groups before and after the interventions. Interventions of 10 sessions (three sessions per week) were conducted for two groups. The significance level in all tests was below 0.05. Results: The results showed that both groups improved all of the symptoms after the intervention (p≤0.05), although there was no statistically significant difference between the two groups in terms of effectiveness (p≥0.05). Conclusion: The results showed that both methods improve symptoms in patients after surgery. According to this study, electroacupuncture is suggested as a new method effective for the treatment of people with post-Total Hip Arthroplasty.Keywords: electroacupuncture, physical performance, total hip arthroplasty, TENS
Procedia PDF Downloads 8214686 Coffee Consumption Has No Acute Effects on Glucose Metabolism in Healthy Men: A Randomized Crossover Clinical Trial
Authors: Caio E. G. Reis, Sara Wassell, Adriana L. Porto, Angélica A. Amato, Leslie J. C. Bluck, Teresa H. M. da Costa
Abstract:
Background: Multiple epidemiologic studies have consistently reported association between increased coffee consumption and a lowered risk of Type 2 Diabetes Mellitus. However, the mechanisms behind this finding have not been fully elucidated. Objective: We investigate the effect of coffee (caffeinated and decaffeinated) on glucose effectiveness and insulin sensitivity using the stable isotope minimal model protocol with oral glucose administration in healthy men. Design: Fifteen healthy men underwent 5 arms randomized crossover single-blinding (researchers) clinical trial. They consumed decaffeinated coffee, caffeinated coffee (with and without sugar), and controls – water (with and without sugar) followed 1 hour by an oral glucose tolerance test (75 g of available carbohydrate) with intravenous labeled dosing interpreted by the two compartment minimal model (225 minutes). One-way ANOVA with Bonferroni adjustment were used to compare the effects of the tested beverages on glucose metabolism parameters. Results: Decaffeinated coffee resulted in 29% and 85% higher insulin sensitivity compared with caffeinated coffee and water, respectively, and the caffeinated coffee showed 15% and 60% higher glucose effectiveness compared with decaffeinated coffee and water, respectively. However, these differences were not significant (p > 0.10). In overall analyze (0 – 225 min) there were no significant differences on glucose effectiveness, insulin sensitivity, and glucose and insulin area under the curve between the groups. The beneficial effects of coffee did not seem to act in the short-term (hours) on glucose metabolism parameters mainly on insulin sensitivity indices. The benefits of coffee consumption occur in the long-term (years) as has been shown in the reduction of Type 2 Diabetes Mellitus risk in epidemiological studies. The clinical relevance of the present findings is that there is no need to avoid coffee as the drink choice for healthy people. Conclusions: The findings of this study demonstrate that the consumption of caffeinated and decaffeinated coffee with or without sugar has no acute effects on glucose metabolism in healthy men. Further researches, including long-term interventional studies, are needed to fully elucidate the mechanisms behind the coffee effects on reduced risk for Type 2 Diabetes Mellitus.Keywords: coffee, diabetes mellitus type 2, glucose, insulin
Procedia PDF Downloads 43614685 An Assessment of Different Blade Tip Timing (BTT) Algorithms Using an Experimentally Validated Finite Element Model Simulator
Authors: Mohamed Mohamed, Philip Bonello, Peter Russhard
Abstract:
Blade Tip Timing (BTT) is a technology concerned with the estimation of both frequency and amplitude of rotating blades. A BTT system comprises two main parts: (a) the arrival time measurement system, and (b) the analysis algorithms. Simulators play an important role in the development of the analysis algorithms since they generate blade tip displacement data from the simulated blade vibration under controlled conditions. This enables an assessment of the performance of the different algorithms with respect to their ability to accurately reproduce the original simulated vibration. Such an assessment is usually not possible with real engine data since there is no practical alternative to BTT for blade vibration measurement. Most simulators used in the literature are based on a simple spring-mass-damper model to determine the vibration. In this work, a more realistic experimentally validated simulator based on the Finite Element (FE) model of a bladed disc (blisk) is first presented. It is then used to generate the necessary data for the assessment of different BTT algorithms. The FE modelling is validated using both a hammer test and two firewire cameras for the mode shapes. A number of autoregressive methods, fitting methods and state-of-the-art inverse methods (i.e. Russhard) are compared. All methods are compared with respect to both synchronous and asynchronous excitations with both single and simultaneous frequencies. The study assesses the applicability of each method for different conditions of vibration, amount of sampling data, and testing facilities, according to its performance and efficiency under these conditions.Keywords: blade tip timing, blisk, finite element, vibration measurement
Procedia PDF Downloads 30514684 Quantum Modelling of AgHMoO4, CsHMoO4 and AgCsMoO4 Chemistry in the Field of Nuclear Power Plant Safety
Authors: Mohamad Saab, Sidi Souvi
Abstract:
In a major nuclear accident, the released fission products (FPs) and the structural materials are likely to influence the transport of iodine in the reactor coolant system (RCS) of a pressurized water reactor (PWR). So far, the thermodynamic data on cesium and silver species used to estimate the magnitude of FP release show some discrepancies, data are scarce and not reliable. For this reason, it is crucial to review the thermodynamic values related to cesium and silver materials. To this end, we have used state-of-the-art quantum chemical methods to compute the formation enthalpies and entropies of AgHMoO₄, CsHMoO₄, and AgCsMoO₄ in the gas phase. Different quantum chemical methods have been investigated (DFT and CCSD(T)) in order to predict the geometrical parameters and the energetics including the correlation energy. The geometries were optimized with TPSSh-5%HF method, followed by a single point calculation of the total electronic energies using the CCSD(T) wave function method. We thus propose with a final uncertainty of about 2 kJmol⁻¹ standard enthalpies of formation of AgHMoO₄, CsHMoO₄, and AgCsMoO₄.Keywords: nuclear accident, ASTEC code, thermochemical database, quantum chemical methods
Procedia PDF Downloads 18814683 Effect of Blanching and Drying Methods on the Degradation Kinetics and Color Stability of Radish (Raphanus sativus) Leaves
Authors: K. Radha Krishnan, Mirajul Alom
Abstract:
Dehydrated powder prepared from fresh radish (Raphanus sativus) leaves were investigated for the color stability by different drying methods (tray, sun and solar). The effect of blanching conditions, drying methods as well as drying temperatures (50 – 90°C) were considered for studying the color degradation kinetics of chlorophyll in the dehydrated powder. The hunter color parameters (L*, a*, b*) and total color difference (TCD) were determined in order to investigate the color degradation kinetics of chlorophyll. Blanching conditions, drying method and drying temperature influenced the changes in L*, a*, b* and TCD values. The changes in color values during processing were described by a first order kinetic model. The temperature dependence of chlorophyll degradation was adequately modeled by Arrhenius equation. To predict the losses in green color, a mathematical model was developed from the steady state kinetic parameters. The results from this study indicated the protective effect of blanching conditions on the color stability of dehydrated radish powder.Keywords: chlorophyll, color stability, degradation kinetics, drying
Procedia PDF Downloads 39814682 Islanding Detection of Wind Turbine by Rate of Change of Frequency (ROCOF) and Rate of change of Power (ROCOP) Method
Authors: Vipulkumar Jagodana
Abstract:
Recently the use of renewable sources has increased, these sources include fuel cell, photo voltaic, and wind turbine. Islanding occurs when one portion of grid is isolated from remaining grid. Use of the renewable sources can provide continuous power to isolated portion in islanding condition. One of the common renewable sources is wind generation using wind turbine. The efficiency of wind generation can be increased in combination with conventional sources. When islanding occurs, few parameters change which may be frequency, voltage, active power, and harmonics. According to large change in one of these parameters islanding is detected. In this paper, two passive methods Rate of Change of Frequency (ROCOF) and Rate of change of Power (ROCOP) have been implemented for islanding detection of small wind-turbine. Islanding detection of both methods have been simulated in PSCAD. Simulation results show at different islanding inception angle response of ROCOF and ROCOP.Keywords: islanding, adopted methods, PSCAD simulation, comparison
Procedia PDF Downloads 22414681 Application Reliability Method for Concrete Dams
Authors: Mustapha Kamel Mihoubi, Mohamed Essadik Kerkar
Abstract:
Probabilistic risk analysis models are used to provide a better understanding of the reliability and structural failure of works, including when calculating the stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of reliability analysis methods including the methods used in engineering. It is in our case, the use of level 2 methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type first order risk method (FORM) and the second order risk method (SORM). By way of comparison, a level three method was used which generates a full analysis of the problem and involves an integration of the probability density function of random variables extended to the field of security using the Monte Carlo simulation method. Taking into account the change in stress following load combinations: normal, exceptional and extreme acting on the dam, calculation of the results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities, thus causing a significant decrease in strength, shear forces then induce a shift that threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case the increase of uplift in a hypothetical default of the drainage system.Keywords: dam, failure, limit-state, monte-carlo, reliability, probability, simulation, sliding, taylor
Procedia PDF Downloads 32414680 Evaluation of Three Digital Graphical Methods of Baseflow Separation Techniques in the Tekeze Water Basin in Ethiopia
Authors: Alebachew Halefom, Navsal Kumar, Arunava Poddar
Abstract:
The purpose of this work is to specify the parameter values, the base flow index (BFI), and to rank the methods that should be used for base flow separation. Three different digital graphical approaches are chosen and used in this study for the purpose of comparison. The daily time series discharge data were collected from the site for a period of 30 years (1986 up to 2015) and were used to evaluate the algorithms. In order to separate the base flow and the surface runoff, daily recorded streamflow (m³/s) data were used to calibrate procedures and get parameter values for the basin. Additionally, the performance of the model was assessed by the use of the standard error (SE), the coefficient of determination (R²), and the flow duration curve (FDC) and baseflow indexes. The findings indicate that, in general, each strategy can be used worldwide to differentiate base flow; however, the Sliding Interval Method (SIM) performs significantly better than the other two techniques in this basin. The average base flow index was calculated to be 0.72 using the local minimum method, 0.76 using the fixed interval method, and 0.78 using the sliding interval method, respectively.Keywords: baseflow index, digital graphical methods, streamflow, Emba Madre Watershed
Procedia PDF Downloads 7614679 Mathematical Reconstruction of an Object Image Using X-Ray Interferometric Fourier Holography Method
Authors: M. K. Balyan
Abstract:
The main principles of X-ray Fourier interferometric holography method are discussed. The object image is reconstructed by the mathematical method of Fourier transformation. The three methods are presented – method of approximation, iteration method and step by step method. As an example the complex amplitude transmission coefficient reconstruction of a beryllium wire is considered. The results reconstructed by three presented methods are compared. The best results are obtained by means of step by step method.Keywords: dynamical diffraction, hologram, object image, X-ray holography
Procedia PDF Downloads 39314678 Numerical Modelling of Crack Initiation around a Wellbore Due to Explosion
Authors: Meysam Lak, Mohammad Fatehi Marji, Alireza Yarahamdi Bafghi, Abolfazl Abdollahipour
Abstract:
A wellbore is a hole that is drilled to aid in the exploration and recovery of natural resources including oil and gas. Occasionally, in order to increase productivity index and porosity of the wellbore and reservoir, the well stimulation methods have been used. Hydraulic fracturing is one of these methods. Moreover, several explosions at the end of the well can stimulate the reservoir and create fractures around it. In this study, crack initiation in rock around the wellbore has been numerically modeled due to explosion. One, two, three, and four pairs of explosion have been set at the end of the wellbore on its wall. After each stage of the explosion, results have been presented and discussed. Results show that this method can initiate and probably propagate several fractures around the wellbore.Keywords: crack initiation, explosion, finite difference modelling, well productivity
Procedia PDF Downloads 28814677 Integrated Grey Rational Analysis-Standard Deviation Method for Handover in Heterogeneous Networks
Authors: Mohanad Alhabo, Naveed Nawaz, Mahmoud Al-Faris
Abstract:
The dense deployment of small cells is a promising solution to enhance the coverage and capacity of the heterogeneous networks (HetNets). However, the unplanned deployment could bring new challenges to the network ranging from interference, unnecessary handovers and handover failures. This will cause a degradation in the quality of service (QoS) delivered to the end user. In this paper, we propose an integrated Grey Rational Analysis Standard Deviation based handover method (GRA-SD) for HetNet. The proposed method integrates the Standard Deviation (SD) technique to acquire the weight of the handover metrics and the GRA method to select the best handover base station. The performance of the GRA-SD method is evaluated and compared with the traditional Multiple Attribute Decision Making (MADM) methods including Simple Additive Weighting (SAW) and VIKOR methods. Results reveal that the proposed method has outperformed the other methods in terms of minimizing the number of frequent unnecessary handovers and handover failures, in addition to improving the energy efficiency.Keywords: energy efficiency, handover, HetNets, MADM, small cells
Procedia PDF Downloads 11514676 Analysis and Comparison of Prototypes of an Ergometric Step in a Multidisciplinary Design Process
Authors: M. B. Ricardo De Oliveira, A. Borghi-Silva, L. Di Thommazo, D. Braatz
Abstract:
Prototypes can be understood as representations of a product concept. Furthermore, prototyping consists in an important stage in product development and results in better team communication, decision making, testing and problem solving through feedback. Although there are several methods of prototyping suggested by recent studies for designers to choose from, some methods present different advantages, such as cost and time reduction, performance and fidelity, which should be taken in account during a product development project. In this multidisciplinary study, involving areas of physiotherapy, engineering and computer science (hardware and software), we compared four developed prototypes of an ergometric step: a virtual prototype, a 3D printed prototype, a bricolage prototype and a prototype manufactured by a third-party company. These prototypes were evaluated in a comparative-qualitative approach for their contribution to the concept’s maturation of the product, the different prototyping methods used and the advantages and disadvantages of each one based on the product’s design specifications (performance, safety, materials, cost, maintenance, usability, ergonomics and portability). Our results indicated that despite prototypes show overall advantages, all of them have limitations, thus being crucial to have different methods of testing and interacting with the product. Additionally, virtual and 3D printed prototypes were essential at early stages of the project due to their low-cost and high-fidelity representation of the product, while the prototype manufactured by a third-party company and bricolage prototype introduced functional tests in real scenarios, allowing more detailed evaluations. This study also resulted in a patent for an ergometric step.Keywords: Product Design, Product Development, Prototypes, Step
Procedia PDF Downloads 11514675 The Impact of Total Dust (LGS) and Mineral Dust (PM 10) in Cardio Vascular and Respiratory System, in Albania: A Longitudinal Study
Authors: Canga Mimoza, Irene Malagnino, Giulia Malagnino, Vito Malagnino
Abstract:
Aim: This study aims at evaluating the impact of total dust (LGS) and mineral dust (PM10), in the cardio vascular and respiratory systems. Also proving that these air polluters are the cause of several diseases, such as bronchopneumonia, pneumonia, bronchitis, angina pectoris and cardiac insufficiency. Material and Method: The study is concentrated in the cities of Fier and Vlora. This is a clinic-epidemiological study conducted during the time period 2014-2019. Some of the data of LGS and PM10 were obtained from the database of the Institute of Public Health. The formula to measure the mean value of LGS and PM10 is ∆X=X (mean)-Xᵢ. Results: Based on the calculations made, we noticed that: The mean value of LGS in the city of Fieri was 227,33, while the mean value of LGS in the city of Vlora was 177,4. Whereas, the mean value of PM10 in the city of Fieri was 105.5 and the mean value of PM10 in the city of Vlore was 77.5. According to, our statistics the values of LGS were 1.2 times higher in Fier than in Vlora and the PM10 values were 1.36 times higher in Fier than in Vlora. Based on the data, in the city of Fier, the incidence of the bronchopneumonia was 56.53 sick patients/1000 inhabitants, but in Vlora, it was 22 sick patients/1000 inhabitants, so the number of the sick patients was 2.5 times higher in the city of Fieri compared with Vlora city, (P=0.001). The number of the patients with bronchitis, in the city of Fier, was 18 patients/1000 inhabitants, whereas, in Vlora, it was 9 patients/1000 inhabitants, (P=0.005). Based on the data, 8 patients/1000 inhabitants in the city of Fier, suffered from the pneumonia disease, while in Vlora city, were 4 patients/1000 inhabitants, (P=0.005). Another disease taken in consideration was angina pectoris. This study can claim that in the city of Fier, 9.5 patients/1000 inhabitants suffered from this disease, while in Vlora city, were only 4 patients /1000 inhabitants, (P=0.001). Findings of the present study proved that 3.7 patients/1000 inhabitants in the city of Fieri, had cardiac insufficiency, whereas in the city of Vlora, were 1.8 patients/1000 inhabitants, (P=0.05). Conclusions: LGS and PM10 have an influential impact on the cardio vascular and respiratory system; that’s why their levels should be kept under control. The pollution levels are 1.2 and 1.4 times higher in Fier than in Vlora; also the incidences of the diseases are 2 times higher in Fier than in Vlora. Recommendations: In order to prevent the cardio vascular and respiratory diseases, we should avoid places where pollution is higher than the norm. This can be achieved by frequenting places where the air pollution is lower, such as parks, gardens, top floors, etc.Keywords: impact of total dust, LGS, mineral dust, PM 10, cardio vascular pathologies, respiratory disease
Procedia PDF Downloads 12614674 A Review of Masonry Buildings Restrengthening Methods
Authors: Negar Sartipzadeh
Abstract:
The historic buildings are generally the ones which have been built by materials like brick, mud, stone, and wood. Some phenomena such as severe earthquakes can be tremendously detrimental to the structures, imposing serious effects and losses on such structures. Hence, it matters a lot to ascertain safety and reliability of the structures under such circumstances. It has been asserted that the major reason for the collapse of Unreinforced Masonry (URM) in various earthquakes is the incapability of resisting the forces and vice versa because such URMs are meant for the gravity load and they fail to withstand the shear forces inside the plate and the bending forces outside the plate. For this reason, restrengthening such structures is a key factor in lowering the seismic loss in developing countries. Seismic reinforcement of the historic buildings with regard to their cultural value on one hand, and exhaustion and damage of many of the structural elements on the other hand, have brought in restricting factors which necessitate the seismic reinforcement methods meant for such buildings to be maximally safe, non-destructive, effective, and non-obvious. Henceforth, it is pinpointed that making use of diverse technologies such as active controlling, Energy dampers, and seismic separators besides the current popular methods would be justifiable for such buildings, notwithstanding their high imposed costs.Keywords: masonry buildings, seismic reinforcement, Unreinforced Masonry (URM), earthquake
Procedia PDF Downloads 27814673 A Mixed Methods Study Aimed at Exploring the Conceptualization of Orthorexia Nervosa on Instagram
Authors: Elena V. Syurina, Sophie Renckens, Martina Valente
Abstract:
Objective: The objective of this study was to investigate the nature of the conversation around orthorexia nervosa (ON) on Instagram. Methods: The present study was conducted using mixed methods, combining a concurrent triangulation and sequential explanatory design. First, 3027 pictures posted on Instagram using #Orthorexia were analyzed. Then, a questionnaire about Instagram use related to ON was completed entirely by 185 respondents. These two quantitative data sources were statistically analyzed and triangulated afterwards. Finally, 9 interviews were conducted, to more deeply investigate what is being said about ON on Instagram and what the motivations to post about it are. Results: Four main categories of pictures were found to be represented in Instagram posts about ON: ‘food’, ‘people’, ‘text’, and ‘other.’ Savory and unprocessed food was most highly represented within the food category, and pictures of people were mostly pictures of the account holder. People who self-identify as having ON were more likely to post about ON, and they were significantly more likely to post about ‘food’, ‘people’ and ‘text.’ The goal of the posts was to raise awareness around ON, as well as to provide support for people who believe to be suffering from it. Conclusion: Since the conversation around ON on Instagram is supportive, it could be beneficial to consider Instagram use in the treatment of ON. However, more research is needed on a larger scale.Keywords: orthorexia nervosa, Instagram, social media, disordered eating
Procedia PDF Downloads 13714672 The Influence of High Temperatures on HVFA Concrete Columns by NDT Methods
Authors: D. Jagath Kumari, K. Srinivasa Rao
Abstract:
Quality assurance of the structures subjected to high temperatures is now enforcing measure for the Structural Engineers. The existing relations between strength and nondestructive measurements have been established under normal conditions are not suitable to concretes that have been exposed to high temperatures. The scope of the work is to investigate the influence of high temperatures of short durations on the residual properties of reinforced HVFA concrete columns that affect the strength by non-destructive tests (NDT). Fly ash concrete is increasingly used in the design of normal strength, high strength and high performance concretes. In this paper, the authors revealed the influence of high temperatures on HVFA concrete columns. These columns are heated from 100oC to 800oC with increments of 100oC and allowed to cool to room temperature by two methods one is air cooling method and the other immediate water quenching method. All the specimens were tested identically, before heating and after heating for compressive strength and material integrity by rebound hammer and ultrasonic pulse velocity (UPV) meter respectively. HVFA concrete retained more residual strength by water quenching method than air-cooling method.Keywords: HVFA concrete, NDT methods, residual strength, non-destructive tests
Procedia PDF Downloads 455