Search results for: interval estimation
2045 Using Support Vector Machines for Measuring Democracy
Authors: Tommy Krieger, Klaus Gruendler
Abstract:
We present a novel approach for measuring democracy, which enables a very detailed and sensitive index. This method is based on Support Vector Machines, a mathematical algorithm for pattern recognition. Our implementation evaluates 188 countries in the period between 1981 and 2011. The Support Vector Machines Democracy Index (SVMDI) is continuously on the 0-1-Interval and robust to variations in the numerical process parameters. The algorithm introduced here can be used for every concept of democracy without additional adjustments, and due to its flexibility it is also a valuable tool for comparison studies.Keywords: democracy, democracy index, machine learning, support vector machines
Procedia PDF Downloads 3782044 An Approach for Detection Efficiency Determination of High Purity Germanium Detector Using Cesium-137
Authors: Abdulsalam M. Alhawsawi
Abstract:
Estimation of a radiation detector's efficiency plays a significant role in calculating the activity of radioactive samples. Detector efficiency is measured using sources that emit a variety of energies from low to high-energy photons along the energy spectrum. Some photon energies are hard to find in lab settings either because check sources are hard to obtain or the sources have short half-lives. This work aims to develop a method to determine the efficiency of a High Purity Germanium Detector (HPGe) based on the 662 keV gamma ray photon emitted from Cs-137. Cesium-137 is readily available in most labs with radiation detection and health physics applications and has a long half-life of ~30 years. Several photon efficiencies were calculated using the MCNP5 simulation code. The simulated efficiency of the 662 keV photon was used as a base to calculate other photon efficiencies in a point source and a Marinelli Beaker form. In the Marinelli Beaker filled with water case, the efficiency of the 59 keV low energy photons from Am-241 was estimated with a 9% error compared to the MCNP5 simulated efficiency. The 1.17 and 1.33 MeV high energy photons emitted by Co-60 had errors of 4% and 5%, respectively. The estimated errors are considered acceptable in calculating the activity of unknown samples as they fall within the 95% confidence level.Keywords: MCNP5, MonteCarlo simulations, efficiency calculation, absolute efficiency, activity estimation, Cs-137
Procedia PDF Downloads 1162043 Effect of the Workpiece Position on the Manufacturing Tolerances
Authors: Rahou Mohamed , Sebaa Fethi, Cheikh Abdelmadjid
Abstract:
Manufacturing tolerancing is intended to determine the intermediate geometrical and dimensional states of the part during its manufacturing process. These manufacturing dimensions also serve to satisfy not only the functional requirements given in the definition drawing but also the manufacturing constraints, for example geometrical defects of the machine, vibration, and the wear of the cutting tool. The choice of positioning has an important influence on the cost and quality of manufacture. To avoid this problem, a two-step approach have been developed. The first step is dedicated to the determination of the optimum position. As for the second step, a study was carried out for the tightening effect on the tolerance interval.Keywords: dispersion, tolerance, manufacturing, position
Procedia PDF Downloads 3362042 Fracture and Fatigue Crack Growth Analysis and Modeling
Authors: Volkmar Nolting
Abstract:
Fatigue crack growth prediction has become an important topic in both engineering and non-destructive evaluation. Crack propagation is influenced by the mechanical properties of the material and is conveniently modelled by the Paris-Erdogan equation. The critical crack size and the total number of load cycles are calculated. From a Larson-Miller plot the maximum operational temperature can for a given stress level be determined so that failure does not occur within a given time interval t. The study is used to determine a reasonable inspection cycle and thus enhances operational safety and reduces costs.Keywords: fracturemechanics, crack growth prediction, lifetime of a component, structural health monitoring
Procedia PDF Downloads 492041 Calculation of Instrumental Results of the Tohoku Earthquake, Japan (Mw 9.0) on March 11, 2011 and Other Destructive Earthquakes during Seismic Hazard Assessment
Authors: J. K. Karapetyan
Abstract:
In this paper seismological-statistical analysis of actual instrumental data on the main tremor of the Great Japan earthquake 11.03.2011 is implemented for finding out the dependence between maximal values of peak ground accelerations (PGA) and epicentric distances. A number of peculiarities of manifestation of accelerations' maximum values at the interval of long epicentric distances are revealed which do not correspond with current scales of seismic intensity.Keywords: earthquakes, instrumental records, seismic hazard, Japan
Procedia PDF Downloads 3652040 Three-Dimensional CFD Modeling of Flow Field and Scouring around Bridge Piers
Authors: P. Deepak Kumar, P. R. Maiti
Abstract:
In recent years, sediment scour near bridge piers and abutment is a serious problem which causes nationwide concern because it has resulted in more bridge failures than other causes. Scour is the formation of scour hole around the structure mounted on and embedded in erodible channel bed due to the erosion of soil by flowing water. The formation of scour hole around the structures depends upon shape and size of the pier, depth of flow as well as angle of attack of flow and sediment characteristics. The flow characteristics around these structures change due to man-made obstruction in the natural flow path which changes the kinetic energy of the flow around these structures. Excessive scour affects the stability of the foundation of the structure by the removal of the bed material. The accurate estimation of scour depth around bridge pier is very difficult. The foundation of bridge piers have to be taken deeper and to provide sufficient anchorage length required for stability of the foundation. In this study, computational model simulations using a 3D Computational Fluid Dynamics (CFD) model were conducted to examine the mechanism of scour around a cylindrical pier. Subsequently, the flow characteristics around these structures are presented for different flow conditions. Mechanism of scouring phenomenon, the formation of vortex and its consequent effect is discussed for a straight channel. Effort was made towards estimation of scour depth around bridge piers under different flow conditions.Keywords: bridge pier, computational fluid dynamics, multigrid, pier shape, scour
Procedia PDF Downloads 2962039 Poverty Dynamics in Thailand: Evidence from Household Panel Data
Authors: Nattabhorn Leamcharaskul
Abstract:
This study aims to examine determining factors of the dynamics of poverty in Thailand by using panel data of 3,567 households in 2007-2017. Four techniques of estimation are employed to analyze the situation of poverty across households and time periods: the multinomial logit model, the sequential logit model, the quantile regression model, and the difference in difference model. Households are categorized based on their experiences into 5 groups, namely chronically poor, falling into poverty, re-entering into poverty, exiting from poverty and never poor households. Estimation results emphasize the effects of demographic and socioeconomic factors as well as unexpected events on the economic status of a household. It is found that remittances have positive impact on household’s economic status in that they are likely to lower the probability of falling into poverty or trapping in poverty while they tend to increase the probability of exiting from poverty. In addition, not only receiving a secondary source of household income can raise the probability of being a never poor household, but it also significantly increases household income per capita of the chronically poor and falling into poverty households. Public work programs are recommended as an important tool to relieve household financial burden and uncertainty and thus consequently increase a chance for households to escape from poverty.Keywords: difference in difference, dynamic, multinomial logit model, panel data, poverty, quantile regression, remittance, sequential logit model, Thailand, transfer
Procedia PDF Downloads 1122038 Determining Water Use Efficiency of Mung Bean (Vigna radiata L.) under Arid Climatic Conditions
Authors: Awais Ahmad, Mostafa Muhammad Selim, Ali Abdullah Alderfasi
Abstract:
Water limitation is undoubtedly a critical environmental constraint limiting the crop production under arid and semiarid areas. Mung bean is susceptible to both drought and water logging stresses. Therefore, present study was conducted to assess the water deficit stress consequences of yield components and water use efficiency in Mung bean. A field experiment was conducted at Educational Farm, Crop Production Department, College of Food and Agricultural Sciences, Kind Saud University, Saudi Arabia. Trail comprised of four irrigation levels — total amount of irrigation divided into irrigation intervals — (3, 5, 7 and 9 days interval) and three Mung bean genotypes; Kawmay-1, VC-2010 and King from Egypt, Thailand and China respectively. Experiment was arranged under split plot design having irrigation as main while genotype as subplot treatment, and replicated thrice. Plant height, 100 seed weight, biological yield, seed yield, harvest index and water use efficiency were recorded at harvesting. Results revealed that decrease in irrigation have significantly hampered all the studied parameters. Mung bean genotypes have also shown significant differences for all parameters, whereas irrigation genotype interaction was highly significant for seed yield, harvest index and water use efficiency (WUE) while it was significant for biological yield. Plant height and 100 seed weight were recorded non-significant for irrigation genotype interaction. A statistically highly significant correlation among recorded parameters was observed. Minimum irrigation interval (3 days) significantly produced maximum values while VC-2010 comparatively performed better under low irrigation levels. It was concluded that Mung bean may be successfully adopted under Saudi Arabian climate but it needs high water or frequent irrigation, however, genotypic differences are a hope to develop some improved varieties with high water use efficiency.Keywords: mung bean, irrigation intervals, water use efficiency, genotypes, yield
Procedia PDF Downloads 2742037 Instant Location Detection of Objects Moving at High Speed in C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
The practical efficient approach is suggested to estimate the high-speed objects instant bounds in C-OTDR monitoring systems. In case of super-dynamic objects (trains, cars) is difficult to obtain the adequate estimate of the instantaneous object localization because of estimation lag. In other words, reliable estimation coordinates of monitored object requires taking some time for data observation collection by means of C-OTDR system, and only if the required sample volume will be collected the final decision could be issued. But it is contrary to requirements of many real applications. For example, in rail traffic management systems we need to get data off the dynamic objects localization in real time. The way to solve this problem is to use the set of statistical independent parameters of C-OTDR signals for obtaining the most reliable solution in real time. The parameters of this type we can call as 'signaling parameters' (SP). There are several the SP’s which carry information about dynamic objects instant localization for each of C-OTDR channels. The problem is that some of these parameters are very sensitive to dynamics of seismoacoustic emission sources but are non-stable. On the other hand, in case the SP is very stable it becomes insensitive as a rule. This report contains describing the method for SP’s co-processing which is designed to get the most effective dynamic objects localization estimates in the C-OTDR monitoring system framework.Keywords: C-OTDR-system, co-processing of signaling parameters, high-speed objects localization, multichannel monitoring systems
Procedia PDF Downloads 4702036 Developing Allometric Equations for More Accurate Aboveground Biomass and Carbon Estimation in Secondary Evergreen Forests, Thailand
Authors: Titinan Pothong, Prasit Wangpakapattanawong, Stephen Elliott
Abstract:
Shifting cultivation is an indigenous agricultural practice among upland people and has long been one of the major land-use systems in Southeast Asia. As a result, fallows and secondary forests have come to cover a large part of the region. However, they are increasingly being replaced by monocultures, such as corn cultivation. This is believed to be a main driver of deforestation and forest degradation, and one of the reasons behind the recurring winter smog crisis in Thailand and around Southeast Asia. Accurate biomass estimation of trees is important to quantify valuable carbon stocks and changes to these stocks in case of land use change. However, presently, Thailand lacks proper tools and optimal equations to quantify its carbon stocks, especially for secondary evergreen forests, including fallow areas after shifting cultivation and smaller trees with a diameter at breast height (DBH) of less than 5 cm. Developing new allometric equations to estimate biomass is urgently needed to accurately estimate and manage carbon storage in tropical secondary forests. This study established new equations using a destructive method at three study sites: approximately 50-year-old secondary forest, 4-year-old fallow, and 7-year-old fallow. Tree biomass was collected by harvesting 136 individual trees (including coppiced trees) from 23 species, with a DBH ranging from 1 to 31 cm. Oven-dried samples were sent for carbon analysis. Wood density was calculated from disk samples and samples collected with an increment borer from 79 species, including 35 species currently missing from the Global Wood Densities database. Several models were developed, showing that aboveground biomass (AGB) was strongly related to DBH, height (H), and wood density (WD). Including WD in the model was found to improve the accuracy of the AGB estimation. This study provides insights for reforestation management, and can be used to prepare baseline data for Thailand’s carbon stocks for the REDD+ and other carbon trading schemes. These may provide monetary incentives to stop illegal logging and deforestation for monoculture.Keywords: aboveground biomass, allometric equation, carbon stock, secondary forest
Procedia PDF Downloads 2842035 Monte Carlo and Biophysics Analysis in a Criminal Trial
Authors: Luca Indovina, Carmela Coppola, Carlo Altucci, Riccardo Barberi, Rocco Romano
Abstract:
In this paper a real court case, held in Italy at the Court of Nola, in which a correct physical description, conducted with both a Monte Carlo and biophysical analysis, would have been sufficient to arrive at conclusions confirmed by documentary evidence, is considered. This will be an example of how forensic physics can be useful in confirming documentary evidence in order to reach hardly questionable conclusions. This was a libel trial in which the defendant, Mr. DS (Defendant for Slander), had falsely accused one of his neighbors, Mr. OP (Offended Person), of having caused him some damages. The damages would have been caused by an external plaster piece that would have detached from the neighbor’s property and would have hit Mr DS while he was in his garden, much more than a meter far away from the facade of the building from which the plaster piece would have detached. In the trial, Mr. DS claimed to have suffered a scratch on his forehead, but he never showed the plaster that had hit him, nor was able to tell from where the plaster would have arrived. Furthermore, Mr. DS presented a medical certificate with a diagnosis of contusion of the cerebral cortex. On the contrary, the images of Mr. OP’s security cameras do not show any movement in the garden of Mr. DS in a long interval of time (about 2 hours) around the time of the alleged accident, nor do they show any people entering or coming out from the house of Mr. DS in the same interval of time. Biophysical analysis shows that both the diagnosis of the medical certificate and the wound declared by the defendant, already in conflict with each other, are not compatible with the fall of external plaster pieces too small to be found. The wind was at a level 1 of the Beaufort scale, that is, unable to raise even dust (level 4 of the Beaufort scale). Therefore, the motion of the plaster pieces can be described as a projectile motion, whereas collisions with the building cornice can be treated using Newtons law of coefficients of restitution. Numerous numerical Monte Carlo simulations show that the pieces of plaster would not have been able to reach even the garden of Mr. DS, let alone a distance over 1.30 meters. Results agree with the documentary evidence (images of Mr. OP’s security cameras) that Mr. DS could not have been hit by plaster pieces coming from Mr. OP’s property.Keywords: biophysics analysis, Monte Carlo simulations, Newton’s law of restitution, projectile motion
Procedia PDF Downloads 1302034 The Role of Human Capital in the Evolution of Inequality and Economic Growth in Latin-America
Authors: Luis Felipe Brito-Gaona, Emma M. Iglesias
Abstract:
There is a growing literature that studies the main determinants and drivers of inequality and economic growth in several countries, using panel data and different estimation methods (fixed effects, Generalized Methods of Moments (GMM) and Two Stages Least Squares (TSLS)). Recently, it was studied the evolution of these variables in the period 1980-2009 in the 18 countries of Latin-America and it was found that one of the main variables that explained their evolution was Foreign Direct Investment (FDI). We extend this study to the year 2015 in the same 18 countries in Latin-America, and we find that FDI does not have a significant role anymore, while we find a significant negative and positive effect of schooling levels on inequality and economic growth respectively. We also find that the point estimates associated with human capital are the largest ones of the variables included in the analysis, and this means that an increase in human capital (measured by schooling levels of secondary education) is the main determinant that can help to reduce inequality and to increase economic growth in Latin-America. Therefore, we advise that economic policies in Latin-America should be directed towards increasing the level of education. We use the methodologies of estimating by fixed effects, GMM and TSLS to check the robustness of our results. Our conclusion is the same regardless of the estimation method we choose. We also find that the international recession in the Latin-American countries in 2008 reduced significantly their economic growth.Keywords: economic growth, human capital, inequality, Latin-America
Procedia PDF Downloads 2262033 Application of KL Divergence for Estimation of Each Metabolic Pathway Genes
Authors: Shohei Maruyama, Yasuo Matsuyama, Sachiyo Aburatani
Abstract:
The development of the method to annotate unknown gene functions is an important task in bioinformatics. One of the approaches for the annotation is The identification of the metabolic pathway that genes are involved in. Gene expression data have been utilized for the identification, since gene expression data reflect various intracellular phenomena. However, it has been difficult to estimate the gene function with high accuracy. It is considered that the low accuracy of the estimation is caused by the difficulty of accurately measuring a gene expression. Even though they are measured under the same condition, the gene expressions will vary usually. In this study, we proposed a feature extraction method focusing on the variability of gene expressions to estimate the genes' metabolic pathway accurately. First, we estimated the distribution of each gene expression from replicate data. Next, we calculated the similarity between all gene pairs by KL divergence, which is a method for calculating the similarity between distributions. Finally, we utilized the similarity vectors as feature vectors and trained the multiclass SVM for identifying the genes' metabolic pathway. To evaluate our developed method, we applied the method to budding yeast and trained the multiclass SVM for identifying the seven metabolic pathways. As a result, the accuracy that calculated by our developed method was higher than the one that calculated from the raw gene expression data. Thus, our developed method combined with KL divergence is useful for identifying the genes' metabolic pathway.Keywords: metabolic pathways, gene expression data, microarray, Kullback–Leibler divergence, KL divergence, support vector machines, SVM, machine learning
Procedia PDF Downloads 4032032 Specification Requirements for a Combined Dehumidifier/Cooling Panel: A Global Scale Analysis
Authors: Damien Gondre, Hatem Ben Maad, Abdelkrim Trabelsi, Frédéric Kuznik, Joseph Virgone
Abstract:
The use of a radiant cooling solution would enable to lower cooling needs which is of great interest when the demand is initially high (hot climate). But, radiant systems are not naturally compatibles with humid climates since a low-temperature surface leads to condensation risks as soon as the surface temperature is close to or lower than the dew point temperature. A radiant cooling system combined to a dehumidification system would enable to remove humidity for the space, thereby lowering the dew point temperature. The humidity removal needs to be especially effective near the cooled surface. This requirement could be fulfilled by a system using a single desiccant fluid for the removal of both excessive heat and moisture. This task aims at providing an estimation of the specification requirements of such system in terms of cooling power and dehumidification rate required to fulfill comfort issues and to prevent any condensation risk on the cool panel surface. The present paper develops a preliminary study on the specification requirements, performances and behavior of a combined dehumidifier/cooling ceiling panel for different operating conditions. This study has been carried using the TRNSYS software which allows nodal calculations of thermal systems. It consists of the dynamic modeling of heat and vapor balances of a 5m x 3m x 2.7m office space. In a first design estimation, this room is equipped with an ideal heating, cooling, humidification and dehumidification system so that the room temperature is always maintained in between 21◦C and 25◦C with a relative humidity in between 40% and 60%. The room is also equipped with a ventilation system that includes a heat recovery heat exchanger and another heat exchanger connected to a heat sink. Main results show that the system should be designed to meet a cooling power of 42W.m−2 and a desiccant rate of 45 gH2O.h−1. In a second time, a parametric study of comfort issues and system performances has been achieved on a more realistic system (that includes a chilled ceiling) under different operating conditions. It enables an estimation of an acceptable range of operating conditions. This preliminary study is intended to provide useful information for the system design.Keywords: dehumidification, nodal calculation, radiant cooling panel, system sizing
Procedia PDF Downloads 1752031 Estimating the Receiver Operating Characteristic Curve from Clustered Data and Case-Control Studies
Authors: Yalda Zarnegarnia, Shari Messinger
Abstract:
Receiver operating characteristic (ROC) curves have been widely used in medical research to illustrate the performance of the biomarker in correctly distinguishing the diseased and non-diseased groups. Correlated biomarker data arises in study designs that include subjects that contain same genetic or environmental factors. The information about correlation might help to identify family members at increased risk of disease development, and may lead to initiating treatment to slow or stop the progression to disease. Approaches appropriate to a case-control design matched by family identification, must be able to accommodate both the correlation inherent in the design in correctly estimating the biomarker’s ability to differentiate between cases and controls, as well as to handle estimation from a matched case control design. This talk will review some developed methods for ROC curve estimation in settings with correlated data from case control design and will discuss the limitations of current methods for analyzing correlated familial paired data. An alternative approach using Conditional ROC curves will be demonstrated, to provide appropriate ROC curves for correlated paired data. The proposed approach will use the information about the correlation among biomarker values, producing conditional ROC curves that evaluate the ability of a biomarker to discriminate between diseased and non-diseased subjects in a familial paired design.Keywords: biomarker, correlation, familial paired design, ROC curve
Procedia PDF Downloads 2392030 Nitrification and Denitrification Kinetic Parameters of a Mature Sanitary Landfill Leachate
Authors: Tânia F. C. V. Silva, Eloísa S. S. Vieira, João Pinto da Costa, Rui A. R. Boaventura, Vitor J. P. Vilar
Abstract:
Sanitary landfill leachates are characterized as a complex mixture of diverse organic and inorganic contaminants, which are usually removed by combining different treatment processes. Due to its simplicity, reliability, high cost-effectiveness and high nitrogen content (mostly under the ammonium form) inherent in this type of effluent, the activated sludge biological process is almost always applied in leachate treatment plants (LTPs). The purpose of this work is to assess the effect of the main nitrification and denitrification variables on the nitrogen's biological removal, from mature leachates. The leachate samples were collected after an aerated lagoon, at a LTP nearby Porto, presenting a high amount of dissolved organic carbon (1.0-1.3 g DOC/L) and ammonium nitrogen (1.1-1.7 g NH4+-N/L). The experiments were carried out in a 1-L lab-scale batch reactor, equipped with a pH, temperature and dissolved oxygen (DO) control system, in order to determine the reaction kinetic constants at unchanging conditions. The nitrification reaction rate was evaluated while varying the (i) operating temperature (15, 20, 25 and 30ºC), (ii) DO concentration interval (0.5-1.0, 1.0-2.0 and 2.0-4.0 mg/L) and (iii) solution pH (not controlled, 7.5-8.5 and 6.5-7.5). At the beginning of most assays, it was verified that the ammonium stripping occurred simultaneously to the nitrification, reaching up to 37% removal of total dissolved nitrogen. The denitrification kinetic constants and the methanol consumptions were calculated for different values of (i) volatile suspended solids (VSS) content (25, 50 and 100 mL of centrifuged sludge in 1 L solution), (ii) pH interval (6.5-7.0, 7.5-8.0 and 8.5-9.0) and (iii) temperature (15, 20, 25 and 30ºC), using effluent previously nitrified. The maximum nitrification rate obtained was 38±2 mg NH4+-N/h/g VSS (25ºC, 0.5-1.0 mg O2/L, pH not controlled), consuming 4.4±0.3 mg CaCO3/mg NH4+-N. The highest denitrification rate achieved was 19±1 mg (NO2--N+NO3--N)/h/g VSS (30ºC, 50 mL of sludge and pH between 7.5 and 8.0), with a C/N consumption ratio of 1.1±0.1 mg CH3OH/mg (NO2--N+NO3--N) and an overall alkalinity production of 3.7±0.3 mg CaCO3/mg (NO2--N+NO3--N). The denitrification process showed to be sensitive to all studied parameters, while the nitrification reaction did not suffered significant change when DO content was changed.Keywords: mature sanitary landfill leachate, nitrogen removal, nitrification and denitrification parameters, lab-scale activated sludge biological reactor
Procedia PDF Downloads 2762029 Bayesian Inference for High Dimensional Dynamic Spatio-Temporal Models
Authors: Sofia M. Karadimitriou, Kostas Triantafyllopoulos, Timothy Heaton
Abstract:
Reduced dimension Dynamic Spatio-Temporal Models (DSTMs) jointly describe the spatial and temporal evolution of a function observed subject to noise. A basic state space model is adopted for the discrete temporal variation, while a continuous autoregressive structure describes the continuous spatial evolution. Application of such a DSTM relies upon the pre-selection of a suitable reduced set of basic functions and this can present a challenge in practice. In this talk, we propose an online estimation method for high dimensional spatio-temporal data based upon DSTM and we attempt to resolve this issue by allowing the basis to adapt to the observed data. Specifically, we present a wavelet decomposition in order to obtain a parsimonious approximation of the spatial continuous process. This parsimony can be achieved by placing a Laplace prior distribution on the wavelet coefficients. The aim of using the Laplace prior, is to filter wavelet coefficients with low contribution, and thus achieve the dimension reduction with significant computation savings. We then propose a Hierarchical Bayesian State Space model, for the estimation of which we offer an appropriate particle filter algorithm. The proposed methodology is illustrated using real environmental data.Keywords: multidimensional Laplace prior, particle filtering, spatio-temporal modelling, wavelets
Procedia PDF Downloads 4272028 Rupture Termination of the 1950 C. E. Earthquake and Recurrent Interval of Great Earthquake in North Eastern Himalaya, India
Authors: Rao Singh Priyanka, Jayangondaperumal R.
Abstract:
The Himalayan active fault has the potential to generate great earthquakes in the future, posing a biggest existential threat to humans in the Himalayan and adjacent region. Quantitative evaluation of accumulated and released interseismic strain is crucial to assess the magnitude and spatio-temporal variability of future great earthquakes along the Himalayan arc. To mitigate the destruction and hazards associated with such earthquakes, it is important to understand their recurrence cycle. The eastern Himalayan and Indo-Burman plate boundary systems offers an oblique convergence across two orthogonal plate boundaries, resulting in a zone of distributed deformation both within and away from the plate boundary and clockwise rotation of fault-bounded blocks. This seismically active region has poorly documented historical archive of the past large earthquakes. Thus, paleoseismologicalstudies confirm the surface rupture evidences of the great continental earthquakes (Mw ≥ 8) along the Himalayan Frontal Thrust (HFT), which along with the Geodetic studies, collectively provide the crucial information to understand and assess the seismic potential. These investigations reveal the rupture of 3/4th of the HFT during great events since medieval time but with debatable opinions for the timing of events due to unclear evidences, ignorance of transverse segment boundaries, and lack of detail studies. Recent paleoseismological investigations in the eastern Himalaya and Mishmi ranges confirms the primary surface ruptures of the 1950 C.E. great earthquake (M>8). However, a seismic gap exists between the 1714 C.E. and 1950 C.E. Assam earthquakes that did not slip since 1697 C.E. event. Unlike the latest large blind 2015 Gorkha earthquake (Mw 7.8), the 1950 C.E. event is not triggered by a large event of 1947 C.E. that occurred near the western edge of the great upper Assam event. Moreover, the western segment of the eastern Himalayadid not witness any surface breaking earthquake along the HFT for over the past 300 yr. The frontal fault excavations reveal that during the 1950 earthquake, ~3.1-m-high scarp along the HFT was formed due to the co-seismic slip of 5.5 ± 0.7 m at Pasighat in the Eastern Himalaya and a 10-m-high-scarp at a Kamlang Nagar along the Mishmi Thrust in the Eastern Himalayan Syntaxis is an outcome of a dip-slip displacement of 24.6 ± 4.6 m along a 25 ± 5°E dipping fault. This event has ruptured along the two orthogonal fault systems in the form of oblique thrust fault mechanism. Approx. 130 km west of Pasighat site, the Himebasti village has witnessed two earthquakes, the historical 1697 Sadiya earthquake, and the 1950 event, with a cumulative dip-slip displacement of 15.32 ± 4.69 m. At Niglok site, Arunachal Pradesh, a cumulative slip of ~12.82 m during at least three events since pre 19585 B.P. has produced ~6.2-m high scarp while the youngest scarp of ~2.4-m height has been produced during 1697 C.E. The site preserves two deformational events along the eastern HFT, providing an idea of last serial ruptures at an interval of ~850 yearswhile the successive surface rupturing earthquakes lacks in the Mishmi Range to estimate the recurrence cycle.Keywords: paleoseismology, surface rupture, recurrence interval, Eastern Himalaya
Procedia PDF Downloads 842027 Challenging the Standard 24 Equal Quarter Tones Theory in Arab Music: A Case Study of Tetrachords Bayyātī and ḤIjāz
Authors: Nabil Shair
Abstract:
Arab music maqām (Arab modal framework) is founded, among other main characteristics, on microtonal intervals. Notwithstanding the importance and multifaceted nature of intonation in Arab music, there is a paucity of studies examining this subject based on scientific and quantitative approaches. The present-day theory concerning the Arab tone system is largely based on the pioneering treatise of Mīkhā’īl Mashāqah (1840), which proposes the theoretical division of the octave into 24 equal quarter tones. This kind of equal-tempered division is incompatible with the performance practice of Arab music, as many professional Arab musicians conceptualize additional pitches beyond the standard 24 notes per octave. In this paper, we refute the standard theory presenting the scale of well-tempered quarter tones by implementing a quantitative analysis of the performed intonation of two prominent tetrachords in Arab music, namely bayyātī and ḥijāz. This analysis was conducted with the help of advanced computer programs, such as Sonic Visualiser and Tony, by which we were able to obtain precise frequency data (Hz) of each tone every 0.01 second. As a result, the value (in cents) of all three intervals of each tetrachord was measured and accordingly compared to the theoretical intervals. As a result, a specific distribution of a range of deviation from the equal-tempered division of the octave was detected, especially the detection of a diminished first interval of bayyātí and diminished second interval of ḥijāz. These types of intonation entail a considerable amount of flexibility, mainly influenced by surrounding tones, direction and function of the measured tone, ornaments, text, personal style of the performer and interaction with the audience. This paper seeks to contribute to the existing literature dealing with intonation in Arab music, as it is a vital part of the performance practice of this musical tradition. In addition, the insights offered by this paper and its novel methodology might also contribute to the broadening of the existing pedagogic methods used to teach Arab music.Keywords: Arab music, intonation, performance practice, music theory, oral music, octave division, tetrachords, music of the middle east, music history, musical intervals
Procedia PDF Downloads 532026 Assessment of DNA Degradation Using Comet Assay: A Versatile Technique for Forensic Application
Authors: Ritesh K. Shukla
Abstract:
Degradation of biological samples in terms of macromolecules (DNA, RNA, and protein) are the major challenges in the forensic investigation which misleads the result interpretation. Currently, there are no precise methods available to circumvent this problem. Therefore, at the preliminary level, some methods are urgently needed to solve this issue. In this order, Comet assay is one of the most versatile, rapid and sensitive molecular biology technique to assess the DNA degradation. This technique helps to assess DNA degradation even at very low amount of sample. Moreover, the expedient part of this method does not require any additional process of DNA extraction and isolation during DNA degradation assessment. Samples directly embedded on agarose pre-coated microscopic slide and electrophoresis perform on the same slide after lysis step. After electrophoresis microscopic slide stained by DNA binding dye and observed under fluorescent microscope equipped with Komet software. With the help of this technique extent of DNA degradation can be assessed which can help to screen the sample before DNA fingerprinting, whether it is appropriate for DNA analysis or not. This technique not only helps to assess degradation of DNA but many other challenges in forensic investigation such as time since deposition estimation of biological fluids, repair of genetic material from degraded biological sample and early time since death estimation could also be resolved. With the help of this study, an attempt was made to explore the application of well-known molecular biology technique that is Comet assay in the field of forensic science. This assay will open avenue in the field of forensic research and development.Keywords: comet assay, DNA degradation, forensic, molecular biology
Procedia PDF Downloads 1552025 Prognosis of Patients with COVID-19 and Hematologic Malignancies
Authors: Elizabeth Behrens, Anne Timmermann, Alexander Yerkan, Joshua Thomas, Deborah Katz, Agne Paner, Melissa Larson, Shivi Jain, Seo-Hyun Kim, Celalettin Ustun, Ankur Varma, Parameswaran Venugopal, Jamile Shammo
Abstract:
Coronavirus Disease-2019 (COVID-19) causes persistent concern for poor outcomes in vulnerable populations. Patients with hematologic malignancies (HM) have been found to have higher COVID-19 case fatality rates compared to those without malignancy. While cytopenias are common in patients with HM, especially in those undergoing chemotherapy treatment, hemoglobin (Hgb) and platelet count have not yet been studied, to our best knowledge, as potential prognostic indicators for patients with HM and COVID-19. The goal of this study is to identify factors that may increase the risk of mortality in patients with HM and COVID-19. In this single-center, retrospective, observational study, 65 patients with HM and laboratory confirmed COVID-19 were identified between March 2020 and January 2021. Information on demographics, laboratory data the day of COVID-19 diagnosis, and prognosis was extracted from the electronic medical record (EMR), chart reviewed, and analyzed using the statistical software SAS version 9.4. Chi-square testing was used for categorical variable analyses. Risk factors associated with mortality were established by logistic regression models. Non-Hodgkin lymphoma (37%), chronic lymphocytic leukemia (20%), and plasma cell dyscrasia (15%) were the most common HM. Higher Hgb level upon COVID-19 diagnosis was related to decreased mortality, odd ratio=0.704 (95% confidence interval [CI]: 0.511-0.969; P = .0263). Platelet count the day of COVID-19 diagnosis was lower in patients who ultimately died (mean 127 ± 72K/uL, n=10) compared to patients who survived (mean 197 ±92K/uL, n=55) (P=.0258). Female sex was related to decreased mortality, odd ratio=0.143 (95% confidence interval [CI]: 0.026-0.785; P = .0353). There was no mortality difference between the patients who were on treatment for HM the day of COVID-19 diagnosis compared to those who were not (P=1.000). Lower Hgb and male sex are independent risk factors associated with increased mortality of HM patients with COVID-19. Clinicians should be especially attentive to patients with HM and COVID-19 who present with cytopenias. Larger multi-center studies are urgently needed to further investigate the impact of anemia, thrombocytopenia, and demographics on outcomes of patients with hematologic malignancies diagnosed with COVID-19.Keywords: anemia, COVID-19, hematologic malignancy, prognosis
Procedia PDF Downloads 1492024 Estimation of Normalized Glandular Doses Using a Three-Layer Mammographic Phantom
Authors: Kuan-Jen Lai, Fang-Yi Lin, Shang-Rong Huang, Yun-Zheng Zeng, Po-Chieh Hsu, Jay Wu
Abstract:
The normalized glandular dose (DgN) estimates the energy deposition of mammography in clinical practice. The Monte Carlo simulations frequently use uniformly mixed phantom for calculating the conversion factor. However, breast tissues are not uniformly distributed, leading to errors of conversion factor estimation. This study constructed a three-layer phantom to estimated more accurate of normalized glandular dose. In this study, MCNP code (Monte Carlo N-Particles code) was used to create the geometric structure. We simulated three types of target/filter combinations (Mo/Mo, Mo/Rh, Rh/Rh), six voltages (25 ~ 35 kVp), six HVL parameters and nine breast phantom thicknesses (2 ~ 10 cm) for the three-layer mammographic phantom. The conversion factor for 25%, 50% and 75% glandularity was calculated. The error of conversion factors compared with the results of the American College of Radiology (ACR) was within 6%. For Rh/Rh, the difference was within 9%. The difference between the 50% average glandularity and the uniform phantom was 7.1% ~ -6.7% for the Mo/Mo combination, voltage of 27 kVp, half value layer of 0.34 mmAl, and breast thickness of 4 cm. According to the simulation results, the regression analysis found that the three-layer mammographic phantom at 0% ~ 100% glandularity can be used to accurately calculate the conversion factors. The difference in glandular tissue distribution leads to errors of conversion factor calculation. The three-layer mammographic phantom can provide accurate estimates of glandular dose in clinical practice.Keywords: Monte Carlo simulation, mammography, normalized glandular dose, glandularity
Procedia PDF Downloads 1892023 Earnings vs Cash Flows: The Valuation Perspective
Authors: Megha Agarwal
Abstract:
The research paper is an effort to compare the earnings based and cash flow based methods of valuation of an enterprise. The theoretically equivalent methods based on either earnings such as Residual Earnings Model (REM), Abnormal Earnings Growth Model (AEGM), Residual Operating Income Method (ReOIM), Abnormal Operating Income Growth Model (AOIGM) and its extensions multipliers such as price/earnings ratio, price/book value ratio; or cash flow based models such as Dividend Valuation Method (DVM) and Free Cash Flow Method (FCFM) all provide different estimates of valuation of the Indian giant corporate Reliance India Limited (RIL). An ex-post analysis of published accounting and financial data for four financial years from 2008-09 to 2011-12 has been conducted. A comparison of these valuation estimates with the actual market capitalization of the company shows that the complex accounting based model AOIGM provides closest forecasts. These different estimates may be derived due to inconsistencies in discount rate, growth rates and the other forecasted variables. Although inputs for earnings based models may be available to the investor and analysts through published statements, precise estimation of free cash flows may be better undertaken by the internal management. The estimation of value from more stable parameters as residual operating income and RNOA could be considered superior to the valuations from more volatile return on equity.Keywords: earnings, cash flows, valuation, Residual Earnings Model (REM)
Procedia PDF Downloads 3762022 State Estimator Performance Enhancement: Methods for Identifying Errors in Modelling and Telemetry
Authors: M. Ananthakrishnan, Sunil K Patil, Koti Naveen, Inuganti Hemanth Kumar
Abstract:
State estimation output of EMS forms the base case for all other advanced applications used in real time by a power system operator. Ensuring tuning of state estimator is a repeated process and cannot be left once a good solution is obtained. This paper attempts to demonstrate methods to improve state estimator solution by identifying incorrect modelling and telemetry inputs to the application. In this work, identification of database topology modelling error by plotting static network using node-to-node connection details is demonstrated with examples. Analytical methods to identify wrong transmission parameters, incorrect limits and mistakes in pseudo load and generator modelling are explained with various cases observed. Further, methods used for active and reactive power tuning using bus summation display, reactive power absorption summary, and transformer tap correction are also described. In a large power system, verifying all network static data and modelling parameter on regular basis is difficult .The proposed tuning methods can be easily used by operators to quickly identify errors to obtain the best possible state estimation performance. This, in turn, can lead to improved decision-support capabilities, ultimately enhancing the safety and reliability of the power grid.Keywords: active power tuning, database modelling, reactive power, state estimator
Procedia PDF Downloads 72021 Immunosupressive Effect of Chloroquine through the Inhibition of Myeloperoxidase
Authors: J. B. Minari, O. B. Oloyede
Abstract:
Polymorphonuclear neutrophils (PMNs) play a crucial role in a variety of infections caused by bacteria, fungi, and parasites. Indeed, the involvement of PMNs in host defence against Plasmodium falciparum is well documented both in vitro and in vivo. Many of the antimalarial drugs such as chloroquine used in the treatment of human malaria significantly reduce the immune response of the host in vitro and in vivo. Myeloperoxidase is the most abundant enzyme found in the polymorphonuclear neutrophil which plays a crucial role in its function. This study was carried out to investigate the effect of chloroquine on the enzyme. In investigating the effects of the drug on myeloperoxidase, the influence of concentration, pH, partition ratio estimation and kinetics of inhibition were studied. This study showed that chloroquine is concentration-dependent inhibitor of myeloperoxidase with an IC50 of 0.03 mM. Partition ratio estimation showed that 40 enzymatic turnover cycles are required for complete inhibition of myeloperoxidase in the presence of chloroquine. The influence of pH on the effect of chloroquine on the enzyme showed significant inhibition of myeloperoxidase at physiological pH. The kinetic inhibition studies showed that chloroquine caused a non-competitive inhibition with an inhibition constant Ki of 0.27mM. The results obtained from this study shows that chloroquine is a potent inhibitor of myeloperoxidase and it is capable of inactivating the enzyme. It is therefore considered that the inhibition of myeloperoxidase in the presence of chloroquine as revealed in this study may partly explain the impairment of polymorphonuclear neutrophil and consequent immunosuppression of the host defence system against secondary infections.Keywords: myeloperoxidase, chloroquine, inhibition, neutrophil, immune
Procedia PDF Downloads 3742020 Exceptional Cost and Time Optimization with Successful Leak Repair and Restoration of Oil Production: West Kuwait Case Study
Authors: Nasser Al-Azmi, Al-Sabea Salem, Abu-Eida Abdullah, Milan Patra, Mohamed Elyas, Daniel Freile, Larisa Tagarieva
Abstract:
Well intervention was done along with Production Logging Tools (PLT) to detect sources of water, and to check well integrity for two West Kuwait oil wells started to produce 100 % water. For the first well, to detect the source of water, PLT was performed to check the perforations, no production observed from the bottom two perforation intervals, and an intake of water was observed from the top most perforation. Then a decision was taken to extend the PLT survey from tag depth to the Y-tool. For the second well, the aim was to detect the source of water and if there was a leak in the 7’’liner in front of the upper zones. Data could not be recorded in flowing conditions due to the casing deformation at almost 8300 ft. For the first well from the interpretation of PLT and well integrity data, there was a hole in the 9 5/8'' casing from 8468 ft to 8494 ft producing almost the majority of water, which is 2478 bbl/d. The upper perforation from 10812 ft to 10854 ft was taking 534 stb/d. For the second well, there was a hole in the 7’’liner from 8303 ft MD to 8324 ft MD producing 8334.0 stb/d of water with an intake zone from10322.9-10380.8 ft MD taking the whole fluid. To restore the oil production, W/O rig was mobilized to prevent dump flooding, and during the W/O, the leaking interval was confirmed for both wells. The leakage was cement squeezed and tested at 900-psi positive pressure and 500-psi drawdown pressure. The cement squeeze job was successful. After W/O, the wells kept producing for cleaning, and eventually, the WC reduced to 0%. Regular PLT and well integrity logs are required to study well performance, and well integrity issues, proper cement behind casing is essential to well longevity and well integrity, and the presence of the Y-tool is essential as monitoring of well parameters and ESP to facilitate well intervention tasks. Cost and time optimization in oil and gas and especially during rig operations is crucial. PLT data quality and the accuracy of the interpretations contributed a lot to identify the leakage interval accurately and, in turn, saved a lot of time and reduced the repair cost with almost 35 to 45 %. The added value here was more related to the cost reduction and effective and quick proper decision making based on the economic environment.Keywords: leak, water shut-off, cement, water leak
Procedia PDF Downloads 1162019 Recurrence of Papillary Thyroid Cancer with an Interval of 40 Years. Report of an Autopsy Case
Authors: Satoshi Furukawa, Satomu Morita, Katsuji Nishi, Masahito Hitosugi
Abstract:
A 75-year-old woman took thyroidectomy forty years previously. Enlarged masses were seen at autopsy just above and below the left clavicle. We proved the diagnosis of papillary thyroid cancer (PTC) and lung metastasis by histological examinations. The prognosis of PTC is excellent; the 10-year survival rate ranges between 85 and 99%. Lung metastases may be found in 10% of the patients with PTC. We report an unusual case of recurrence of PTC with metastasis to the lung.Keywords: papillary thyroid cancer, lung metastasis, autopsy, histopathological findings
Procedia PDF Downloads 3402018 Effect Analysis of an Improved Adaptive Speech Noise Reduction Algorithm in Online Communication Scenarios
Authors: Xingxing Peng
Abstract:
With the development of society, there are more and more online communication scenarios such as teleconference and online education. In the process of conference communication, the quality of voice communication is a very important part, and noise may cause the communication effect of participants to be greatly reduced. Therefore, voice noise reduction has an important impact on scenarios such as voice calls. This research focuses on the key technologies of the sound transmission process. The purpose is to maintain the audio quality to the maximum so that the listener can hear clearer and smoother sound. Firstly, to solve the problem that the traditional speech enhancement algorithm is not ideal when dealing with non-stationary noise, an adaptive speech noise reduction algorithm is studied in this paper. Traditional noise estimation methods are mainly used to deal with stationary noise. In this chapter, we study the spectral characteristics of different noise types, especially the characteristics of non-stationary Burst noise, and design a noise estimator module to deal with non-stationary noise. Noise features are extracted from non-speech segments, and the noise estimation module is adjusted in real time according to different noise characteristics. This adaptive algorithm can enhance speech according to different noise characteristics, improve the performance of traditional algorithms to deal with non-stationary noise, so as to achieve better enhancement effect. The experimental results show that the algorithm proposed in this chapter is effective and can better adapt to different types of noise, so as to obtain better speech enhancement effect.Keywords: speech noise reduction, speech enhancement, self-adaptation, Wiener filter algorithm
Procedia PDF Downloads 582017 Modeling Default Probabilities of the Chosen Czech Banks in the Time of the Financial Crisis
Authors: Petr Gurný
Abstract:
One of the most important tasks in the risk management is the correct determination of probability of default (PD) of particular financial subjects. In this paper a possibility of determination of financial institution’s PD according to the credit-scoring models is discussed. The paper is divided into the two parts. The first part is devoted to the estimation of the three different models (based on the linear discriminant analysis, logit regression and probit regression) from the sample of almost three hundred US commercial banks. Afterwards these models are compared and verified on the control sample with the view to choose the best one. The second part of the paper is aimed at the application of the chosen model on the portfolio of three key Czech banks to estimate their present financial stability. However, it is not less important to be able to estimate the evolution of PD in the future. For this reason, the second task in this paper is to estimate the probability distribution of the future PD for the Czech banks. So, there are sampled randomly the values of particular indicators and estimated the PDs’ distribution, while it’s assumed that the indicators are distributed according to the multidimensional subordinated Lévy model (Variance Gamma model and Normal Inverse Gaussian model, particularly). Although the obtained results show that all banks are relatively healthy, there is still high chance that “a financial crisis” will occur, at least in terms of probability. This is indicated by estimation of the various quantiles in the estimated distributions. Finally, it should be noted that the applicability of the estimated model (with respect to the used data) is limited to the recessionary phase of the financial market.Keywords: credit-scoring models, multidimensional subordinated Lévy model, probability of default
Procedia PDF Downloads 4562016 The Increasing of Perception of Consumers’ Awareness about Sustainability Brands during Pandemic: A Multi Mediation Model
Authors: Silvia Platania, Martina Morando, Giuseppe Santisi
Abstract:
Introduction: In the last thirty years, there is constant talk of sustainable consumption and a "transition" of consumer lifestyles towards greater awareness of consumer choices (United Nation, 1992). The 2019 coronavirus (COVID-19) epidemic that has hit the world population since 2020 has had significant consequences in all areas of people's lives; individuals have been forced to change their behaviors, to redefine their owngoals, priorities, practices, and lifestyles, to rebuild themselves in the new situation dictated by the pandemic. Method(Participants and procedure ): The data were collected through an online survey; moreover, we used convenience sampling from the general population. The participants were 669 Italians consumers (Female= 514, 76.8%; Male=155, 23.2%) that choice sustainability brands, aged between 18 and 65 years (Mₐ𝓰ₑ = 35.45; Standard Deviation, SD = 9.51).(Measure ): The following measures were used: The Muncy–Vitell Consumer Ethics Scale; Attitude Toward Business Scale; Perceived Consumer Effectiveness Scale; Consumers Perception on Sustainable Brand Attitudes. Results: Preliminary analyses were conducted to test our model. Pearson's bivariate correlation between variables shows that all variables of our model correlate significantly and positively, PCE with CPSBA (r = .56, p <.001). Furthermore, a CFA, according to Harman's single-factor test, was used to diagnose the extent to which common-method variance was a problem. A comparison between the hypothesised model and a model with one factor (with all items loading on a unique factor) revealed that the former provided a better fit for the data in all the CFA fit measures [χ² [6, n = 669] = 7.228, p = 0.024, χ² / df = 1.20, RMSEA = 0.07 (CI = 0.051-0.067), CFI = 0.95, GFI = 0.95, SRMR = 0.04, AIC = 66.501; BIC = 132,150). Next, amulti mediation was conducted to test our hypotheses. The results show that there is a direct effect of PCE on ethical consumption behavior (β = .38) and on ATB (β = .23); furthermore, there is a direct effect on the CPSBA outcome (β = .34). In addition, there is a mediating effect by ATB (C.I. =. 022-.119, 95% interval confidence) and by CES (C.I. =. 136-.328, 95% interval confidence). Conclusion: The spread of the COVID-19 pandemic has affected consumer consumption styles and has led to an increase in online shopping and purchases of sustainable products. Several theoretical and practical considerations emerge from the results of the study.Keywords: decision making, sustainability, pandemic, multimediation model
Procedia PDF Downloads 110