Search results for: molecular dynamics simulation.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8797

Search results for: molecular dynamics simulation.

1627 Early Warning System of Financial Distress Based On Credit Cycle Index

Authors: Bi-Huei Tsai

Abstract:

Previous studies on financial distress prediction choose the conventional failing and non-failing dichotomy; however, the distressed extent differs substantially among different financial distress events. To solve the problem, “non-distressed”, “slightly-distressed” and “reorganization and bankruptcy” are used in our article to approximate the continuum of corporate financial health. This paper explains different financial distress events using the two-stage method. First, this investigation adopts firm-specific financial ratios, corporate governance and market factors to measure the probability of various financial distress events based on multinomial logit models. Specifically, the bootstrapping simulation is performed to examine the difference of estimated misclassifying cost (EMC). Second, this work further applies macroeconomic factors to establish the credit cycle index and determines the distressed cut-off indicator of the two-stage models using such index. Two different models, one-stage and two-stage prediction models, are developed to forecast financial distress, and the results acquired from different models are compared with each other, and with the collected data. The findings show that the two-stage model incorporating financial ratios, corporate governance and market factors has the lowest misclassification error rate. The two-stage model is more accurate than the one-stage model as its distressed cut-off indicators are adjusted according to the macroeconomic-based credit cycle index.

Keywords: Multinomial logit model, corporate governance, company failure, reorganization, bankruptcy

Procedia PDF Downloads 361
1626 Understanding the Complexities of Consumer Financial Spinning

Authors: Olivier Mesly

Abstract:

This research presents a conceptual framework termed “Consumer Financial Spinning” (CFS) to analyze consumer behavior in the financial/economic markets. This phenomenon occurs when consumers of high-stakes financial products accumulate unsustainable debt, leading them to detach from their initial financial hierarchy of needs, wealth-related goals, and preferences regarding their household portfolio of assets. The daring actions of these consumers, forming a dark financial triangle, are characterized by three behaviors: overconfidence, the use of rationed rationality, and deceitfulness. We show that we can incorporate CFS into the traditional CAPM and Markovitz’ portfolio optimization models to create a framework that explains such market phenomena as the global financial crisis, highlighting the antecedents and consequences of ill-conceived speculation. Because this is a conceptual paper, there is no methodology with respect to ground studies. However, we apply modeling principles derived from the data percolation methodology, which contains tenets explicating how to structure concepts. A simulation test of the proposed framework is conducted; it demonstrates the conditions under which the relationship between expected returns and risk may deviate from linearity. The analysis and conceptual findings are particularly relevant both theoretically and pragmatically as they shed light on the psychological conditions that drive intense speculation, which can lead to market turmoil. Armed with such understanding, regulators are better equipped to propose solutions before the economic problems become out of control.

Keywords: consumer financial spinning, rationality, deceitfulness, overconfidence, CAPM

Procedia PDF Downloads 25
1625 A Multi-Objective Programming Model to Supplier Selection and Order Allocation Problem in Stochastic Environment

Authors: Rouhallah Bagheri, Morteza Mahmoudi, Hadi Moheb-Alizadeh

Abstract:

This paper aims at developing a multi-objective model for supplier selection and order allocation problem in stochastic environment, where purchasing cost, percentage of delivered items with delay and percentage of rejected items provided by each supplier are supposed to be stochastic parameters following any arbitrary probability distribution. In this regard, dependent chance programming is used which maximizes probability of the event that total purchasing cost, total delivered items with delay and total rejected items are less than or equal to pre-determined values given by decision maker. The abovementioned stochastic multi-objective programming problem is then transformed into a stochastic single objective programming problem using minimum deviation method. In the next step, the further problem is solved applying a genetic algorithm, which performs a simulation process in order to calculate the stochastic objective function as its fitness function. Finally, the impact of stochastic parameters on the given solution is examined via a sensitivity analysis exploiting coefficient of variation. The results show that whatever stochastic parameters have greater coefficients of variation, the value of the objective function in the stochastic single objective programming problem is deteriorated.

Keywords: supplier selection, order allocation, dependent chance programming, genetic algorithm

Procedia PDF Downloads 294
1624 Encoded Fiber Optic Sensors for Simultaneous Multipoint Sensing

Authors: C. Babu Rao, Pandian Chelliah

Abstract:

Owing to their reliability, a number of fluorescent spectra based fiber optic sensors have been developed for detection and identification of hazardous chemicals such as explosives, narcotics etc. In High security regions, such as airports, it is important to monitor simultaneously multiple locations. This calls for deployment of a portable sensor at each location. However, the selectivity and sensitivity of these techniques depends on the spectral resolution of the spectral analyzer. The better the resolution the larger the repertoire of chemicals that can be detected. A portable unit will have limitations in meeting these requirements. Optical fibers can be employed for collecting and transmitting spectral signal from the portable sensor head to a sensitive central spectral analyzer (CSA). For multipoint sensing, optical multiplexing of multiple sensor heads with CSA has to be adopted. However with multiplexing, when one sensor head is connected to CSA, the rest may remain unconnected for the turn-around period. The larger the number of sensor heads the larger this turn-around time will be. To circumvent this imitation, we propose in this paper, an optical encoding methodology to use multiple portable sensor heads connected to a single CSA. Each portable sensor head is assigned an unique address. Spectra of every chemical detected through this sensor head, are encoded by its unique address and can be identified at the CSA end. The methodology proposed is demonstrated through a simulation using Matlab SIMULINK.

Keywords: optical encoding, fluorescence, multipoint sensing

Procedia PDF Downloads 698
1623 Structure-Guided Optimization of Sulphonamide as Gamma–Secretase Inhibitors for the Treatment of Alzheimer’s Disease

Authors: Vaishali Patil, Neeraj Masand

Abstract:

In older people, Alzheimer’s disease (AD) is turning out to be a lethal disease. According to the amyloid hypothesis, aggregation of the amyloid β–protein (Aβ), particularly its 42-residue variant (Aβ42), plays direct role in the pathogenesis of AD. Aβ is generated through sequential cleavage of amyloid precursor protein (APP) by β–secretase (BACE) and γ–secretase (GS). Thus in the treatment of AD, γ-secretase modulators (GSMs) are potential disease-modifying as they selectively lower pathogenic Aβ42 levels by shifting the enzyme cleavage sites without inhibiting γ–secretase activity. This possibly avoids known adverse effects observed with complete inhibition of the enzyme complex. Virtual screening, via drug-like ADMET filter, QSAR and molecular docking analyses, has been utilized to identify novel γ–secretase modulators with sulphonamide nucleus. Based on QSAR analyses and docking score, some novel analogs have been synthesized. The results obtained by in silico studies have been validated by performing in vivo analysis. In the first step, behavioral assessment has been carried out using Scopolamine induced amnesia methodology. Later the same series has been evaluated for neuroprotective potential against the oxidative stress induced by Scopolamine. Biochemical estimation was performed to evaluate the changes in biochemical markers of Alzheimer’s disease such as lipid peroxidation (LPO), Glutathione reductase (GSH), and Catalase. The Scopolamine induced amnesia model has shown increased Acetylcholinesterase (AChE) levels and the inhibitory effect of test compounds in the brain AChE levels have been evaluated. In all the studies Donapezil (Dose: 50µg/kg) has been used as reference drug. The reduced AChE activity is shown by compounds 3f, 3c, and 3e. In the later stage, the most potent compounds have been evaluated for Aβ42 inhibitory profile. It can be hypothesized that this series of alkyl-aryl sulphonamides exhibit anti-AD activity by inhibition of Acetylcholinesterase (AChE) enzyme as well as inhibition of plaque formation on prolong dosage along with neuroprotection from oxidative stress.

Keywords: gamma-secretase inhibitors, Alzzheimer's disease, sulphonamides, QSAR

Procedia PDF Downloads 234
1622 Swastika Shape Multiband Patch Antenna for Wireless Applications on Low Cost Substrate

Authors: Md. Samsuzzaman, M. T. Islam, J. S. Mandeep, N. Misran

Abstract:

In this article, a compact simple structure modified Swastika shape patch multiband antenna on a substrate of available low cost polymer resin composite material is designed for Wi-Fi and WiMAX applications. The substrate material consists of an epoxy matrix reinforced by woven glass. The designed micro-strip line fed compact antenna comprises of a planar wide square slot ground with four slits and Swastika shape radiation patch with a rectangular slot. The effect of the different substrate materials on the reflection coefficients of the proposed antennas was also analyzed. It can be clearly seen that the proposed antenna provides a wider bandwidth and acceptable return loss value compared to other reported materials. The simulation results exhibits that the antenna has an impedance bandwidth with -10 dB return loss at 3.01-3.89 GHz and 4.88-6.10 GHz which can cover both the WLAN, WiMAX and public safety WLAN bands. The proposed swastika shape antenna was designed and analyzed by using a finite element method based simulator HFSS and designed on a low cost FR4 (polymer resin composite material) printed circuit board. The electrical performances and superior frequency characteristics make the proposed material antenna desirable for wireless communications.

Keywords: epoxy resin polymer, multiband, swastika shaped, wide slot, WLAN/WiMAX

Procedia PDF Downloads 433
1621 NOx Prediction by Quasi-Dimensional Combustion Model of Hydrogen Enriched Compressed Natural Gas Engine

Authors: Anas Rao, Hao Duan, Fanhua Ma

Abstract:

The dependency on the fossil fuels can be minimized by using the hydrogen enriched compressed natural gas (HCNG) in the transportation vehicles. However, the NOx emissions of HCNG engines are significantly higher, and this turned to be its major drawback. Therefore, the study of NOx emission of HCNG engines is a very important area of research. In this context, the experiments have been performed at the different hydrogen percentage, ignition timing, air-fuel ratio, manifold-absolute pressure, load and engine speed. Afterwards, the simulation has been accomplished by the quasi-dimensional combustion model of HCNG engine. In order to investigate the NOx emission, the NO mechanism has been coupled to the quasi-dimensional combustion model of HCNG engine. The three NOx mechanism: the thermal NOx, prompt NOx and N2O mechanism have been used to predict NOx emission. For the validation purpose, NO curve has been transformed into NO packets based on the temperature difference of 100 K for the lean-burn and 60 K for stoichiometric condition. While, the width of the packet has been taken as the ratio of crank duration of the packet to the total burnt duration. The combustion chamber of the engine has been divided into three zones, with the zone equal to the product of summation of NO packets and space. In order to check the accuracy of the model, the percentage error of NOx emission has been evaluated, and it lies in the range of ±6% and ±10% for the lean-burn and stoichiometric conditions respectively. Finally, the percentage contribution of each NO formation has been evaluated.

Keywords: quasi-dimensional combustion , thermal NO, prompt NO, NO packet

Procedia PDF Downloads 231
1620 Numerical Study of Natural Convection in Isothermal Open Cavities

Authors: Gaurav Prabhudesai, Gaetan Brill

Abstract:

The sun's energy source comes from a hydrogen-to-helium thermonuclear reaction, generating a temperature of about 5760 K on its outer layer. On account of this high temperature, energy is radiated by the sun, a part of which reaches the earth. This sunlight, even after losing part of its energy en-route to scattering and absorption, provides a time and space averaged solar flux of 174.7 W/m^2 striking the earth’s surface. According to one study, the solar energy striking earth’s surface in one and a half hour is more than the energy consumption that was recorded in the year 2001 from all sources combined. Thus, technology for extraction of solar energy holds much promise for solving energy crisis. Of the many technologies developed in this regard, Concentrating Solar Power (CSP) plants with central solar tower and receiver system are very impressive because of their capability to provide a renewable energy that can be stored in the form of heat. One design of central receiver towers is an open cavity where sunlight is concentrated into by using mirrors (also called heliostats). This concentrated solar flux produces high temperature inside the cavity which can be utilized in an energy conversion process. The amount of energy captured is reduced by losses occurring at the cavity through all three modes viz., radiation to the atmosphere, conduction to the adjoining structure and convection. This study investigates the natural convection losses to the environment from the receiver. Computational fluid dynamics were used to simulate the fluid flow and heat transfer of the receiver; since no analytical solution can be obtained and no empirical correlations exist for the given geometry. The results provide guide lines for predicting natural convection losses for hexagonal and circular shaped open cavities. Additionally, correlations are given for various inclination angles and aspect ratios. These results provide methods to minimize natural convection through careful design of receiver geometry and modification of the inclination angle, and aspect ratio of the cavity.

Keywords: concentrated solar power (CSP), central receivers, natural convection, CFD, open cavities

Procedia PDF Downloads 269
1619 A Normalized Non-Stationary Wavelet Based Analysis Approach for a Computer Assisted Classification of Laryngoscopic High-Speed Video Recordings

Authors: Mona K. Fehling, Jakob Unger, Dietmar J. Hecker, Bernhard Schick, Joerg Lohscheller

Abstract:

Voice disorders origin from disturbances of the vibration patterns of the two vocal folds located within the human larynx. Consequently, the visual examination of vocal fold vibrations is an integral part within the clinical diagnostic process. For an objective analysis of the vocal fold vibration patterns, the two-dimensional vocal fold dynamics are captured during sustained phonation using an endoscopic high-speed camera. In this work, we present an approach allowing a fully automatic analysis of the high-speed video data including a computerized classification of healthy and pathological voices. The approach bases on a wavelet-based analysis of so-called phonovibrograms (PVG), which are extracted from the high-speed videos and comprise the entire two-dimensional vibration pattern of each vocal fold individually. Using a principal component analysis (PCA) strategy a low-dimensional feature set is computed from each phonovibrogram. From the PCA-space clinically relevant measures can be derived that quantify objectively vibration abnormalities. In the first part of the work it will be shown that, using a machine learning approach, the derived measures are suitable to distinguish automatically between healthy and pathological voices. Within the approach the formation of the PCA-space and consequently the extracted quantitative measures depend on the clinical data, which were used to compute the principle components. Therefore, in the second part of the work we proposed a strategy to achieve a normalization of the PCA-space by registering the PCA-space to a coordinate system using a set of synthetically generated vibration patterns. The results show that owing to the normalization step potential ambiguousness of the parameter space can be eliminated. The normalization further allows a direct comparison of research results, which bases on PCA-spaces obtained from different clinical subjects.

Keywords: Wavelet-based analysis, Multiscale product, normalization, computer assisted classification, high-speed laryngoscopy, vocal fold analysis, phonovibrogram

Procedia PDF Downloads 248
1618 Use of Didactic Bibliographic Resources to Improve the Teaching and Learning Processes of Animal Reproduction in Veterinary Science

Authors: Yasser Y. Lenis, Amy Jo Montgomery, Diego F. Carrillo-Gonzalez

Abstract:

Introduction: The use of didactic instruments in different learning environments plays a pivotal role in enhancing the level of knowledge in veterinary science students. The direct instruction of basic animal reproduction concepts in students enrolled in veterinary medicine programs allows them to elucidate the biological and molecular mechanisms that perpetuate the animal species in an ecosystem. Therefore, universities must implement didactic strategies that facilitate the teaching and learning processes for students and, in turn, enrich learning environments. Objective: to evaluate the effect of the use of a didactic textbook on the level of theoretical knowledge in embryo-maternal recognition for veterinary medicine students. Methods: the participants (n=24) were divided into two experimental groups: control (Ctrl) and treatment (Treat). Both groups received 4 hours of theoretical training regarding the basic concepts in bovine embryo-maternal recognition. However, the Treat group was also exposed to a guided lecture and the activity play-to-learn from a cow reproduction didactic textbook. A pre-test and a post-test were applied to assess the prior and subsequent knowledge in the participants. Descriptive statistics were applied to identify the success rates for each of the tests. Afterwards, a repeated measures model was applied where the effect of the intervention was considered. Results: no significant difference (p>0,05) was observed in the number of right answers for groups Ctrl (54,2%±12,7) and Treat (40,8%±16,8) in the pre-test. There was no difference (p>0,05) compering the number of right answers in Ctrl pre-test (54,2%±12,7) and post-test (60,8±18,8). However, the Treat group showed a significant (p>0,05) difference in the number of right answers when comparing pre-test (40,8%±16,8) and post-test (71,7%±14,7). Finally, after the theoretical training and the didactic activity in the Treat group, an increase of 10.9% (p<0,05) in the number of right answers was found when compared with the Ctrl group. Conclusion: the use of didactic tools that include guided lectures and activities like play-to-learn from a didactic textbook enhances the level of knowledge in an animal reproduction course for veterinary medicine students.

Keywords: animal reproduction, pedagogic, level of knowledge, learning environment

Procedia PDF Downloads 46
1617 Estimation of a Finite Population Mean under Random Non Response Using Improved Nadaraya and Watson Kernel Weights

Authors: Nelson Bii, Christopher Ouma, John Odhiambo

Abstract:

Non-response is a potential source of errors in sample surveys. It introduces bias and large variance in the estimation of finite population parameters. Regression models have been recognized as one of the techniques of reducing bias and variance due to random non-response using auxiliary data. In this study, it is assumed that random non-response occurs in the survey variable in the second stage of cluster sampling, assuming full auxiliary information is available throughout. Auxiliary information is used at the estimation stage via a regression model to address the problem of random non-response. In particular, the auxiliary information is used via an improved Nadaraya-Watson kernel regression technique to compensate for random non-response. The asymptotic bias and mean squared error of the estimator proposed are derived. Besides, a simulation study conducted indicates that the proposed estimator has smaller values of the bias and smaller mean squared error values compared to existing estimators of finite population mean. The proposed estimator is also shown to have tighter confidence interval lengths at a 95% coverage rate. The results obtained in this study are useful, for instance, in choosing efficient estimators of the finite population mean in demographic sample surveys.

Keywords: mean squared error, random non-response, two-stage cluster sampling, confidence interval lengths

Procedia PDF Downloads 117
1616 Bayesian Borrowing Methods for Count Data: Analysis of Incontinence Episodes in Patients with Overactive Bladder

Authors: Akalu Banbeta, Emmanuel Lesaffre, Reynaldo Martina, Joost Van Rosmalen

Abstract:

Including data from previous studies (historical data) in the analysis of the current study may reduce the sample size requirement and/or increase the power of analysis. The most common example is incorporating historical control data in the analysis of a current clinical trial. However, this only applies when the historical control dataare similar enough to the current control data. Recently, several Bayesian approaches for incorporating historical data have been proposed, such as the meta-analytic-predictive (MAP) prior and the modified power prior (MPP) both for single control as well as for multiple historical control arms. Here, we examine the performance of the MAP and the MPP approaches for the analysis of (over-dispersed) count data. To this end, we propose a computational method for the MPP approach for the Poisson and the negative binomial models. We conducted an extensive simulation study to assess the performance of Bayesian approaches. Additionally, we illustrate our approaches on an overactive bladder data set. For similar data across the control arms, the MPP approach outperformed the MAP approach with respect to thestatistical power. When the means across the control arms are different, the MPP yielded a slightly inflated type I error (TIE) rate, whereas the MAP did not. In contrast, when the dispersion parameters are different, the MAP gave an inflated TIE rate, whereas the MPP did not.We conclude that the MPP approach is more promising than the MAP approach for incorporating historical count data.

Keywords: count data, meta-analytic prior, negative binomial, poisson

Procedia PDF Downloads 103
1615 Bioeconomic Modelling for Barramundi (Lates calcarifer) in Queensland: Implications for Recreational Fishing Following Recent Gill Netting Closures

Authors: Sabiha S. Marine, Nicole Flint, John Rolfe

Abstract:

The Queensland state government introduced commercial gill net fishing closures in Cairns, Mackay, and Rockhampton in November 2015 to increase the recreational fishing opportunities, nature-based tourism, and economic benefits in these three regional areas. This management change is likely to improve the potential for more desirable stock structures through natural recruitment. Barramundi (Lates calcarifer) is one of the popular target fish for recreational and commercial fishers in Northern Australia. This investigation examines the effects of reduced commercial fishing from both biological and economic perspectives, particularly on the local Barramundi population of the Fitzroy River in Rockhampton, the largest river catchment flowing to the eastern coast of Australia. Data on different parameters of biological and economic aspects have been collated from secondary sources for analysis through a system simulation approach to identify the effectiveness of the commercial netting closures on recreational fishing effort, especially for the Barramundi population. The results have the potential to explain certain consequences of the netting closures in Queensland, which could serve to inform future fisheries management decisions. The study output as a whole will help in the better management of fisheries resources by evaluating recreational fishing opportunities in Queensland, where the potential for increases in recreation is high.

Keywords: Barramundi, bioeconomic model, fishery management, recreational fishing

Procedia PDF Downloads 148
1614 A Micro-Scale of Electromechanical System Micro-Sensor Resonator Based on UNO-Microcontroller for Low Magnetic Field Detection

Authors: Waddah Abdelbagi Talha, Mohammed Abdullah Elmaleeh, John Ojur Dennis

Abstract:

This paper focuses on the simulation and implementation of a resonator micro-sensor for low magnetic field sensing based on a U-shaped cantilever and piezoresistive configuration, which works based on Lorentz force physical phenomena. The resonance frequency is an important parameter that depends upon the highest response and sensitivity through the frequency domain (frequency response) of any vibrated micro-scale of an electromechanical system (MEMS) device. And it is important to determine the direction of the detected magnetic field. The deflection of the cantilever is considered for vibrated mode with different frequencies in the range of (0 Hz to 7000 Hz); for the purpose of observing the frequency response. A simple electronic circuit-based polysilicon piezoresistors in Wheatstone's bridge configuration are used to transduce the response of the cantilever to electrical measurements at various voltages. Microcontroller-based Arduino program and PROTEUS electronic software are used to analyze the output signals from the sensor. The highest output voltage amplitude of about 4.7 mV is spotted at about 3 kHz of the frequency domain, indicating the highest sensitivity, which can be called resonant sensitivity. Based on the resonant frequency value, the mode of vibration is determined (up-down vibration), and based on that, the vector of the magnetic field is also determined.

Keywords: resonant frequency, sensitivity, Wheatstone bridge, UNO-microcontroller

Procedia PDF Downloads 101
1613 Adsorptive Media Selection for Bilirubin Removal: An Adsorption Equilibrium Study

Authors: Vincenzo Piemonte

Abstract:

The liver is a complex, large-scale biochemical reactor which plays a unique role in the human physiology. When liver ceases to perform its physiological activity, a functional replacement is required. Actually, liver transplantation is the only clinically effective method of treating severe liver disease. Anyway, the aforementioned therapeutic approach is hampered by the disparity between organ availability and the number of patients on the waiting list. In order to overcome this critical issue, research activities focused on liver support device systems (LSDs) designed to bridging patients to transplantation or to keep them alive until the recovery of native liver function. In recirculating albumin dialysis devices, such as MARS (Molecular Adsorbed Recirculating System), adsorption is one of the fundamental steps in albumin-dialysate regeneration. Among the albumin-bound toxins that must be removed from blood during liver-failure therapy, bilirubin and tryptophan can be considered as representative of two different toxin classes. The first one, not water soluble at physiological blood pH and strongly bounded to albumin, the second one, loosely albumin bound and partially water soluble at pH 7.4. Fixed bed units are normally used for this task, and the design of such units requires information both on toxin adsorption equilibrium and kinetics. The most common adsorptive media used in LSDs are activated carbon, non-ionic polymeric resins and anionic resins. In this paper, bilirubin adsorption isotherms on different adsorptive media, such as polymeric resin, albumin-coated resin, anionic resin, activated carbon and alginate beads with entrapped albumin are presented. By comparing all the results, it can be stated that the adsorption capacity for bilirubin of the five different media increases in the following order: Alginate beads < Polymeric resin < Albumin-coated resin < Activated carbon < Anionic resin. The main focus of this paper is to provide useful guidelines for the optimization of liver support devices which implement adsorption columns to remove albumin-bound toxins from albumin dialysate solutions.

Keywords: adsorptive media, adsorption equilibrium, artificial liver devices, bilirubin, mathematical modelling

Procedia PDF Downloads 245
1612 Evaluation of Azo Dye Toxicity Using Some Haematological and Histopathological Alterations in Fish Catla Catla

Authors: Jagruti Barot

Abstract:

The textile industry plays a major role in the economy of India and on the other side of the coin it is the major source for water pollution. As azo dyes is the largest dye class they are extensively used in many fields such as textile industry, leather tanning industry, paper production, food, colour photography, pharmaceuticals and medicine, cosmetic, hair colourings, wood staining, agricultural, biological and chemical research etc. In addition to these, they can have acute and/or chronic effects on organisms depending on their concentration and length of exposure when they discharged as effluent in the environment. The aim of this study was to assess the genotoxic and histotoxic potentials of environmentally relevant concentrations of RR 120 on Catla catla, important edible freshwater fingerlings. For this, healthy Catla catla fingerlings were procured from the Government Fish Farm and acclimatized in 100 L capacity and continuously aerated glass aquarium in laboratory for 15 days. According to APHA some physic-chemical parameters were measured and maintained such as temperature, pH, dissolve oxygen, alkalinity, total hardness. Water along with excreta had been changed every 24 hrs. All fingerlings were fed artificial food palates once a day @ body weight. After 15 days fingerlings were grouped in 5 (10 in each) and exposed to various concentrations of RR 120 (Control, 10, 20, 30 and 40 mg/L) and samples (peripheral blood and gills, kidney) were collected and analyzed at 96 hrs. of interval. All results were compared with the control. Micronuclei (MN), nuclear buds (NB), fragmented-apoptotic (FA) and bi-nucleated (BN) cells in blood cells and in tissues (gills and kidney cells) were observed. Prominent histopathological alterations were noticed in gills such as aneurism, hyperplasia, degenerated central axis, lifting of gill epithelium, curved secondary gill lamellae etc. Similarly kidney showed some detrimental changes like shrunken glomeruli with increased periglomerular space, degenerated renal tubules etc. Both haematological and histopathological changes clearly reveal the toxic potential of RR 120. This work concludes that water pollution assessment can be done by these two biomarkers which provide baseline to the further chromosomal or molecular work.

Keywords: micronuclei, genotoxicity, RR 120, Catla catla

Procedia PDF Downloads 193
1611 Estimation of the Road Traffic Emissions and Dispersion in the Developing Countries Conditions

Authors: Hicham Gourgue, Ahmed Aharoune, Ahmed Ihlal

Abstract:

We present in this work our model of road traffic emissions (line sources) and dispersion of these emissions, named DISPOLSPEM (Dispersion of Poly Sources and Pollutants Emission Model). In its emission part, this model was designed to keep the consistent bottom-up and top-down approaches. It also allows to generate emission inventories from reduced input parameters being adapted to existing conditions in Morocco and in the other developing countries. While several simplifications are made, all the performance of the model results are kept. A further important advantage of the model is that it allows the uncertainty calculation and emission rate uncertainty according to each of the input parameters. In the dispersion part of the model, an improved line source model has been developed, implemented and tested against a reference solution. It provides improvement in accuracy over previous formulas of line source Gaussian plume model, without being too demanding in terms of computational resources. In the case study presented here, the biggest errors were associated with the ends of line source sections; these errors will be canceled by adjacent sections of line sources during the simulation of a road network. In cases where the wind is parallel to the source line, the use of the combination discretized source and analytical line source formulas minimizes remarkably the error. Because this combination is applied only for a small number of wind directions, it should not excessively increase the calculation time.

Keywords: air pollution, dispersion, emissions, line sources, road traffic, urban transport

Procedia PDF Downloads 423
1610 Investigating the Shear Behaviour of Fouled Ballast Using Discrete Element Modelling

Authors: Ngoc Trung Ngo, Buddhima Indraratna, Cholachat Rujikiathmakjornr

Abstract:

For several hundred years, the design of railway tracks has practically remained unchanged. Traditionally, rail tracks are placed on a ballast layer due to several reasons, including economy, rapid drainage, and high load bearing capacity. The primary function of ballast is to distributing dynamic track loads to sub-ballast and subgrade layers, while also providing lateral resistance and allowing for rapid drainage. Upon repeated trainloads, the ballast becomes fouled due to ballast degradation and the intrusion of fines which adversely affects the strength and deformation behaviour of ballast. This paper presents the use of three-dimensional discrete element method (DEM) in studying the shear behaviour of the fouled ballast subjected to direct shear loading. Irregularly shaped particles of ballast were modelled by grouping many spherical balls together in appropriate sizes to simulate representative ballast aggregates. Fouled ballast was modelled by injecting a specified number of miniature spherical particles into the void spaces. The DEM simulation highlights that the peak shear stress of the ballast assembly decreases and the dilation of fouled ballast increases with an increase level of fouling. Additionally, the distributions of contact force chain and particle displacement vectors were captured during shearing progress, explaining the formation of shear band and the evolutions of volumetric change of fouled ballast.

Keywords: railway ballast, coal fouling, discrete element modelling, discrete element method

Procedia PDF Downloads 434
1609 Design of a Real Time Closed Loop Simulation Test Bed on a General Purpose Operating System: Practical Approaches

Authors: Pratibha Srivastava, Chithra V. J., Sudhakar S., Nitin K. D.

Abstract:

A closed-loop system comprises of a controller, a response system, and an actuating system. The controller, which is the system under test for us, excites the actuators based on feedback from the sensors in a periodic manner. The sensors should provide the feedback to the System Under Test (SUT) within a deterministic time post excitation of the actuators. Any delay or miss in the generation of response or acquisition of excitation pulses may lead to control loop controller computation errors, which can be catastrophic in certain cases. Such systems categorised as hard real-time systems that need special strategies. The real-time operating systems available in the market may be the best solutions for such kind of simulations, but they pose limitations like the availability of the X Windows system, graphical interfaces, other user tools. In this paper, we present strategies that can be used on a general purpose operating system (Bare Linux Kernel) to achieve a deterministic deadline and hence have the added advantages of a GPOS with real-time features. Techniques shall be discussed how to make the time-critical application run with the highest priority in an uninterrupted manner, reduced network latency for distributed architecture, real-time data acquisition, data storage, and retrieval, user interactions, etc.

Keywords: real time data acquisition, real time kernel preemption, scheduling, network latency

Procedia PDF Downloads 128
1608 Uranoplasty Using Tongue Flap for Bilateral Clefts

Authors: Saidasanov Saidazal Shokhmurodovich, Topolnickiy Orest Zinovyevich, Afaunova Olga Arturovna

Abstract:

Relevance: Bilateral congenital cleft is one of the most complex forms of all clefts, which makes it difficult to choose a surgical method of treatment. During primary operations to close the hard and soft palate, there is a shortage of soft tissues and their lack during standard uranoplasty, and these factors aggravate the period of rehabilitation of patients. Materials and methods: The results of surgical treatment of children with bilateral cleft, who underwent uranoplasty using a flap from the tongue, were analyzed. The study used methods: clinical and statistical, which allowed us to solve the tasks, based on the principles of evidence-based medicine. Results and discussion: in our study, 15 patients were studied, who underwent surgical treatment in the following volume: uranoplasty using a flap from the tongue in two stages. Of these, 9 boys and 6 girls aged 2.5 to 6 years. The first stage was surgical treatment in the volume: veloplasty. The second stage was a surgical intervention in volume: uranoplasty using a flap from the tongue. In all patients, the width of the cleft ranged from 1.6-2.8 cm. All patients in this group were orthodontically prepared. Using this method, the surgeon can achieve the following results: maximum narrowing of the palatopharyngeal ring, long soft palate, complete closure of the hard palate, alveolar process, and the mucous membrane of the nasal cavity is also sutured, which creates good conditions for the next stage of osteoplastic surgery. Based on the result obtained, patients have positive results of working with a speech therapist. In all patients, the dynamics were positive without complications. Conclusions: Based on our observation, tongue flap uranoplasty is one of the effective techniques for patients with wide clefts of the hard and soft palate. The use of a flap from the tongue makes it possible to reduce the number of repeated reoperations and improve the quality of social adaptation of this group of patients, which is one of the important stages of rehabilitation. Upon completion of the stages of rehabilitation, all patients had the maximum improvement in functional, anatomical and social indicators.

Keywords: congenital cleft lips and palate, bilateral cleft, child surgery, maxillofacial surgery

Procedia PDF Downloads 106
1607 Gender Justice and Empowerment: A Study of Chhara Bootlegger Women of Ahmedabad

Authors: Neeta Khurana, Ritu Sharma

Abstract:

This paper is an impact assessment study of the rehabilitation work done for Chhara women in the rural precincts of Ahmedabad. The Chharas constitute a denotified tribe and live in abject poverty. The women of this community are infamous absconders of law and active bootleggers of locally made liquor. As part of a psychological study with a local NGO, the authors headed a training program aimed at rehabilitating and providing these women alternate modes of employment, thereby driving them away from a life of crime. The paper centers on the idea of women entrepreneurship and women empowerment. It notes the importance of handholding in a conflict situation. Most of the research on Chharas is either focused on victimising them for state-sponsored violence or mostly makes a plea on reconditioning them in the mainstream. Going against this trend, this paper which documents the study argues that making these poor women self-dependent is a panacea for their sluggish development. The alienation caused due to the demonisation of the community has made them abandon traditional modes of employment. This has further led the community astray into making illegal country liquor causing further damage to their reputation. Women are at the centre of this vicious circle facing much repression and ostracisation. The study conducted by the PDPU team was an attempt to change this dogmatic alienation of these poor women. It was found that with consistent support and reformist approach towards law, it is possible to drive these women away from a life of penury repression and crime. The aforementioned study uses empirical tools to verify this claim. Placed at the confluence of the sociology of gender and psychology, this paper is a good way to argue that law enforcement cannot be effective without sensitisation to the ground realities of conflict. The study conducted from which the paper borrows was a scientific survey focused on markers of gender and caste realities of the Chharas. The paper mentions various dynamics involved in the training program that paved the way for the successful employment of the women. In an attempt to explain its uniqueness, the paper also has a section on comparing similar social experiments.

Keywords: employment, gender, handholding, rehabilitation

Procedia PDF Downloads 110
1606 Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator

Authors: Wedad Albalawi

Abstract:

The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is defined as a closed subset contains real numbers. Then the inequalities of time scales version have received a lot of attention and has had a major field in both pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on double integrals to obtain new time-scale inequalities of Copson driven by Steklov operator. They will be applied in the solution of the Cauchy problem for the wave equation. The proof can be done by introducing restriction on the operator in several cases. In addition, the obtained inequalities done by using some concepts in time scale version such as time scales calculus, theorem of Fubini and the inequality of H¨older.

Keywords: time scales, inequality of Hardy, inequality of Coposon, Steklov operator

Procedia PDF Downloads 58
1605 Evaluation of Deformation for Deep Excavations in the Greater Vancouver Area Through Case Studies

Authors: Boris Kolev, Matt Kokan, Mohammad Deriszadeh, Farshid Bateni

Abstract:

Due to the increasing demand for real estate and the need for efficient land utilization in Greater Vancouver, developers have been increasingly considering the construction of high-rise structures with multiple below-grade parking. The temporary excavations required to allow for the construction of underground levels have recently reached up to 40 meters in depth. One of the challenges with deep excavations is the prediction of wall displacements and ground settlements due to their effect on the integrity of City utilities, infrastructure, and adjacent buildings. A large database of survey monitoring data has been collected for deep excavations in various soil conditions and shoring systems. The majority of the data collected is for tie-back anchors and shotcrete lagging systems. The data were categorized, analyzed and the results were evaluated to find a relationship between the most dominant parameters controlling the displacement, such as depth of excavation, soil properties, and the tie-back anchor loading and arrangement. For a select number of deep excavations, finite element modeling was considered for analyses. The lateral displacements from the simulation results were compared to the recorded survey monitoring data. The study concludes with a discussion and comparison of the available empirical and numerical modeling methodologies for evaluating lateral displacements in deep excavations.

Keywords: deep excavations, lateral displacements, numerical modeling, shoring walls, tieback anchors

Procedia PDF Downloads 162
1604 Analysis of Nonlinear Dynamic Systems Excited by Combined Colored and White Noise Excitations

Authors: Siu-Siu Guo, Qingxuan Shi

Abstract:

In this paper, single-degree-of-freedom (SDOF) systems to white noise and colored noise excitations are investigated. By expressing colored noise excitation as a second-order filtered white noise process and introducing colored noise as an additional state variable, the equation of motion for SDOF system under colored noise is then transferred artificially to multi-degree-of-freedom (MDOF) system under white noise excitations. As a consequence, corresponding Fokker-Planck-Kolmogorov (FPK) equation governing the joint probabilistic density function (PDF) of state variables increases to 4-dimension (4-D). Solution procedure and computer programme become much more sophisticated. The exponential-polynomial closure (EPC) method, widely applied for cases of SDOF systems under white noise excitations, is developed and improved for cases of systems under colored noise excitations and for solving the complex 4-D FPK equation. On the other hand, Monte Carlo simulation (MCS) method is performed to test the approximate EPC solutions. Two examples associated with Gaussian and non-Gaussian colored noise excitations are considered. Corresponding band-limited power spectral densities (PSDs) for colored noise excitations are separately given. Numerical studies show that the developed EPC method provides relatively accurate estimates of the stationary probabilistic solutions. Moreover, statistical parameter of mean-up crossing rate (MCR) is taken into account, which is important for reliability and failure analysis.

Keywords: filtered noise, narrow-banded noise, nonlinear dynamic, random vibration

Procedia PDF Downloads 210
1603 A Review Study on the Importance and Correlation of Crisis Literacy and Media Communications for Vulnerable Marginalized People During Crisis

Authors: Maryam Jabeen

Abstract:

In recent times, there has been a notable surge in attention towards diverse literacy concepts such as media literacy, information literacy, and digital literacy. These concepts have garnered escalating interest, spurring the emergence of novel approaches, particularly in the aftermath of the Covid-19 crisis. However, amidst discussions of crises, the domain of crisis literacy remains largely uncharted within academic exploration. Crisis literacy, also referred to as disaster literacy, denotes an individual's aptitude to not only comprehend but also effectively apply information, enabling well-informed decision-making and adherence to instructions about disaster mitigation, preparedness, response, and recovery. This theoretical and descriptive study seeks to transcend foundational literacy concepts, underscoring the urgency for an in-depth exploration of crisis literacy and its interplay with the realm of media communication. Given the profound impact of the pandemic experience and the looming uncertainty of potential future crises, there arises a pressing need to elevate crisis literacy, or disaster literacy, towards heightened autonomy and active involvement within the spheres of critical disaster preparedness, recovery initiatives, and media communication domains. This research paper is part of my ongoing Ph.D. research study, which explores on a broader level the Encoding and decoding of media communications in relation to crisis literacy. The primary objective of this research paper is to expound upon a descriptive, theoretical research endeavor delving into this domain. The emphasis lies in highlighting the paramount significance of media communications in literacy of crisis, coupled with an accentuated focus on its role in providing information to marginalized populations amidst crises. In conclusion, this research bridges the gap in crisis literacy correlation to media communications exploration, advocating for a comprehensive understanding of its dynamics and its symbiotic relationship with media communications. It intends to foster a heightened sense of crisis literacy, particularly within marginalized communities, catalyzing proactive participation in disaster preparedness, recovery processes, and adept media interactions.

Keywords: covid-19, crisis literacy, crisis, marginalized, media and communications, pandemic, vulnerable people

Procedia PDF Downloads 44
1602 Influential Health Care System Rankings Can Conceal Maximal Inequities: A Simulation Study

Authors: Samuel Reisman

Abstract:

Background: Comparative rankings are increasingly used to evaluate health care systems. These rankings combine discrete attribute rankings into a composite overall ranking. Health care equity is a component of overall rankings, but excelling in other categories can counterbalance low inequity grades. Highly ranked inequitable health care would commend systems that disregard human rights. We simulated the ranking of a maximally inequitable health care system using a published, influential ranking methodology. Methods: We used The Commonwealth Fund’s ranking of eleven health care systems to simulate the rank of a maximally inequitable system. Eighty performance indicators were simulated, assuming maximal ineptitude in equity benchmarks. Maximal rankings in all non-equity subcategories were assumed. Subsequent stepwise simulations lowered all non-equity rank positions by one. Results: The maximally non-equitable health care system ranked first overall. Three subsequent stepwise simulations, lowering non-equity rankings by one, each resulted in an overall ranking within the top three. Discussion: Our results demonstrate that grossly inequitable health care systems can rank highly in comparative health care system rankings. These findings challenge the validity of ranking methodologies that subsume equity under broader benchmarks. We advocate limiting maximum overall rankings of health care systems to their individual equity rankings. Such limits are logical given the insignificance of health care system improvements to those lacking adequate health care.

Keywords: global health, health equity, healthcare systems, international health

Procedia PDF Downloads 374
1601 Biochar-induced Metals Immobilization in the Soil as Affected by Citric Acid

Authors: Md. Shoffikul Islam, Hongqing Hu

Abstract:

Reducing trace elements' mobility and bioavailability through amendment addition, especially biochar (BC), is a cost-effective and efficient method to address their toxicity in the soil environment. However, the low molecular weight organic acids (LMWOAs) in the rhizosphere could affect BC's efficiency to stabilize trace metals as the LMWOAs could either mobilize or fix metals in the soils. Therefore, understanding the BC's and LMWOAs' interaction mechanisms on metals stabilization in the rhizosphere is crucial. The present study explored the impact of BC derived from rice husk and citric acid (CA) and the combination of BC and CA on the redistribution of cadmium (Cd), lead (Pb), and zinc (Zn) among their geochemical forms through incubation experiment. The changes of zeta potential and X-ray diffraction (XRD) pattern of BC and BC-amended soils to investigate the probable mechanisms of trace elements' immobilization by BC under the CA attack were also examined. The rice husk BC at 5% (w/w) was mixed with the air-dry soil (an Anthrosols) contaminated with Cd, Pb, and Zn in the plastic pot. The 2, 5, 10, and 20 mM kg-1 (w/v) of CA were added separately into the pot. All the ingredients were mixed thoroughly with the soil. A control (CK) treatment was also prepared without BC and CA addition. After 7, 15, and 60 days of incubation with 60% (w/v) moisture level at 25 °C, the incubated soils were determined for pH and EC and were sequentially extracted to assess the metals' transformation in soil. The electronegative charges and XRD peaks of BC and BC-amended soils were also measured. Compared to CK, the application of BC, low level of CA (2 mM kg-1 soil) (CA2), and BC plus the low concentration of CA (BC-CA2) considerably declined the acid-soluble Cd, Pb, and Zn in which BC-CA2 was found to be the most effective treatment. The reversed trends were observed concerning the high levels of CA (>5-20 mM kg-1 soil) and the BC plus high concentrations of CA treatments. BC-CA2 changed the highest amounts of acid-soluble and reducible metals to the oxidizable and residual forms with time. The most increased electronegative charges of BC-CA2 indicate its (BC-CA2) highest Cd, Pb, and Zn immobilizing efficiency, probably through metals adsorption and fixation with the negative charge sites. The XRD study revealed the presence of P, O, CO32-, and Cl1- in BC, which might be responsible for the precipitation of CdCO3, pyromorphite, and hopeite in the case of Cd, Pb, and Zn immobilization, respectively. The findings depicted that the low concentration of CA increased metals' stabilization, whereas the high levels of CA enhanced their mobilization. The BC-CA2 emerged as the best amendment among treatments for metals stabilization in contaminated soils.

Keywords: Biochar, citric acid, immobilization, trace elements contaminated soil

Procedia PDF Downloads 61
1600 Creating Smart and Healthy Cities by Exploring the Potentials of Emerging Technologies and Social Innovation for Urban Efficiency: Lessons from the Innovative City of Boston

Authors: Mohammed Agbali, Claudia Trillo, Yusuf Arayici, Terrence Fernando

Abstract:

The wide-spread adoption of the Smart City concept has introduced a new era of computing paradigm with opportunities for city administrators and stakeholders in various sectors to re-think the concept of urbanization and development of healthy cities. With the world population rapidly becoming urban-centric especially amongst the emerging economies, social innovation will assist greatly in deploying emerging technologies to address the development challenges in core sectors of the future cities. In this context, sustainable health-care delivery and improved quality of life of the people is considered at the heart of the healthy city agenda. This paper examines the Boston innovation landscape from the perspective of smart services and innovation ecosystem for sustainable development, especially in transportation and healthcare. It investigates the policy implementation process of the Healthy City agenda and eHealth economy innovation based on the experience of Massachusetts’s City of Boston initiatives. For this purpose, three emerging areas are emphasized, namely the eHealth concept, the innovation hubs, and the emerging technologies that drive innovation. This was carried out through empirical analysis on results of public sector and industry-wide interviews/survey about Boston’s current initiatives and the enabling environment. The paper highlights few potential research directions for service integration and social innovation for deploying emerging technologies in the healthy city agenda. The study therefore suggests the need to prioritize social innovation as an overarching strategy to build sustainable Smart Cities in order to avoid technology lock-in. Finally, it concludes that the Boston example of innovation economy is unique in view of the existing platforms for innovation and proper understanding of its dynamics, which is imperative in building smart and healthy cities where quality of life of the citizenry can be improved.

Keywords: computing paradigm, emerging technologies, equitable healthcare, healthy cities, open data, smart city, social innovation

Procedia PDF Downloads 315
1599 Flame Volume Prediction and Validation for Lean Blowout of Gas Turbine Combustor

Authors: Ejaz Ahmed, Huang Yong

Abstract:

The operation of aero engines has a critical importance in the vicinity of lean blowout (LBO) limits. Lefebvre’s model of LBO based on empirical correlation has been extended to flame volume concept by the authors. The flame volume takes into account the effects of geometric configuration, the complex spatial interaction of mixing, turbulence, heat transfer and combustion processes inside the gas turbine combustion chamber. For these reasons, flame volume based LBO predictions are more accurate. Although LBO prediction accuracy has improved, it poses a challenge associated with Vf estimation in real gas turbine combustors. This work extends the approach of flame volume prediction previously based on fuel iterative approximation with cold flow simulations to reactive flow simulations. Flame volume for 11 combustor configurations has been simulated and validated against experimental data. To make prediction methodology robust as required in the preliminary design stage, reactive flow simulations were carried out with the combination of probability density function (PDF) and discrete phase model (DPM) in FLUENT 15.0. The criterion for flame identification was defined. Two important parameters i.e. critical injection diameter (Dp,crit) and critical temperature (Tcrit) were identified, and their influence on reactive flow simulation was studied for Vf estimation. Obtained results exhibit ±15% error in Vf estimation with experimental data.

Keywords: CFD, combustion, gas turbine combustor, lean blowout

Procedia PDF Downloads 250
1598 The Relationship between Central Bank Independence and Inflation: Evidence from Africa

Authors: R. Bhattu Babajee, Marie Sandrine Estelle Benoit

Abstract:

The past decades have witnessed a considerable institutional shift towards Central Bank Independence across economies of the world. The motivation behind such a change is the acceptance that increased central bank autonomy has the power of alleviating inflation bias. Hence, studying whether Central Bank Independence acts as a significant factor behind the price stability in the African economies or whether this macroeconomic aim in these countries result from other economic, political or social factors is a pertinent issue. The main research objective of this paper is to assess the relationship between central bank autonomy and inflation in African economies where inflation has proved to be a serious problem. In this optic, we shall measure the degree of CBI in Africa by computing the turnover rates of central banks governors thereby studying whether decisions made by African central banks are affected by external forces. The purpose of this study is to investigate empirically the association between Central Bank Independence (CBI) and inflation for 10 African economies over a period of 17 years, from 1995 to 2012. The sample includes Botswana, Egypt, Ghana, Kenya, Madagascar, Mauritius, Mozambique, Nigeria, South Africa, and Uganda. In contrast to empirical research, we have not been using the usual static panel model for it is associated with potential mis specification arising from the absence of dynamics. To this issue a dynamic panel data model which integrates several control variables has been used. Firstly, the analysis includes dynamic terms to explain the tenacity of inflation. Given the confirmation of inflation inertia, that is very likely in African countries there exists the need for including lagged inflation in the empirical model. Secondly, due to known reverse causality between Central Bank Independence and inflation, the system generalized method of moments (GMM) is employed. With GMM estimators, the presence of unknown forms of heteroskedasticity is admissible as well as auto correlation in the error term. Thirdly, control variables have been used to enhance the efficiency of the model. The main finding of this paper is that central bank independence is negatively associated with inflation even after including control variables.

Keywords: central bank independence, inflation, macroeconomic variables, price stability

Procedia PDF Downloads 353