Search results for: ageing model
11655 Determinants of Foreign Direct Investment in Tourism: A Panel Data Analysis of Developing Countries
Authors: Malraj Bharatha Kiriella
Abstract:
The purpose of this paper is to investigate the determinants of tourism foreign direct investment (TFDI) to selected developing countries during 1978-2017. The study used pooled panel data to estimate an econometric model. The findings show that market size and institutional barriers are determining factors for TFDI in countries, while other variables of positive country conditions, FDI-related government policy, tourism-related infrastructure and labor conditions are insignificant. The result shows that institutional effects are positive, while market size negatively affects TFDI inflows. The research is limited to eight developing countries. The results can be used to support government policy on TFDI. The paper makes the following contributions: First, it provides important insight and understanding into the TFDI decision-making process in developing countries. Second, both TFDI theory and evidence are minimal, and an econometric model developed on the basis of available literature has been empirically tested.Keywords: determinants, developing countries, FDI in tourism, panel data
Procedia PDF Downloads 11011654 Theoretical Stress-Strain Model for Confined Concrete by Rectangular Reinforcement
Authors: Mizam Dogan, Hande Gökdemir
Abstract:
In reinforced concrete elements, reinforcement steel bars are placed in concrete both longitudinal and lateral directions. The lateral reinforcement (called as confinement) which is used for confining circular RC elements is in a spiral shape. If the cross section of RC element is rectangular, stirrups should be rectangular too. At very high compressive stresses concrete will reach its limit strain value and therefore concrete outside the lateral reinforcement, which is not confined, will crush and start to spell. At this stage, concrete core of the RC element tries to expand laterally as a reason of high Poisson’s ratio value of concrete. Such a deformation is prevented by the lateral reinforcement which applies lateral passive pressure on concrete. At very high compressive stresses, the strength of reinforced column member rises to four times σ 2. This increase in strength of member is related to the properties of rectangular stirrups. In this paper, effect of stirrup step spacing to column behavior is calculated and presented confined concrete model is proved by numerical solutions.Keywords: confined concrete, concrete column, stress-strain, stirrup, solid, frame
Procedia PDF Downloads 45211653 Dye Removal from Aqueous Solution by Regenerated Spent Bleaching Earth
Authors: Ahmed I. Shehab, Sabah M. Abdel Basir, M. A. Abdel Khalek, M. H. Soliman, G. Elgemeie
Abstract:
Spent bleaching earth (SBE) recycling and utilization as an adsorbent to eliminate dyes from aqueous solution was studied. Organic solvents and subsequent thermal treatment were carried out to recover and reactivate the SBE. The effect of pH, temperature, dye’s initial concentration, and contact time on the dye removal using recycled spent bleaching earth (RSBE) was investigated. Recycled SBE showed better removal affinity of cationic than anionic dyes. The maximum removal was achieved at pH 2 and 8 for anionic and cationic dyes, respectively. Kinetic data matched with the pseudo second-order model. The adsorption phenomenon governing this process was identified by the Langmuir and Freundlich isotherms for anionic dye while Freundlich model represented the sorption process for cationic dye. The changes of Gibbs free energy (ΔG°), enthalpy (ΔH°), and entropy (ΔS°) were computed and compared through thermodynamic study for both dyes.Keywords: Spent bleaching earth, reactivation, regeneration, thermal treatment, dye removal, thermodynamic
Procedia PDF Downloads 18611652 A Machine Learning-Based Model to Screen Antituberculosis Compound Targeted against LprG Lipoprotein of Mycobacterium tuberculosis
Authors: Syed Asif Hassan, Syed Atif Hassan
Abstract:
Multidrug-resistant Tuberculosis (MDR-TB) is an infection caused by the resistant strains of Mycobacterium tuberculosis that do not respond either to isoniazid or rifampicin, which are the most important anti-TB drugs. The increase in the occurrence of a drug-resistance strain of MTB calls for an intensive search of novel target-based therapeutics. In this context LprG (Rv1411c) a lipoprotein from MTB plays a pivotal role in the immune evasion of Mtb leading to survival and propagation of the bacterium within the host cell. Therefore, a machine learning method will be developed for generating a computational model that could predict for a potential anti LprG activity of the novel antituberculosis compound. The present study will utilize dataset from PubChem database maintained by National Center for Biotechnology Information (NCBI). The dataset involves compounds screened against MTB were categorized as active and inactive based upon PubChem activity score. PowerMV, a molecular descriptor generator, and visualization tool will be used to generate the 2D molecular descriptors for the actives and inactive compounds present in the dataset. The 2D molecular descriptors generated from PowerMV will be used as features. We feed these features into three different classifiers, namely, random forest, a deep neural network, and a recurring neural network, to build separate predictive models and choosing the best performing model based on the accuracy of predicting novel antituberculosis compound with an anti LprG activity. Additionally, the efficacy of predicted active compounds will be screened using SMARTS filter to choose molecule with drug-like features.Keywords: antituberculosis drug, classifier, machine learning, molecular descriptors, prediction
Procedia PDF Downloads 39311651 Survival Analysis after a First Ischaemic Stroke Event: A Case-Control Study in the Adult Population of England.
Authors: Padma Chutoo, Elena Kulinskaya, Ilyas Bakbergenuly, Nicholas Steel, Dmitri Pchejetski
Abstract:
Stroke is associated with a significant risk of morbidity and mortality. There is scarcity of research on the long-term survival after first-ever ischaemic stroke (IS) events in England with regards to effects of different medical therapies and comorbidities. The objective of this study was to model the all-cause mortality after an IS diagnosis in the adult population of England. Using a retrospective case-control design, we extracted the electronic medical records of patients born prior to or in year 1960 in England with a first-ever ischaemic stroke diagnosis from January 1986 to January 2017 within the Health and Improvement Network (THIN) database. Participants with a history of ischaemic stroke were matched to 3 controls by sex and age at diagnosis and general practice. The primary outcome was the all-cause mortality. The hazards of the all-cause mortality were estimated using a Weibull-Cox survival model which included both scale and shape effects and a shared random effect of general practice. The model included sex, birth cohort, socio-economic status, comorbidities and medical therapies. 20,250 patients with a history of IS (cases) and 55,519 controls were followed up to 30 years. From 2008 to 2015, the one-year all-cause mortality for the IS patients declined with an absolute change of -0.5%. Preventive treatments to cases increased considerably over time. These included prescriptions of statins and antihypertensives. However, prescriptions for antiplatelet drugs decreased in the routine general practice since 2010. The survival model revealed a survival benefit of antiplatelet treatment to stroke survivors with hazard ratio (HR) of 0.92 (0.90 – 0.94). IS diagnosis had significant interactions with gender and age at entry and hypertension diagnosis. IS diagnosis was associated with high risk of all-cause mortality with HR= 3.39 (3.05-3.72) for cases compared to controls. Hypertension was associated with poor survival with HR = 4.79 (4.49 - 5.09) for hypertensive cases relative to non-hypertensive controls, though the detrimental effect of hypertension has not reached significance for hypertensive controls, HR = 1.19(0.82-1.56). This study of English primary care data showed that between 2008 and 2015, the rates of prescriptions of stroke preventive treatments increased, and a short-term all-cause mortality after IS stroke declined. However, stroke resulted in poor long-term survival. Hypertension, a modifiable risk factor, was found to be associated with poor survival outcomes in IS patients. Antiplatelet drugs were found to be protective to survival. Better efforts are required to reduce the burden of stroke through health service development and primary prevention.Keywords: general practice, hazard ratio, health improvement network (THIN), ischaemic stroke, multiple imputation, Weibull-Cox model.
Procedia PDF Downloads 18911650 Development of a Regression Based Model to Predict Subjective Perception of Squeak and Rattle Noise
Authors: Ramkumar R., Gaurav Shinde, Pratik Shroff, Sachin Kumar Jain, Nagesh Walke
Abstract:
Advancements in electric vehicles have significantly reduced the powertrain noise and moving components of vehicles. As a result, in-cab noises have become more noticeable to passengers inside the car. To ensure a comfortable ride for drivers and other passengers, it has become crucial to eliminate undesirable component noises during the development phase. Standard practices are followed to identify the severity of noises based on subjective ratings, but it can be a tedious process to identify the severity of each development sample and make changes to reduce it. Additionally, the severity rating can vary from jury to jury, making it challenging to arrive at a definitive conclusion. To address this, an automotive component was identified to evaluate squeak and rattle noise issue. Physical tests were carried out for random and sine excitation profiles. Aim was to subjectively assess the noise using jury rating method and objectively evaluate the same by measuring the noise. Suitable jury evaluation method was selected for the said activity, and recorded sounds were replayed for jury rating. Objective data sound quality metrics viz., loudness, sharpness, roughness, fluctuation strength and overall Sound Pressure Level (SPL) were measured. Based on this, correlation co-efficients was established to identify the most relevant sound quality metrics that are contributing to particular identified noise issue. Regression analysis was then performed to establish the correlation between subjective and objective data. Mathematical model was prepared using artificial intelligence and machine learning algorithm. The developed model was able to predict the subjective rating with good accuracy.Keywords: BSR, noise, correlation, regression
Procedia PDF Downloads 8311649 A Geospatial Consumer Marketing Campaign Optimization Strategy: Case of Fuzzy Approach in Nigeria Mobile Market
Authors: Adeolu O. Dairo
Abstract:
Getting the consumer marketing strategy right is a crucial and complex task for firms with a large customer base such as mobile operators in a competitive mobile market. While empirical studies have made efforts to identify key constructs, no geospatial model has been developed to comprehensively assess the viability and interdependency of ground realities regarding the customer, competition, channel and the network quality of mobile operators. With this research, a geo-analytic framework is proposed for strategy formulation and allocation for mobile operators. Firstly, a fuzzy analytic network using a self-organizing feature map clustering technique based on inputs from managers and literature, which depicts the interrelationships amongst ground realities is developed. The model is tested with a mobile operator in the Nigeria mobile market. As a result, a customer-centric geospatial and visualization solution is developed. This provides a consolidated and integrated insight that serves as a transparent, logical and practical guide for strategic, tactical and operational decision making.Keywords: geospatial, geo-analytics, self-organizing map, customer-centric
Procedia PDF Downloads 18511648 A Multi-Criteria Decision Making Approach for Disassembly-To-Order Systems under Uncertainty
Authors: Ammar Y. Alqahtani
Abstract:
In order to minimize the negative impact on the environment, it is essential to manage the waste that generated from the premature disposal of end-of-life (EOL) products properly. Consequently, government and international organizations introduced new policies and regulations to minimize the amount of waste being sent to landfills. Moreover, the consumers’ awareness regards environment has forced original equipment manufacturers to consider being more environmentally conscious. Therefore, manufacturers have thought of different ways to deal with waste generated from EOL products viz., remanufacturing, reusing, recycling, or disposing of EOL products. The rate of depletion of virgin natural resources and their dependency on the natural resources can be reduced by manufacturers when EOL products are treated as remanufactured, reused, or recycled, as well as this will cut on the amount of harmful waste sent to landfills. However, disposal of EOL products contributes to the problem and therefore is used as a last option. Number of EOL need to be estimated in order to fulfill the components demand. Then, disassembly process needs to be performed to extract individual components and subassemblies. Smart products, built with sensors embedded and network connectivity to enable the collection and exchange of data, utilize sensors that are implanted into products during production. These sensors are used for remanufacturers to predict an optimal warranty policy and time period that should be offered to customers who purchase remanufactured components and products. Sensor-provided data can help to evaluate the overall condition of a product, as well as the remaining lives of product components, prior to perform a disassembly process. In this paper, a multi-period disassembly-to-order (DTO) model is developed that takes into consideration the different system uncertainties. The DTO model is solved using Nonlinear Programming (NLP) in multiple periods. A DTO system is considered where a variety of EOL products are purchased for disassembly. The model’s main objective is to determine the best combination of EOL products to be purchased from every supplier in each period which maximized the total profit of the system while satisfying the demand. This paper also addressed the impact of sensor embedded products on the cost of warranties. Lastly, this paper presented and analyzed a case study involving various simulation conditions to illustrate the applicability of the model.Keywords: closed-loop supply chains, environmentally conscious manufacturing, product recovery, reverse logistics
Procedia PDF Downloads 13911647 Investigation of Unconventional Fuels in Co-Axial Engines
Authors: Arya Pirooz
Abstract:
The effects of different fuels (DME, RME B100, and SME B100) on barrel engines were studied as a general, single dimensional investigation for characterization of these types of engines. A base computational model was created as reference point to be used as a point of comparison with different cases. The models were computed using the commercial computational fluid dynamics program, Diesel-RK. The base model was created using basic dimensions of the PAMAR-3 engine with inline unit injectors. Four fuel cases were considered. Optimized models were also considered for diesel and DME cases with respect to injection duration, fuel, injection timing, exhaust and intake port opening, CR, angular offset. These factors were optimized for highest BMEP, combined PM and NOx emissions, and highest SFC. Results included mechanical efficiency (eta_m), efficiency and power, emission characteristics, combustion characteristics. DME proved to have the highest performing characteristics in relation to diesel and RME fuels for this type of barrel engine.Keywords: DME, RME, Diesel-RK, characterization, inline unit injector
Procedia PDF Downloads 47911646 Transient/Steady Natural Convective Flow of Reactive Viscous Fluid in Vertical Porous Pipe
Authors: Ahmad K. Samaila, Basant K. Jha
Abstract:
This paper presents the effects of suction/injection of transient/steady natural convection flow of reactive viscous fluid in a vertical porous pipe. The mathematical model capturing the time dependent flow of viscous reactive fluid is solved using implicit finite difference method while the corresponding steady state model is solved using regular perturbation technique. Results of analytical and numerical solutions are reported for various parametric conditions to illustrate special features of the solutions. The coefficient of skin friction and rate of heat transfer are obtained and illustrated graphically. The numerical solution is shown to be in excellent agreement with the closed form analytical solution. It is interesting to note that time required to reach steady state is higher in case of injection in comparison to suction.Keywords: porous pipe, reactive viscous fluid, transient natural-convective flow, analytical solution
Procedia PDF Downloads 29811645 Seismic Behavior of Self-Balancing Post-Tensioned Reinforced Concrete Spatial Structure
Authors: Mircea Pastrav, Horia Constantinescu
Abstract:
The construction industry is currently trying to develop sustainable reinforced concrete structures. In trying to aid in the effort, the research presented in this paper aims to prove the efficiency of modified special hybrid moment frames composed of discretely jointed precast and post-tensioned concrete members. This aim is due to the fact that current design standards do not cover the spatial design of moment frame structures assembled by post-tensioning with special hybrid joints. This lack of standardization is coupled with the fact that previous experimental programs, available in scientific literature, deal mainly with plane structures and offer little information regarding spatial behavior. A spatial model of a modified hybrid moment frame is experimentally analyzed. The experimental results of a natural scale model test of a corner column-beams sub-structure, cut from an actual multilevel building tested to seismic type loading are presented in order to highlight the behavior of this type of structure. The test is performed under alternative cycles of imposed lateral displacements, up to a storey drift ratio of 0.035. Seismic response of the spatial model is discussed considering the acceptance criteria for reinforced concrete frame structures designed based on experimental tests, as well as some of its major sustainability features. The results obtained show an overall excellent behavior of the system. The joint detailing allows for quick and cheap repairs after an accidental event and a self-balancing behavior of the system that ensures it can be used almost immediately after an accidental event it.Keywords: modified hybrid joint, seismic type loading response, self-balancing structure, acceptance criteria
Procedia PDF Downloads 24211644 Evaluating the Diagnostic Accuracy of the ctDNA Methylation for Liver Cancer
Authors: Maomao Cao
Abstract:
Objective: To test the performance of ctDNA methylation for the detection of liver cancer. Methods: A total of 1233 individuals have been recruited in 2017. 15 male and 15 female samples (including 10 cases of liver cancer) were randomly selected in the present study. CfDNA was extracted by MagPure Circulating DNA Maxi Kit. The concentration of cfDNA was obtained by Qubit™ dsDNA HS Assay Kit. A pre-constructed predictive model was used to analyze methylation data and to give a predictive score for each cfDNA sample. Individuals with a predictive score greater than or equal to 80 were classified as having liver cancer. CT tests were considered the gold standard. Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for the diagnosis of liver cancer were calculated. Results: 9 patients were diagnosed with liver cancer according to the prediction model (with high sensitivity and threshold of 80 points), with scores of 99.2, 91.9, 96.6, 92.4, 91.3, 92.5, 96.8, 91.1, and 92.2, respectively. The sensitivity, specificity, positive predictive value, and negative predictive value of ctDNA methylation for the diagnosis of liver cancer were 0.70, 0.90, 0.78, and 0.86, respectively. Conclusions: ctDNA methylation could be an acceptable diagnostic modality for the detection of liver cancer.Keywords: liver cancer, ctDNA methylation, detection, diagnostic performance
Procedia PDF Downloads 15611643 A Proposal of Multi-modal Teaching Model for College English
Authors: Huang Yajing
Abstract:
Multimodal discourse refers to the phenomenon of using various senses such as hearing, vision, and touch to communicate through various means and symbolic resources such as language, images, sounds, and movements. With the development of modern technology and multimedia, language and technology have become inseparable, and foreign language teaching is becoming more and more modal. Teacher-student communication resorts to multiple senses and uses multiple symbol systems to construct and interpret meaning. The classroom is a semiotic space where multimodal discourses are intertwined. College English multi-modal teaching is to rationally utilize traditional teaching methods while mobilizing and coordinating various modern teaching methods to form a joint force to promote teaching and learning. Multimodal teaching makes full and reasonable use of various meaning resources and can maximize the advantages of multimedia and network environments. Based upon the above theories about multimodal discourse and multimedia technology, the present paper will propose a multi-modal teaching model for college English in China.Keywords: multimodal discourse, multimedia technology, English education, applied linguistics
Procedia PDF Downloads 7111642 Identification and Force Control of a Two Chambers Pneumatic Soft Actuator
Authors: Najib K. Dankadai, Ahmad 'Athif Mohd Faudzi, Khairuddin Osman, Muhammad Rusydi Muhammad Razif, IIi Najaa Aimi Mohd Nordin
Abstract:
Researches in soft actuators are now growing rapidly because of their adequacy to be applied in sectors like medical, agriculture, biological and welfare. This paper presents system identification (SI) and control of the force generated by a two chambers pneumatic soft actuator (PSA). A force mathematical model for the actuator was identified experimentally using data acquisition card and MATLAB SI toolbox. Two control techniques; a predictive functional control (PFC) and conventional proportional integral and derivative (PID) schemes are proposed and compared based on the identified model for the soft actuator flexible mechanism. Results of this study showed that both of the proposed controllers ensure accurate tracking when the closed loop system was tested with the step, sinusoidal and multi step reference input through MATLAB simulation although the PFC provides a better response than the PID.Keywords: predictive functional control (PFC), proportional integral and derivative (PID), soft actuator, system identification
Procedia PDF Downloads 32711641 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization
Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman
Abstract:
In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization
Procedia PDF Downloads 24311640 Evaluation of NH3-Slip from Diesel Vehicles Equipped with Selective Catalytic Reduction Systems by Neural Networks Approach
Authors: Mona Lisa M. Oliveira, Nara A. Policarpo, Ana Luiza B. P. Barros, Carla A. Silva
Abstract:
Selective catalytic reduction systems for nitrogen oxides reduction by ammonia has been the chosen technology by most of diesel vehicle (i.e. bus and truck) manufacturers in Brazil, as also in Europe. Furthermore, at some conditions, over-stoichiometric ammonia availability is also needed that increases the NH3 slips even more. Ammonia (NH3) by this vehicle exhaust aftertreatment system provides a maximum efficiency of NOx removal if a significant amount of NH3 is stored on its catalyst surface. In the other words, the practice shows that slightly less than 100% of the NOx conversion is usually targeted, so that the aqueous urea solution hydrolyzes to NH3 via other species formation, under relatively low temperatures. This paper presents a model based on neural networks integrated with a road vehicle simulator that allows to estimate NH3-slip emission factors for different driving conditions and patterns. The proposed model generates high NH3slips which are not also limited in Brazil, but more efforts needed to be made to elucidate the contribution of vehicle-emitted NH3 to the urban atmosphere.Keywords: ammonia slip, neural-network, vehicles emissions, SCR-NOx
Procedia PDF Downloads 21711639 Dynamics of a Reaction-Diffusion Problems Modeling Two Predators Competing for a Prey
Authors: Owolabi Kolade Matthew
Abstract:
In this work, we investigate both the analytical and numerical studies of the dynamical model comprising of three species system. We analyze the linear stability of stationary solutions in the one-dimensional multi-system modeling the interactions of two predators and one prey species. The stability analysis has a lot of implications for understanding the various spatiotemporal and chaotic behaviors of the species in the spatial domain. The analysis results presented have established the possibility of the three interacting species to coexist harmoniously, this feat is achieved by combining the local and global analyzes to determine the global dynamics of the system. In the presence of diffusion, a viable exponential time differencing method is applied to multi-species nonlinear time-dependent partial differential equation to address the points and queries that may naturally arise. The scheme is described in detail, and justified by a number of computational experiments.Keywords: asymptotically stable, coexistence, exponential time differencing method, global and local stability, predator-prey model, nonlinear, reaction-diffusion system
Procedia PDF Downloads 41511638 Profit-Based Artificial Neural Network (ANN) Trained by Migrating Birds Optimization: A Case Study in Credit Card Fraud Detection
Authors: Ashkan Zakaryazad, Ekrem Duman
Abstract:
A typical classification technique ranks the instances in a data set according to the likelihood of belonging to one (positive) class. A credit card (CC) fraud detection model ranks the transactions in terms of probability of being fraud. In fact, this approach is often criticized, because firms do not care about fraud probability but about the profitability or costliness of detecting a fraudulent transaction. The key contribution in this study is to focus on the profit maximization in the model building step. The artificial neural network proposed in this study works based on profit maximization instead of minimizing the error of prediction. Moreover, some studies have shown that the back propagation algorithm, similar to other gradient–based algorithms, usually gets trapped in local optima and swarm-based algorithms are more successful in this respect. In this study, we train our profit maximization ANN using the Migrating Birds optimization (MBO) which is introduced to literature recently.Keywords: neural network, profit-based neural network, sum of squared errors (SSE), MBO, gradient descent
Procedia PDF Downloads 47611637 Estimation of Coefficient of Discharge of Side Trapezoidal Labyrinth Weir Using Group Method of Data Handling Technique
Authors: M. A. Ansari, A. Hussain, A. Uddin
Abstract:
A side weir is a flow diversion structure provided in the side wall of a channel to divert water from the main channel to a branch channel. The trapezoidal labyrinth weir is a special type of weir in which crest length of the weir is increased to pass higher discharge. Experimental and numerical studies related to the coefficient of discharge of trapezoidal labyrinth weir in an open channel have been presented in the present study. Group Method of Data Handling (GMDH) with the transfer function of quadratic polynomial has been used to predict the coefficient of discharge for the side trapezoidal labyrinth weir. A new model is developed for coefficient of discharge of labyrinth weir by regression method. Generalized models for predicting the coefficient of discharge for labyrinth weir using Group Method of Data Handling (GMDH) network have also been developed. The prediction based on GMDH model is more satisfactory than those given by traditional regression equations.Keywords: discharge coefficient, group method of data handling, open channel, side labyrinth weir
Procedia PDF Downloads 16211636 Steady State Charge Transport in Quantum Dots: Nonequilibrium Green's Function (NEGF) vs. Single Electron Analysis
Authors: Mahesh Koti
Abstract:
In this paper, we present a quantum transport study of a quantum dot in steady state in the presence of static gate potential. We consider a quantum dot coupled to the two metallic leads. The quantum dot under study is modeled through Anderson Impurity Model (AIM) with hopping parameter modulated through voltage drop between leads and the central dot region. Based on the Landauer's formula derived from Nonequilibrium Green's Function and Single Electron Theory, the essential ingredients of transport properties are revealed. We show that the results out of two approaches closely agree with each other. We demonstrate that Landauer current response derived from single electron approach converges with non-zero interaction through gate potential whereas Landauer current response derived from Nonequilibrium Green's Function (NEGF) hits a pole.Keywords: Anderson impurity model (AIM), nonequilibrium Green's function (NEGF), Landauer's formula, single electron analysis
Procedia PDF Downloads 47411635 Instability Index Method and Logistic Regression to Assess Landslide Susceptibility in County Route 89, Taiwan
Authors: Y. H. Wu, Ji-Yuan Lin, Yu-Ming Liou
Abstract:
This study aims to set up the landslide susceptibility map of County Route 89 at Ren-Ai Township in Nantou County using the Instability Index Method and Logistic regression. Seven susceptibility factors including Slope Angle, Aspect, Elevation, Distance to fold, Distance to River, Distance to Road and Accumulated Rainfall were obtained by GIS based on the Typhoon Toraji landslide area identified by Industrial Technology Research Institute in 2001. To calculate the landslide percentage of each factor and acquire the weight and grade the grid by means of Instability Index Method. In this study, landslide susceptibility can be classified into four grades: high, medium high, medium low and low, in order to determine the advantages and disadvantages of the two models. The precision of this model is verified by classification error matrix and SRC curve. These results suggest that the logistic regression model is a preferred method than instability index in the assessment of landslide susceptibility. It is suitable for the landslide prediction and precaution in this area in the future.Keywords: instability index method, logistic regression, landslide susceptibility, SRC curve
Procedia PDF Downloads 29511634 Evaluating Oman's Green Transition: A Dynamic Stochastic General Equilibrium Analysis of Climate Policy Effects
Authors: Mohamed Chakroun
Abstract:
In this paper, we analyze the macroeconomic impacts of Oman’s strategy to transition to a green economy by 2050. Our objective is to determine the most effective climate policy instrument to facilitate this transition. By utilizing a Dynamic Stochastic General Equilibrium (DSGE) model, we assess the effectiveness of three climate policy tools: a carbon tax, subsidies to green assets, and taxes on brown assets. Our results indicate that a combination of a carbon tax, along with differentiated taxes and subsidies on green and brown assets, proves to the most effective policy in reducing emissions while maintaining macroeconomic stability. The findings of this study demonstrate the need for policymakers to balance the immediate goals of reducing emissions with the economic costs involved. Implementing a gradual transition strategy may be preferable as it allows for mitigating the negative economic impacts while facilitating the shift towards a green economy.Keywords: green economy, carbon tax, DSGE model, climate policy, sustainable growth
Procedia PDF Downloads 3411633 Fault Detection of Pipeline in Water Distribution Network System
Authors: Shin Je Lee, Go Bong Choi, Jeong Cheol Seo, Jong Min Lee, Gibaek Lee
Abstract:
Water pipe network is installed underground and once equipped; it is difficult to recognize the state of pipes when the leak or burst happens. Accordingly, post management is often delayed after the fault occurs. Therefore, the systematic fault management system of water pipe network is required to prevent the accident and minimize the loss. In this work, we develop online fault detection system of water pipe network using data of pipes such as flow rate or pressure. The transient model describing water flow in pipelines is presented and simulated using Matlab. The fault situations such as the leak or burst can be also simulated and flow rate or pressure data when the fault happens are collected. Faults are detected using statistical methods of fast Fourier transform and discrete wavelet transform, and they are compared to find which method shows the better fault detection performance.Keywords: fault detection, water pipeline model, fast Fourier transform, discrete wavelet transform
Procedia PDF Downloads 51311632 Demand Response from Residential Air Conditioning Load Using a Programmable Communication Thermostat
Authors: Saurabh Chanana, Monika Arora
Abstract:
Demand response is getting increased attention these days due to the increase in electricity demand and introduction of renewable resources in the existing power grid. Traditionally demand response programs involve large industrial consumers but with technological advancement, demand response is being implemented for small residential and commercial consumers also. In this paper, demand response program aims to reduce the peak demand as well as overall energy consumption of the residential customers. Air conditioners are the major reason of peak load in residential sector in summer, so a dynamic model of air conditioning load with thermostat action has been considered for applying demand response programs. A programmable communicating thermostat (PCT) is a device that uses real time pricing (RTP) signals to control the thermostat setting. A new model incorporating PCT in air conditioning load has been proposed in this paper. Results show that introduction of PCT in air conditioner is useful in reducing the electricity payments of customers as well as reducing the peak demand.Keywords: demand response, home energy management, programmable communicating thermostat, thermostatically controlled appliances
Procedia PDF Downloads 61011631 Automatic Classification of Periodic Heart Sounds Using Convolutional Neural Network
Authors: Jia Xin Low, Keng Wah Choo
Abstract:
This paper presents an automatic normal and abnormal heart sound classification model developed based on deep learning algorithm. MITHSDB heart sounds datasets obtained from the 2016 PhysioNet/Computing in Cardiology Challenge database were used in this research with the assumption that the electrocardiograms (ECG) were recorded simultaneously with the heart sounds (phonocardiogram, PCG). The PCG time series are segmented per heart beat, and each sub-segment is converted to form a square intensity matrix, and classified using convolutional neural network (CNN) models. This approach removes the need to provide classification features for the supervised machine learning algorithm. Instead, the features are determined automatically through training, from the time series provided. The result proves that the prediction model is able to provide reasonable and comparable classification accuracy despite simple implementation. This approach can be used for real-time classification of heart sounds in Internet of Medical Things (IoMT), e.g. remote monitoring applications of PCG signal.Keywords: convolutional neural network, discrete wavelet transform, deep learning, heart sound classification
Procedia PDF Downloads 35111630 Study of Flow-Induced Noise Control Effects on Flat Plate through Biomimetic Mucus Injection
Authors: Chen Niu, Xuesong Zhang, Dejiang Shang, Yongwei Liu
Abstract:
Fishes can secrete high molecular weight fluid on their body skin to enable their rapid movement in the water. In this work, we employ a hybrid method that combines Computational Fluid Dynamics (CFD) and Finite Element Method (FEM) to investigate the effects of different mucus viscosities and injection velocities on fluctuation pressure in the boundary layer and flow-induced structural vibration noise of a flat plate model. To accurately capture the transient flow distribution on the plate surface, we use Large Eddy Simulation (LES) while the mucus inlet is positioned at a sufficient distance from the model to ensure effective coverage. Mucus injection is modeled using the Volume of Fluid (VOF) method for multiphase flow calculations. The results demonstrate that mucus control of pulsating pressure effectively reduces flow-induced structural vibration noise, providing an approach for controlling flow-induced noise in underwater vehicles.Keywords: mucus, flow control, noise control, flow-induced noise
Procedia PDF Downloads 14911629 A General Iterative Nonlinear Programming Method to Synthesize Heat Exchanger Network
Authors: Rupu Yang, Cong Toan Tran, Assaad Zoughaib
Abstract:
The work provides an iterative nonlinear programming method to synthesize a heat exchanger network by manipulating the trade-offs between the heat load of process heat exchangers (HEs) and utilities. We consider for the synthesis problem two cases, the first one without fixed cost for HEs, and the second one with fixed cost. For the no fixed cost problem, the nonlinear programming (NLP) model with all the potential HEs is optimized to obtain the global optimum. For the case with fixed cost, the NLP model is iterated through adding/removing HEs. The method was applied in five case studies and illustrated quite well effectiveness. Among which, the approach reaches the lowest TAC (2,904,026$/year) compared with the best record for the famous Aromatic plants problem. It also locates a slightly better design than records in literature for a 10 streams case without fixed cost with only 1/9 computational time. Moreover, compared to the traditional mixed-integer nonlinear programming approach, the iterative NLP method opens a possibility to consider constraints (such as controllability or dynamic performances) that require knowing the structure of the network to be calculated.Keywords: heat exchanger network, synthesis, NLP, optimization
Procedia PDF Downloads 16711628 Investigations into Effect of Neural Network Predictive Control of UPFC for Improving Transient Stability Performance of Multimachine Power System
Authors: Sheela Tiwari, R. Naresh, R. Jha
Abstract:
The paper presents an investigation into the effect of neural network predictive control of UPFC on the transient stability performance of a multi-machine power system. The proposed controller consists of a neural network model of the test system. This model is used to predict the future control inputs using the damped Gauss-Newton method which employs ‘backtracking’ as the line search method for step selection. The benchmark 2 area, 4 machine system that mimics the behavior of large power systems is taken as the test system for the study and is subjected to three phase short circuit faults at different locations over a wide range of operating conditions. The simulation results clearly establish the robustness of the proposed controller to the fault location, an increase in the critical clearing time for the circuit breakers and an improved damping of the power oscillations as compared to the conventional PI controller.Keywords: identification, neural networks, predictive control, transient stability, UPFC
Procedia PDF Downloads 37611627 Preparation of Papers - Developing a Leukemia Diagnostic System Based on Hybrid Deep Learning Architectures in Actual Clinical Environments
Authors: Skyler Kim
Abstract:
An early diagnosis of leukemia has always been a challenge to doctors and hematologists. On a worldwide basis, it was reported that there were approximately 350,000 new cases in 2012, and diagnosing leukemia was time-consuming and inefficient because of an endemic shortage of flow cytometry equipment in current clinical practice. As the number of medical diagnosis tools increased and a large volume of high-quality data was produced, there was an urgent need for more advanced data analysis methods. One of these methods was the AI approach. This approach has become a major trend in recent years, and several research groups have been working on developing these diagnostic models. However, designing and implementing a leukemia diagnostic system in real clinical environments based on a deep learning approach with larger sets remains complex. Leukemia is a major hematological malignancy that results in mortality and morbidity throughout different ages. We decided to select acute lymphocytic leukemia to develop our diagnostic system since acute lymphocytic leukemia is the most common type of leukemia, accounting for 74% of all children diagnosed with leukemia. The results from this development work can be applied to all other types of leukemia. To develop our model, the Kaggle dataset was used, which consists of 15135 total images, 8491 of these are images of abnormal cells, and 5398 images are normal. In this paper, we design and implement a leukemia diagnostic system in a real clinical environment based on deep learning approaches with larger sets. The proposed diagnostic system has the function of detecting and classifying leukemia. Different from other AI approaches, we explore hybrid architectures to improve the current performance. First, we developed two independent convolutional neural network models: VGG19 and ResNet50. Then, using both VGG19 and ResNet50, we developed a hybrid deep learning architecture employing transfer learning techniques to extract features from each input image. In our approach, fusing the features from specific abstraction layers can be deemed as auxiliary features and lead to further improvement of the classification accuracy. In this approach, features extracted from the lower levels are combined into higher dimension feature maps to help improve the discriminative capability of intermediate features and also overcome the problem of network gradient vanishing or exploding. By comparing VGG19 and ResNet50 and the proposed hybrid model, we concluded that the hybrid model had a significant advantage in accuracy. The detailed results of each model’s performance and their pros and cons will be presented in the conference.Keywords: acute lymphoblastic leukemia, hybrid model, leukemia diagnostic system, machine learning
Procedia PDF Downloads 18811626 An Empirical Study for the Data-Driven Digital Transformation of the Indian Telecommunication Service Providers
Authors: S. Jigna, K. Nanda Kumar, T. Anna
Abstract:
Being a major contributor to the Indian economy and a critical facilitator for the country’s digital India vision, the Indian telecommunications industry is also a major source of employment for the country. Since the last few years, the Indian telecommunication service providers (TSPs), however, are facing business challenges related to increasing competition, losses, debts, and decreasing revenue. The strategic use of digital technologies for a successful digital transformation has the potential to equip organizations to meet these business challenges. Despite an increased focus on digital transformation, the telecom service providers globally, including Indian TSPs, have seen limited success so far. The purpose of this research was thus to identify the factors that are critical for the digital transformation and to what extent they influence the successful digital transformation of the Indian TSPs. The literature review of more than 300 digital transformation-related articles, mostly from 2013-2019, demonstrated a lack of an empirical model consisting of factors for the successful digital transformation of the TSPs. This study theorizes a research framework grounded in multiple theories, and a research model consisting of 7 constructs that may be influencing business success during the digital transformation of the organization was proposed. The questionnaire survey of senior managers in the Indian telecommunications industry was seeking to validate the research model. Based on 294 survey responses, the validation of the Structural equation model using the statistical tool ADANCO 2.1.1 was found to be robust. Results indicate that Digital Capabilities, Digital Strategy, and Corporate Level Data Strategy in that order has a strong influence on the successful Business Performance, followed by IT Function Transformation, Digital Innovation, and Transformation Management respectively. Even though Digital Organization did not have a direct significance on Business Performance outcomes, it had a strong influence on IT Function Transformation, thus affecting the Business Performance outcomes indirectly. Amongst numerous practical and theoretical contributions of the study, the main contribution for the Indian TSPs is a validated reference for prioritizing the transformation initiatives in their strategic roadmap. Also, the main contribution to the theory is the possibility to use the research framework artifact of the present research for quantitative validation in different industries and geographies.Keywords: corporate level data strategy, digital capabilities, digital innovation, digital strategy
Procedia PDF Downloads 132