Search results for: energy demand model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24553

Search results for: energy demand model

19243 Study on the Voltage Induced Wrinkling of Elastomer with Different Electrode Areas

Authors: Zhende Hou, Fan Yang, Guoli Zhang

Abstract:

Dielectric elastomer is a promising class of Electroactive polymers which can deform in response to an applied electric field. Comparing general smart material, the Dielectric elastomer is more compliance and can achieve higher energy density, which can be for diverse applications such as actuators, artificial muscles, soft robotics, and energy harvesters. The coupling of the Electroactive polymers and the electric field is that the elastomer is sandwiched between two compliant electrodes and when the electrodes are subjected to a voltage, the positive and negative charges on the two electrodes compress the polymer, so that the polymer reduces in thickness and expands in area. However, the pre-stretched dielectric elastomer film not only can achieve large electric-field induced deformation but also is prone to wrinkling, under the interaction of its own strain energy and the applied electric field energy. For a uniaxially pre-stretched dielectric elastomer film, the electrode area is an important parameter to the electric-field induced deformation and may also be a key factor affecting the film wrinkling. To determine and quantify the effect experimentally, VHB 9473 tapes were employed and compliant electrodes with different areas were pant on each of them. The tape was first tensed to a uniaxial stretch of 8. Then a DC voltage was applied to the electrodes and increased gradually until wrinkling occurred in the film. Then, the critical wrinkling voltages of the film with different electrode areas were obtained, and the wrinkle wavelengths were obtained simultaneously for analyzing the wrinkling characteristics. Experimental results indicate when the electrode area is smaller the wrinkling voltage is higher, and with the increases of electrode area, the wrinkling voltage decreases rapidly until a specific area. Beyond that, the wrinkling voltage becomes larger gradually with the increases of the area. While the wrinkle wavelength decreases gradually with the increase of voltage monotonically. That is, the relation between the critical wrinkling voltage and the electrode areas is U-shaped. Analysis believes that the film wrinkling is a kind of local effect, the interaction and the energy transfer between electrode region and non-electrode region have great influence on wrinkling. In the experiment, very thin copper wires are used as the electrode leads that just contact with the electrodes, which can avoid the stiffness of the leads affecting the wrinkling.

Keywords: elastomers, uniaxial stretch, electrode area, wrinkling

Procedia PDF Downloads 228
19242 A Novel Machining Method and Tool-Path Generation for Bent Mandrel

Authors: Hong Lu, Yongquan Zhang, Wei Fan, Xiangang Su

Abstract:

Bent mandrel has been widely used as precise mould in automobile industry, shipping industry and aviation industry. To improve the versatility and efficiency of turning method of bent mandrel with fixed rotational center, an instantaneous machining model based on cutting parameters and machine dimension is prospered in this paper. The spiral-like tool path generation approach in non-axisymmetric turning process of bent mandrel is developed as well to deal with the error of part-to-part repeatability in existed turning model. The actual cutter-location points are calculated by cutter-contact points, which are obtained from the approach of spiral sweep process using equal-arc-length segment principle in polar coordinate system. The tool offset is set to avoid the interference between tool and work piece is also considered in the machining model. Depend on the spindle rotational angle, synchronization control of X-axis, Z-axis and C-axis is adopted to generate the tool-path of the turning process. The simulation method is developed to generate NC program according to the presented model, which includes calculation of cutter-location points and generation of tool-path of cutting process. With the approach of a bent mandrel taken as an example, the maximum offset of center axis is 4mm in the 3D space. Experiment results verify that the machining model and turning method are appropriate for the characteristics of bent mandrel.

Keywords: bent mandrel, instantaneous machining model, simulation method, tool-path generation

Procedia PDF Downloads 323
19241 Assessing Effects of an Intervention on Bottle-Weaning and Reducing Daily Milk Intake from Bottles in Toddlers Using Two-Part Random Effects Models

Authors: Yungtai Lo

Abstract:

Two-part random effects models have been used to fit semi-continuous longitudinal data where the response variable has a point mass at 0 and a continuous right-skewed distribution for positive values. We review methods proposed in the literature for analyzing data with excess zeros. A two-part logit-log-normal random effects model, a two-part logit-truncated normal random effects model, a two-part logit-gamma random effects model, and a two-part logit-skew normal random effects model were used to examine effects of a bottle-weaning intervention on reducing bottle use and daily milk intake from bottles in toddlers aged 11 to 13 months in a randomized controlled trial. We show in all four two-part models that the intervention promoted bottle-weaning and reduced daily milk intake from bottles in toddlers drinking from a bottle. We also show that there are no differences in model fit using either the logit link function or the probit link function for modeling the probability of bottle-weaning in all four models. Furthermore, prediction accuracy of the logit or probit link function is not sensitive to the distribution assumption on daily milk intake from bottles in toddlers not off bottles.

Keywords: two-part model, semi-continuous variable, truncated normal, gamma regression, skew normal, Pearson residual, receiver operating characteristic curve

Procedia PDF Downloads 335
19240 Optimal Selling Prices for Small Sized Poultry Farmers

Authors: Hidefumi Kawakatsu, Dong Li, Kosuke Kato

Abstract:

In Japan, meat-type chickens are mainly classified into three categories: (1) Broilers, (2) Branded chickens, and (3) Jidori (Free-range local traditional pedigree chickens). The Jidori chickens are certified by the Japanese Ministry of Agriculture, whilst, for the Branded chickens, there is no regulation with respect to their breed (genotype) or methods for rearing them. It is, therefore, relatively easy for poultry farmers to introduce Branded than Jidori chickens. The Branded chickens are normally fed a low-calorie diet with ingredients such as herbs, which lengthens their breeding period (compared with that of the Broilers) and increases their market value. In the field of inventory management, fast-growing animals such as broilers are categorised as ameliorating items. To the best of our knowledge, there are no previous studies that have explicitly considered smaller sized poultry farmers with limited breeding areas. This study develops an inventory model for a small sized poultry farmer that produces both the Broilers (Product 1) and the Branded chickens (Product 2) with different amelioration rates. The poultry farmer’s total profit per unit of time is formulated as a function of selling prices by using a price-dependent demand function. The existence of a unique optimal selling price for each product, which maximises the total profit, established. It has also been confirmed through numerical examples that, when the breeding area is fixed, the total profit could increase if the poultry farmer reduced the product quantity of Product 1 to introduce Product 2.

Keywords: amelioration, deterioration, small sized poultry farmers, optimal price

Procedia PDF Downloads 196
19239 DenseNet and Autoencoder Architecture for COVID-19 Chest X-Ray Image Classification and Improved U-Net Lung X-Ray Segmentation

Authors: Jonathan Gong

Abstract:

Purpose AI-driven solutions are at the forefront of many pathology and medical imaging methods. Using algorithms designed to better the experience of medical professionals within their respective fields, the efficiency and accuracy of diagnosis can improve. In particular, X-rays are a fast and relatively inexpensive test that can diagnose diseases. In recent years, X-rays have not been widely used to detect and diagnose COVID-19. The under use of Xrays is mainly due to the low diagnostic accuracy and confounding with pneumonia, another respiratory disease. However, research in this field has expressed a possibility that artificial neural networks can successfully diagnose COVID-19 with high accuracy. Models and Data The dataset used is the COVID-19 Radiography Database. This dataset includes images and masks of chest X-rays under the labels of COVID-19, normal, and pneumonia. The classification model developed uses an autoencoder and a pre-trained convolutional neural network (DenseNet201) to provide transfer learning to the model. The model then uses a deep neural network to finalize the feature extraction and predict the diagnosis for the input image. This model was trained on 4035 images and validated on 807 separate images from the ones used for training. The images used to train the classification model include an important feature: the pictures are cropped beforehand to eliminate distractions when training the model. The image segmentation model uses an improved U-Net architecture. This model is used to extract the lung mask from the chest X-ray image. The model is trained on 8577 images and validated on a validation split of 20%. These models are calculated using the external dataset for validation. The models’ accuracy, precision, recall, f1-score, IOU, and loss are calculated. Results The classification model achieved an accuracy of 97.65% and a loss of 0.1234 when differentiating COVID19-infected, pneumonia-infected, and normal lung X-rays. The segmentation model achieved an accuracy of 97.31% and an IOU of 0.928. Conclusion The models proposed can detect COVID-19, pneumonia, and normal lungs with high accuracy and derive the lung mask from a chest X-ray with similarly high accuracy. The hope is for these models to elevate the experience of medical professionals and provide insight into the future of the methods used.

Keywords: artificial intelligence, convolutional neural networks, deep learning, image processing, machine learning

Procedia PDF Downloads 114
19238 Analytical Solution for Stellar Distance Based on Photon Dominated Cosmic Expansion Model

Authors: Xiaoyun Li, Suoang Longzhou

Abstract:

This paper derives the analytical solution of stellar distance according to its redshift based on the photon-dominated universe expansion model. Firstly, it calculates stellar separation speed and the farthest distance of observable stars via simulation. Then the analytical solution of stellar distance according to its redshift is derived. It shows that when the redshift is large, the stellar distance (and its separation speed) is not proportional to its redshift due to the relativity effect. It also reveals the relationship between stellar age and its redshift. The correctness of the analytical solution is verified by the latest astronomic observations of Ia supernovas in 2020.

Keywords: redshift, cosmic expansion model, analytical solution, stellar distance

Procedia PDF Downloads 150
19237 Knowledge Audit Model for Requirement Elicitation Process

Authors: Laleh Taheri, Noraini C. Pa, Rusli Abdullah, Salfarina Abdullah

Abstract:

Knowledge plays an important role to the success of any organization. Software development organizations are highly knowledge-intensive organizations especially in their Requirement Elicitation Process (REP). There are several problems regarding communicating and using the knowledge in REP such as misunderstanding, being out of scope, conflicting information and changes of requirements. All of these problems occurred in transmitting the requirements knowledge during REP. Several researches have been done in REP in order to solve the problem towards requirements. Knowledge Audit (KA) approaches were proposed in order to solve managing knowledge in human resources, financial, and manufacturing. There is lack of study applying the KA in requirements elicitation process. Therefore, this paper proposes a KA model for REP in supporting to acquire good requirements.

Keywords: knowledge audit, requirement elicitation process, KA model, knowledge in requirement elicitation

Procedia PDF Downloads 328
19236 Preference for Housing Services and Rational House Price Bubbles

Authors: Stefanie Jeanette Huber

Abstract:

This paper explores the relevance and implications of preferences for housing services on house price fluctuations through the lens of an overlapping generation’s model. The model implies that an economy whose agents have lower preferences for housing services is characterized with lower expenditure shares on housing services and will tend to experience more frequent and more volatile housing bubbles. These model predictions are tested empirically in the companion paper Housing Booms and Busts - Convergences and Divergences across OECD countries. Between 1970 - 2013, countries who spend less on housing services as a share of total income experienced significantly more housing cycles and the associated housing boom-bust cycles were more violent. Finally, the model is used to study the impact of rental subsidies and help-to-buy schemes on rational housing bubbles. Rental subsidies are found to contribute to the control of housing bubbles, whereas help-to- buy scheme makes the economy more bubble-prone.

Keywords: housing bubbles, housing booms and busts, preference for housing services, expenditure shares for housing services, rental and purchase subsidies

Procedia PDF Downloads 284
19235 Autonomous Quantum Competitive Learning

Authors: Mohammed A. Zidan, Alaa Sagheer, Nasser Metwally

Abstract:

Real-time learning is an important goal that most of artificial intelligence researches try to achieve it. There are a lot of problems and applications which require low cost learning such as learn a robot to be able to classify and recognize patterns in real time and real-time recall. In this contribution, we suggest a model of quantum competitive learning based on a series of quantum gates and additional operator. The proposed model enables to recognize any incomplete patterns, where we can increase the probability of recognizing the pattern at the expense of the undesired ones. Moreover, these undesired ones could be utilized as new patterns for the system. The proposed model is much better compared with classical approaches and more powerful than the current quantum competitive learning approaches.

Keywords: competitive learning, quantum gates, quantum gates, winner-take-all

Procedia PDF Downloads 448
19234 Predicting Indonesia External Debt Crisis: An Artificial Neural Network Approach

Authors: Riznaldi Akbar

Abstract:

In this study, we compared the performance of the Artificial Neural Network (ANN) model with back-propagation algorithm in correctly predicting in-sample and out-of-sample external debt crisis in Indonesia. We found that exchange rate, foreign reserves, and exports are the major determinants to experiencing external debt crisis. The ANN in-sample performance provides relatively superior results. The ANN model is able to classify correctly crisis of 89.12 per cent with reasonably low false alarms of 7.01 per cent. In out-of-sample, the prediction performance fairly deteriorates compared to their in-sample performances. It could be explained as the ANN model tends to over-fit the data in the in-sample, but it could not fit the out-of-sample very well. The 10-fold cross-validation has been used to improve the out-of-sample prediction accuracy. The results also offer policy implications. The out-of-sample performance could be very sensitive to the size of the samples, as it could yield a higher total misclassification error and lower prediction accuracy. The ANN model could be used to identify past crisis episodes with some accuracy, but predicting crisis outside the estimation sample is much more challenging because of the presence of uncertainty.

Keywords: debt crisis, external debt, artificial neural network, ANN

Procedia PDF Downloads 426
19233 Optimal Trajectory Finding of IDP Ventilation Control with Outdoor Air Information and Indoor Health Risk Index

Authors: Minjeong Kim, Seungchul Lee, Iman Janghorban Esfahani, Jeong Tai Kim, ChangKyoo Yoo

Abstract:

A trajectory of set-point of ventilation control systems plays an important role for efficient ventilation inside subway stations since it affects the level of indoor air pollutants and ventilation energy consumption. To maintain indoor air quality (IAQ) at a comfortable range with lower ventilation energy consumption, the optimal trajectory of the ventilation control system needs to be determined. The concentration of air pollutants inside the station shows a diurnal variation in accordance with the variations in the number of passengers and subway frequency. To consider the diurnal variation of IAQ, an iterative dynamic programming (IDP) that searches for a piecewise control policy by separating whole duration into several stages is used. When outdoor air is contaminated by pollutants, it enters the subway station through the ventilation system, which results in the deteriorated IAQ and adverse effects on passenger health. In this study, to consider the influence of outdoor air quality (OAQ), a new performance index of the IDP with the passenger health risk and OAQ is proposed. This study was carried out for an underground subway station at Seoul Metro, Korea. The optimal set-points of the ventilation control system are determined every 3 hours, then, the ventilation controller adjusts the ventilation fan speed according to the optimal set-point changes. Compared to manual ventilation system which is operated irrespective of the OAQ, the IDP-based ventilation control system saves 3.7% of the energy consumption. Compared to the fixed set-point controller which is operated irrespective of the IAQ diurnal variation, the IDP-based controller shows better performance with a 2% decrease in energy consumption, maintaining the comfortable IAQ range inside the station.

Keywords: indoor air quality, iterative dynamic algorithm, outdoor air information, ventilation control system

Procedia PDF Downloads 491
19232 Failure Inference and Optimization for Step Stress Model Based on Bivariate Wiener Model

Authors: Soudabeh Shemehsavar

Abstract:

In this paper, we consider the situation under a life test, in which the failure time of the test units are not related deterministically to an observable stochastic time varying covariate. In such a case, the joint distribution of failure time and a marker value would be useful for modeling the step stress life test. The problem of accelerating such an experiment is considered as the main aim of this paper. We present a step stress accelerated model based on a bivariate Wiener process with one component as the latent (unobservable) degradation process, which determines the failure times and the other as a marker process, the degradation values of which are recorded at times of failure. Parametric inference based on the proposed model is discussed and the optimization procedure for obtaining the optimal time for changing the stress level is presented. The optimization criterion is to minimize the approximate variance of the maximum likelihood estimator of a percentile of the products’ lifetime distribution.

Keywords: bivariate normal, Fisher information matrix, inverse Gaussian distribution, Wiener process

Procedia PDF Downloads 309
19231 The Effects of Different Parameters of Wood Floating Debris on Scour Rate Around Bridge Piers

Authors: Muhanad Al-Jubouri

Abstract:

A local scour is the most important of the several scours impacting bridge performance and security. Even though scour is widespread in bridges, especially during flood seasons, the experimental tests could not be applied to many standard highway bridges. A computational fluid dynamics numerical model was used to solve the problem of calculating local scouring and deposition for non-cohesive silt and clear water conditions near single and double cylindrical piers with the effect of floating debris. When FLOW-3D software is employed with the Rang turbulence model, the Nilsson bed-load transfer equation and fine mesh size are considered. The numerical findings of single cylindrical piers correspond pretty well with the physical model's results. Furthermore, after parameter effectiveness investigates the range of outcomes based on predicted user inputs such as the bed-load equation, mesh cell size, and turbulence model, the final numerical predictions are compared to experimental data. When the findings are compared, the error rate for the deepest point of the scour is equivalent to 3.8% for the single pier example.

Keywords: local scouring, non-cohesive, clear water, computational fluid dynamics, turbulence model, bed-load equation, debris

Procedia PDF Downloads 56
19230 Application of RS and GIS Technique for Identifying Groundwater Potential Zone in Gomukhi Nadhi Sub Basin, South India

Authors: Punitha Periyasamy, Mahalingam Sudalaimuthu, Sachikanta Nanda, Arasu Sundaram

Abstract:

India holds 17.5% of the world’s population but has only 2% of the total geographical area of the world where 27.35% of the area is categorized as wasteland due to lack of or less groundwater. So there is a demand for excessive groundwater for agricultural and non agricultural activities to balance its growth rate. With this in mind, an attempt is made to find the groundwater potential zone in Gomukhi river sub basin of Vellar River basin, TamilNadu, India covering an area of 1146.6 Sq.Km consists of 9 blocks from Peddanaickanpalayam to Villupuram fall in the sub basin. The thematic maps such as Geology, Geomorphology, Lineament, Landuse, and Landcover and Drainage are prepared for the study area using IRS P6 data. The collateral data includes rainfall, water level, soil map are collected for analysis and inference. The digital elevation model (DEM) is generated using Shuttle Radar Topographic Mission (SRTM) and the slope of the study area is obtained. ArcGIS 10.1 acts as a powerful spatial analysis tool to find out the ground water potential zones in the study area by means of weighted overlay analysis. Each individual parameter of the thematic maps are ranked and weighted in accordance with their influence to increase the water level in the ground. The potential zones in the study area are classified viz., Very Good, Good, Moderate, Poor with its aerial extent of 15.67, 381.06, 575.38, 174.49 Sq.Km respectively.

Keywords: ArcGIS, DEM, groundwater, recharge, weighted overlay

Procedia PDF Downloads 427
19229 Simulation of Optimum Sculling Angle for Adaptive Rowing

Authors: Pornthep Rachnavy

Abstract:

The purpose of this paper is twofold. First, we believe that there are a significant relationship between sculling angle and sculling style among adaptive rowing. Second, we introduce a methodology used for adaptive rowing, namely simulation, to identify effectiveness of adaptive rowing. For our study we simulate the arms only single scull of adaptive rowing. The method for rowing fastest under the 1000 meter was investigated by study sculling angle using the simulation modeling. A simulation model of a rowing system was developed using the Matlab software package base on equations of motion consist of many variation for moving the boat such as oars length, blade velocity and sculling style. The boat speed, power and energy consumption on the system were compute. This simulation modeling can predict the force acting on the boat. The optimum sculling angle was performing by computer simulation for compute the solution. Input to the model are sculling style of each rower and sculling angle. Outputs of the model are boat velocity at 1000 meter. The present study suggests that the optimum sculling angle exist depends on sculling styles. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the first style is -57.00 and 22.0 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the second style is -57.00 and 22.0 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the third style is -51.57 and 28.65 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the fourth style is -45.84 and 34.38 degree. A theoretical simulation for rowing has been developed and presented. The results suggest that it may be advantageous for the rowers to select the sculling angles proper to sculling styles. The optimum sculling angles of the rower depends on the sculling styles made by each rower. The investigated of this paper can be concludes in three directions: 1;. There is the optimum sculling angle in arms only single scull of adaptive rowing. 2. The optimum sculling angles depend on the sculling styles. 3. Computer simulation of rowing can identify opportunities for improving rowing performance by utilizing the kinematic description of rowing. The freedom to explore alternatives in speed, thrust and timing with the computer simulation will provide the coach with a tool for systematic assessments of rowing technique In addition, the ability to use the computer to examine the very complex movements during rowing will help both the rower and the coach to conceptualize the components of movements that may have been previously unclear or even undefined.

Keywords: simulation, sculling, adaptive, rowing

Procedia PDF Downloads 453
19228 Acetic Acid Assisted Phytoextraction of Chromium (Cr) by Energy Crop (Arundo donax L.) in Cr Contaminated Soils

Authors: Muhammad Iqbal, Hafiz Muhammad Tauqeer, Hamza Rafaqat, Muhammad Naveed, Muhammad Awais Irshad

Abstract:

Soil pollution with chromium (Cr) has become one of the most important concerns due to its toxicity for humans. To date, various remediation approaches have been employed for the remediation and management of Cr contaminated soils. Phytoextraction is an eco-friendly and emerging remediation approach which has gained attention due to several advantages over conventional remediation approach. The use of energy crops for phytoremediation is an emerging trend worldwide. These energy crops have high tolerance against various environmental stresses, the potential to grow in diverse ecosystems and high biomass production make them a suitable candidate for phytoremediation of contaminated soils. The removal efficiency of plants in phytoextraction depends upon several soil and plant factors including solubility, bioavailability and metal speciation in soils. A pot scale experiment was conducted to evaluate the phytoextraction potential of Arundo donax L. with the application of acetic acid (A.A) in Cr contaminated soils. Plants were grown in pots filled with 5 kg soils for 90 days. After 30 days plants acclimatization in pot conditions, plants were treated with various levels of Cr (2.5 mM, 5 mM, 7.5 mM, 10 mM) and A.A (Cr 2.5 mM + A.A 2.5 mM, Cr 5 mM + A.A 2.5 mM, Cr 7.5 mM + A.A 2.5 mM, Cr 10 mM + A.A 2.5 mM). The application of A.A significantly increased metal uptake and in roots and shoots of A. donax. This increase was observed at Cr 7.5 mM + A.A 2.5 mM but at high concentrations, visual symptoms of Cr toxicity were observed on leaves. Similarly, A.A applications also affect the activities of key enzymes including catalase (CAT), superoxidase dismutase (SOD), and ascorbate peroxidase (APX) in leaves of A. donax. Based on results it is concluded that the applications of A.A acid for phytoextraction is an alternative approach for the management of Cr affected soils and synthetic chelators should be replaced with organic acids.

Keywords: acetic acid, A. donax, chromium, energy crop, phytoextraction

Procedia PDF Downloads 368
19227 Biogas from Cover Crops and Field Residues: Effects on Soil, Water, Climate and Ecological Footprint

Authors: Manfred Szerencsits, Christine Weinberger, Maximilian Kuderna, Franz Feichtinger, Eva Erhart, Stephan Maier

Abstract:

Cover or catch crops have beneficial effects for soil, water, erosion, etc. If harvested, they also provide feedstock for biogas without competition for arable land in regions, where only one main crop can be produced per year. On average gross energy yields of approx. 1300 m³ methane (CH4) ha-1 can be expected from 4.5 tonnes (t) of cover crop dry matter (DM) in Austria. Considering the total energy invested from cultivation to compression for biofuel use a net energy yield of about 1000 m³ CH4 ha-1 is remaining. With the straw of grain maize or Corn Cob Mix (CCM) similar energy yields can be achieved. In comparison to catch crops remaining on the field as green manure or to complete fallow between main crops the effects on soil, water and climate can be improved if cover crops are harvested without soil compaction and digestate is returned to the field in an amount equivalent to cover crop removal. In this way, the risk of nitrate leaching can be reduced approx. by 25% in comparison to full fallow. The risk of nitrous oxide emissions may be reduced up to 50% by contrast with cover crops serving as green manure. The effects on humus content and erosion are similar or better than those of cover crops used as green manure when the same amount of biomass was produced. With higher biomass production the positive effects increase even if cover crops are harvested and the only digestate is brought back to the fields. The ecological footprint of arable farming can be reduced by approx. 50% considering the substitution of natural gas with CH4 produced from cover crops.

Keywords: biogas, cover crops, catch crops, land use competition, sustainable agriculture

Procedia PDF Downloads 533
19226 The Role of Group Size, Public Employees’ Wages and Control Corruption Institutions in a Game-Theoretical Model of Public Corruption

Authors: Pablo J. Valverde, Jaime E. Fernandez

Abstract:

This paper shows under which conditions public corruption can emerge. The theoretical model includes variables such as the public employee wage (w), a control corruption parameter (c), and the group size of interactions (GS) between clusters of public officers and contractors. The system behavior is analyzed using phase diagrams based on combinations of such parameters (c, w, GS). Numerical simulations are implemented in order to contrast analytic results based on Nash equilibria of the theoretical model. Major findings include the functional relationship between wages and network topology, which attempts to reduce the emergence of corrupt behavior.

Keywords: public corruption, game theory, complex systems, Nash equilibrium.

Procedia PDF Downloads 226
19225 Fluidised Bed Gasification of Multiple Agricultural Biomass-Derived Briquettes

Authors: Rukayya Ibrahim Muazu, Aiduan Li Borrion, Julia A. Stegemann

Abstract:

Biomass briquette gasification is regarded as a promising route for efficient briquette use in energy generation, fuels and other useful chemicals, however, previous research work has focused on briquette gasification in fixed bed gasifiers such as updraft and downdraft gasifiers. Fluidised bed gasifier has the potential to be effectively sized for medium or large scale. This study investigated the use of fuel briquettes produced from blends of rice husks and corn cobs biomass residues, in a bubbling fluidised bed gasifier. The study adopted a combination of numerical equations and Aspen Plus simulation software to predict the product gas (syngas) composition based on briquette's density and biomass composition (blend ratio of rice husks to corn cobs). The Aspen Plus model was based on an experimentally validated model from the literature. The results based on a briquette size of 32 mm diameter and relaxed density range of 500 to 650 kg/m3 indicated that fluidisation air required in the gasifier increased with an increase in briquette density, and the fluidisation air showed to be the controlling factor compared with the actual air required for gasification of the biomass briquettes. The mass flowrate of CO2 in the predicted syngas composition, increased with an increase in the air flow rate, while CO production decreased and H2 was almost constant. The H2/CO ratio for various blends of rice husks and corn cobs did not significantly change at the designed process air, but a significant difference of 1.0 for H2/CO ratio was observed at higher air flow rate, and between 10/90 to 90/10 blend ratio of rice husks to corn cobs. This implies the need for further understanding of biomass variability and hydrodynamic parameters on syngas composition in biomass briquette gasification.

Keywords: aspen plus, briquettes, fluidised bed, gasification, syngas

Procedia PDF Downloads 438
19224 Solving Transient Conduction and Radiation using Finite Volume Method

Authors: Ashok K. Satapathy, Prerana Nashine

Abstract:

Radiative heat transfer in participating medium was anticipated using the finite volume method. The radiative transfer equations are formulated for absorbing and anisotropically scattering and emitting medium. The solution strategy is discussed and the conditions for computational stability are conferred. The equations have been solved for transient radiative medium and transient radiation incorporated with transient conduction. Results have been obtained for irradiation and corresponding heat fluxes for both the cases. The solutions can be used to conclude incident energy and surface heat flux. Transient solutions were obtained for a slab of heat conducting in slab by thermal radiation. The effect of heat conduction during the transient phase is to partially equalize the internal temperature distribution. The solution procedure provides accurate temperature distributions in these regions. A finite volume procedure with variable space and time increments is used to solve the transient energy equation. The medium in the enclosure absorbs, emits, and anisotropically scatters radiative energy. The incident radiations and the radiative heat fluxes are presented in graphical forms. The phase function anisotropy plays a significant role in the radiation heat transfer when the boundary condition is non-symmetric.

Keywords: participating media, finite volume method, radiation coupled with conduction, heat transfer

Procedia PDF Downloads 366
19223 Evaluating the Suitability and Performance of Dynamic Modulus Predictive Models for North Dakota’s Asphalt Mixtures

Authors: Duncan Oteki, Andebut Yeneneh, Daba Gedafa, Nabil Suleiman

Abstract:

Most agencies lack the equipment required to measure the dynamic modulus (|E*|) of asphalt mixtures, necessitating the need to use predictive models. This study compared measured |E*| values for nine North Dakota asphalt mixes using the original Witczak, modified Witczak, and Hirsch models. The influence of temperature on the |E*| models was investigated, and Pavement ME simulations were conducted using measured |E*| and predictions from the most accurate |E*| model. The results revealed that the original Witczak model yielded the lowest Se/Sy and highest R² values, indicating the lowest bias and highest accuracy, while the poorest overall performance was exhibited by the Hirsch model. Using predicted |E*| as inputs in the Pavement ME generated conservative distress predictions compared to using measured |E*|. The original Witczak model was recommended for predicting |E*| for low-reliability pavements in North Dakota.

Keywords: asphalt mixture, binder, dynamic modulus, MEPDG, pavement ME, performance, prediction

Procedia PDF Downloads 31
19222 Efficiency of Secondary Schools by ICT Intervention in Sylhet Division of Bangladesh

Authors: Azizul Baten, Kamrul Hossain, Abdullah-Al-Zabir

Abstract:

The objective of this study is to develop an appropriate stochastic frontier secondary schools efficiency model by ICT Intervention and to examine the impact of ICT challenges on secondary schools efficiency in the Sylhet division in Bangladesh using stochastic frontier analysis. The Translog stochastic frontier model was found an appropriate than the Cobb-Douglas model in secondary schools efficiency by ICT Intervention. Based on the results of the Cobb-Douglas model, it is found that the coefficient of the number of teachers, the number of students, and teaching ability had a positive effect on increasing the level of efficiency. It indicated that these are related to technical efficiency. In the case of inefficiency effects for both Cobb-Douglas and Translog models, the coefficient of the ICT lab decreased secondary school inefficiency, but the online class in school was found to increase the level of inefficiency. The coefficients of teacher’s preference for ICT tools like multimedia projectors played a contributor role in decreasing the secondary school inefficiency in the Sylhet division of Bangladesh. The interaction effects of the number of teachers and the classrooms, and the number of students and the number of classrooms, the number of students and teaching ability, and the classrooms and teaching ability of the teachers were recorded with the positive values and these have a positive impact on increasing the secondary school efficiency. The overall mean efficiency of urban secondary schools was found at 84.66% for the Translog model, while it was 83.63% for the Cobb-Douglas model. The overall mean efficiency of rural secondary schools was found at 80.98% for the Translog model, while it was 81.24% for the Cobb-Douglas model. So, the urban secondary schools performed better than the rural secondary schools in the Sylhet division. It is observed from the results of the Tobit model that the teacher-student ratio had a positive influence on secondary school efficiency. The teaching experiences of those who have 1 to 5 years and 10 years above, MPO type school, conventional teaching method have had a negative and significant influence on secondary school efficiency. The estimated value of σ-square (0.0625) was different from Zero, indicating a good fit. The value of γ (0.9872) was recorded as positive and it can be interpreted as follows: 98.72 percent of random variation around in secondary school outcomes due to inefficiency.

Keywords: efficiency, secondary schools, ICT, stochastic frontier analysis

Procedia PDF Downloads 132
19221 Distangling Biological Noise in Cellular Images with a Focus on Explainability

Authors: Manik Sharma, Ganapathy Krishnamurthi

Abstract:

The cost of some drugs and medical treatments has risen in recent years, that many patients are having to go without. A classification project could make researchers more efficient. One of the more surprising reasons behind the cost is how long it takes to bring new treatments to market. Despite improvements in technology and science, research and development continues to lag. In fact, finding new treatment takes, on average, more than 10 years and costs hundreds of millions of dollars. If successful, we could dramatically improve the industry's ability to model cellular images according to their relevant biology. In turn, greatly decreasing the cost of treatments and ensure these treatments get to patients faster. This work aims at solving a part of this problem by creating a cellular image classification model which can decipher the genetic perturbations in cell (occurring naturally or artificially). Another interesting question addressed is what makes the deep-learning model decide in a particular fashion, which can further help in demystifying the mechanism of action of certain perturbations and paves a way towards the explainability of the deep-learning model.

Keywords: cellular images, genetic perturbations, deep-learning, explainability

Procedia PDF Downloads 96
19220 Cognitive Model of Analogy Based on Operation of the Brain Cells: Glial, Axons and Neurons

Authors: Ozgu Hafizoglu

Abstract:

Analogy is an essential tool of human cognition that enables connecting diffuse and diverse systems with attributional, deep structural, casual relations that are essential to learning, to innovation in artificial worlds, and to discovery in science. Cognitive Model of Analogy (CMA) leads and creates information pattern transfer within and between domains and disciplines in science. This paper demonstrates the Cognitive Model of Analogy (CMA) as an evolutionary approach to scientific research. The model puts forward the challenges of deep uncertainty about the future, emphasizing the need for flexibility of the system in order to enable reasoning methodology to adapt to changing conditions. In this paper, the model of analogical reasoning is created based on brain cells, their fractal, and operational forms within the system itself. Visualization techniques are used to show correspondences. Distinct phases of the problem-solving processes are divided thusly: encoding, mapping, inference, and response. The system is revealed relevant to brain activation considering each of these phases with an emphasis on achieving a better visualization of the brain cells: glial cells, axons, axon terminals, and neurons, relative to matching conditions of analogical reasoning and relational information. It’s found that encoding, mapping, inference, and response processes in four-term analogical reasoning are corresponding with the fractal and operational forms of brain cells: glial, axons, and neurons.

Keywords: analogy, analogical reasoning, cognitive model, brain and glials

Procedia PDF Downloads 170
19219 Efficient Model Selection in Linear and Non-Linear Quantile Regression by Cross-Validation

Authors: Yoonsuh Jung, Steven N. MacEachern

Abstract:

Check loss function is used to define quantile regression. In the prospect of cross validation, it is also employed as a validation function when underlying truth is unknown. However, our empirical study indicates that the validation with check loss often leads to choosing an over estimated fits. In this work, we suggest a modified or L2-adjusted check loss which rounds the sharp corner in the middle of check loss. It has a large effect of guarding against over fitted model in some extent. Through various simulation settings of linear and non-linear regressions, the improvement of check loss by L2 adjustment is empirically examined. This adjustment is devised to shrink to zero as sample size grows.

Keywords: cross-validation, model selection, quantile regression, tuning parameter selection

Procedia PDF Downloads 421
19218 Thermodynamics of Random Copolymers in Solution

Authors: Maria Bercea, Bernhard A. Wolf

Abstract:

The thermodynamic behavior for solutions of poly (methyl methacrylate-ran-t-butyl methacrylate) of variable composition as compared with the corresponding homopolymers was investigated by light scattering measurements carried out for dilute solutions and vapor pressure measurements of concentrated solutions. The complex dependencies of the Flory Huggins interaction parameter on concentration and copolymer composition in solvents of different polarity (toluene and chloroform) can be understood by taking into account the ability of the polymers to rearrange in a response to changes in their molecular surrounding. A recent unified thermodynamic approach was used for modeling the experimental data, being able to describe the behavior of the different solutions by means of two adjustable parameters, one representing the effective number of solvent segments and another one accounting for the interactions between the components. Thus, it was investigated how the solvent quality changes with the composition of the copolymers through the Gibbs energy of mixing as a function of polymer concentration. The largest reduction of the Gibbs energy at a given composition of the system was observed for the best solvent. The present investigation proves that the new unified thermodynamic approach is a general concept applicable to homo- and copolymers, independent of the chain conformation or shape, molecular and chemical architecture of the components and of other dissimilarities, such as electrical charges.

Keywords: random copolymers, Flory Huggins interaction parameter, Gibbs energy of mixing, chemical architecture

Procedia PDF Downloads 269
19217 Stabilization of the Bernoulli-Euler Plate Equation: Numerical Analysis

Authors: Carla E. O. de Moraes, Gladson O. Antunes, Mauro A. Rincon

Abstract:

The aim of this paper is to study the internal stabilization of the Bernoulli-Euler equation numerically. For this, we consider a square plate subjected to a feedback/damping force distributed only in a subdomain. An algorithm for obtaining an approximate solution to this problem was proposed and implemented. The numerical method used was the Finite Difference Method. Numerical simulations were performed and showed the behavior of the solution, confirming the theoretical results that have already been proved in the literature. In addition, we studied the validation of the numerical scheme proposed, followed by an analysis of the numerical error; and we conducted a study on the decay of the energy associated.

Keywords: Bernoulli-Euler plate equation, numerical simulations, stability, energy decay, finite difference method

Procedia PDF Downloads 400
19216 Uncertainty in Risk Modeling

Authors: Mueller Jann, Hoffmann Christian Hugo

Abstract:

Conventional quantitative risk management in banking is a risk factor of its own, because it rests on assumptions such as independence and availability of data which do not hold when rare events of extreme consequences are involved. There is a growing recognition of the need for alternative risk measures that do not make these assumptions. We propose a novel method for modeling the risk associated with investment products, in particular derivatives, by using a formal language for specifying financial contracts. Expressions in this language are interpreted in the category of values annotated with (a formal representation of) uncertainty. The choice of uncertainty formalism thus becomes a parameter of the model, so it can be adapted to the particular application and it is not constrained to classical probabilities. We demonstrate our approach using a simple logic-based uncertainty model and a case study in which we assess the risk of counter party default in a portfolio of collateralized loans.

Keywords: risk model, uncertainty monad, derivatives, contract algebra

Procedia PDF Downloads 561
19215 Site Selection in Adaptive Reuse Architecture for Social Housing in Johannesburg, South Africa

Authors: Setapo Moloi, Jun-Ichiro Giorgos Tsutsumi

Abstract:

South Africa’s need for the provision of housing within its major city centres, specifically Gauteng Province (GP), is a major concern. Initiatives for converting misused/ unused buildings to suitable housing for residents who work in the city as well as prospective citizens are currently underway, one aspect that is needed currently, is the re-possession of these buildings repurposing, into housing communities for quality low cost mixed density housing and for this process to have minimal strain on existing infrastructure like energy, emission reduction etc. Unfortunately, there are instances in Johannesburg, the country’s economic capital, with 2017 estimates claiming that 700 buildings lay unused or misused due to issues that will be discussed in this paper, these then become hubs for illegal activity and are an unacceptable form of shelter. It can be argued that the provision of inner-city social housing is lacking, but not due to the unavailability of funding or usable land and buildings, but that these assets are not being used appropriately nor to their full potential. Currently the GP government has mandated the re-purposing of all buildings that meet their criteria (structural stability, feasibility, adaptability, etc.) with the intention of inviting interested parties to propose conversions of the buildings into densified social housing. Going forward, the proposed focus is creation of social housing communities within existing buildings which may be retrofitted with sustainable technologies, green design strategies and principles, aiming for the finished buildings to achieve ‘Net-Zero/Positive’ status. A Net-Zero building, according to The Green Building Council of South Africa (GBCSA) is a building which manages to produce resources it needs to function, and reduces wastage, emissions and demand of these resources during its lifespan. The categories which GBCSA includes are carbon, water, waste and ecology, this may include material selection, construction methods, etc.

Keywords: adaptive reuse, conversion, net-zero, social housing, sustainable communities

Procedia PDF Downloads 117
19214 Analysis of Co2 Emission from Thailand's Thermal Power Sector by Divisia Decomposition Approach

Authors: Isara Muangthai, Lin Sue Jane

Abstract:

Electricity is vital to every country’s economy in the world. For Thailand, the electricity generation sector plays an important role in the economic system, and it is the largest source of CO2 emissions. The aim of this paper is to use the decomposition analysis to investigate the key factors contributing to the changes of CO2 emissions from the electricity sector. The decomposition analysis has been widely used to identify and assess the contributors to the changes in emission trends. Our study adopted the Divisia index decomposition to identify the key factors affecting the evolution of CO2 emissions from Thailand’s thermal power sector during 2000-2011. The change of CO2 emissions were decomposed into five factors, including: Emission coefficient, heat rate, fuel intensity, electricity intensity, and economic growth. Results have shown that CO2 emission in Thailand’s thermal power sector increased 29,173 thousand tons during 2000-2011. Economic growth was found to be the primary factor for increasing CO2 emissions, while the electricity intensity played a dominant role in decreasing CO2 emissions. The increasing effect of economic growth was up to 55,924 million tons of CO2 emissions because the growth and development of the economy relied on a large electricity supply. On the other hand, the shifting of fuel structure towards a lower-carbon content resulted in CO2 emission decline. Since the CO2 emissions released from Thailand’s electricity generation are rapidly increasing, the Thailand government will be required to implement a CO2 reduction plan in the future. In order to cope with the impact of CO2 emissions related to the power sector and to achieve sustainable development, this study suggests that Thailand’s government should focus on restructuring the fuel supply in power generation towards low carbon fuels by promoting the use of renewable energy for electricity, improving the efficiency of electricity use by reducing electricity transmission and the distribution of line losses, implementing energy conservation strategies by enhancing the purchase of energy-saving products, substituting the new power plant technology in the old power plants, promoting a shift of economic structure towards less energy-intensive services and orienting Thailand’s power industry towards low carbon electricity generation.

Keywords: co2 emission, decomposition analysis, electricity generation, energy consumption

Procedia PDF Downloads 465