Search results for: reduce order aeroelastic model (ROAM)
28452 The Influence of Music Education and the Order of Sounds on the Grouping of Sounds into Sequences of Six Tones
Authors: Adam Rosiński
Abstract:
This paper discusses an experiment conducted with two groups of participants, composed of musicians and non-musicians, in order to investigate the impact of the speed of a sound sequence and the order of sounds on the grouping of sounds into sequences of six tones. Significant differences were observed between musicians and non-musicians with respect to the threshold sequence speed at which the sequence was split into two streams. The differences in the results for the two groups suggest that the musical education of the participating listeners may be a vital factor. The criterion of musical education should be taken into account during experiments so that the results obtained are reliable, uniform, and free from interpretive errors.Keywords: auditory scene analysis, education, hearing, psychoacoustics
Procedia PDF Downloads 10328451 Evaluated Nuclear Data Based Photon Induced Nuclear Reaction Model of GEANT4
Authors: Jae Won Shin
Abstract:
We develop an evaluated nuclear data based photonuclear reaction model of GEANT4 for a more accurate simulation of photon-induced neutron production. The evaluated photonuclear data libraries from the ENDF/B-VII.1 are taken as input. Incident photon energies up to 140 MeV which is the threshold energy for the pion production are considered. For checking the validity of the use of the data-based model, we calculate the photoneutron production cross-sections and yields and compared them with experimental data. The results obtained from the developed model are found to be in good agreement with the experimental data for (γ,xn) reactions.Keywords: ENDF/B-VII.1, GEANT4, photoneutron, photonuclear reaction
Procedia PDF Downloads 27528450 Architectural Building Safety and Health Performance Model for Stratified Low-Cost Housing: Education and Management Tool for Building Managers
Authors: Zainal Abidin Akasah, Maizam Alias, Azuin Ramli
Abstract:
The safety and health performances aspects of a building are the most challenging aspect of facility management. It requires a deep understanding by the building managers on the factors that contribute to health and safety performances. This study attempted to develop an explanatory architectural safety performance model for stratified low-cost housing in Malaysia. The proposed Building Safety and Health Performance (BSHP) model was tested empirically through a survey on 308 construction practitioners using Partial Least Squares (PLS) and Structural Equation Modelling (SEM) tool. Statistical analysis results supports the conclusion that architecture, building services, external environment, management approaches and maintenance management have positive influence on safety and health performance of stratified low-cost housing in Malaysia. The findings provide valuable insights for construction industry to introduce BSHP model in the future where the model could be used as a guideline for training purposes of managers and better planning and implementation of building management.Keywords: building management, stratified low-cost housing, safety, health model
Procedia PDF Downloads 55928449 Optimization Aluminium Design for the Facade Second Skin toward Visual Comfort: Case Studies & Dialux Daylighting Simulation Model
Authors: Yaseri Dahlia Apritasari
Abstract:
Visual comfort is important for the building occupants to need. Visual comfort can be fulfilled through natural lighting (daylighting) and artificial lighting. One strategy to optimize natural lighting can be achieved through the facade second skin design. This strategy can reduce glare, and fulfill visual comfort need. However, the design strategy cannot achieve light intensity for visual comfort. Because the materials, design and opening percentage of the facade of second skin blocked sunlight. This paper discusses aluminum material for the facade second skin design that can fulfill the optimal visual comfort with the case studies Multi Media Tower building. The methodology of the research is combination quantitative and qualitative through field study observed, lighting measurement and visual comfort questionnaire. Then it used too simulation modeling (DIALUX 4.13, 2016) for three facades second skin design model. Through following steps; (1) Measuring visual comfort factor: light intensity indoor and outdoor; (2) Taking visual comfort data from building occupants; (3) Making models with different facade second skin design; (3) Simulating and analyzing the light intensity value for each models that meet occupants visual comfort standard: 350 lux (Indonesia National Standard, 2010). The result shows that optimization of aluminum material for the facade second skin design can meet optimal visual comfort for building occupants. The result can give recommendation aluminum opening percentage of the facade second skin can meet optimal visual comfort for building occupants.Keywords: aluminium material, Facade, second skin, visual comfort
Procedia PDF Downloads 35328448 Knowledge Sharing within a Team: Exploring the Antecedents and Role of Trust
Authors: Li Yan Hei, Au Wing Tung
Abstract:
Knowledge sharing is a process in which individuals mutually exchange existing knowledge and co-create new knowledge. Previous research has confirmed that trust is positively associated with knowledge sharing. However, only few studies systematically examined the antecedents of trust and these antecedents’ impacts on knowledge sharing. In order to explore and understand the relationships between trust and knowledge sharing in depth, this study proposed a relationship maintenance-based model to examine the antecedents of trust in knowledge sharing in project teams. Three critical elements within a project team were measured, including the environment, project team partner and interaction. It was hypothesized that the trust would lead to knowledge sharing and in turn result in perceived good team performance. With a sample of 200 Hong Kong employees, the proposed model was evaluated with structural equation modeling. Expected findings are trust will contribute to knowledge sharing, resulting in better team performance. The results will also offer insights into antecedents of trust that play a heavy role in the focal relationship. The present study contributes to a more holistic understanding of relationship between trust and knowledge sharing by linking the antecedents and outcomes. The findings will raise the awareness of project managers on ways to promote knowledge sharing.Keywords: knowledge sharing, project management, team, trust
Procedia PDF Downloads 61828447 Development of an Elastic Functionally Graded Interphase Model for the Micromechanics Response of Composites
Authors: Trevor Sabiston, Mohsen Mohammadi, Mohammed Cherkaoui, Kaan Inal
Abstract:
A new micromechanics framework is developed for long fibre reinforced composites using a single fibre surrounded by a functionally graded interphase and matrix as a representative unit cell. The unit cell is formulated to represent any number of aligned fibres by a single fibre. Using this model the elastic response of long fibre composites is predicted in all directions. The model is calibrated to experimental results and shows very good agreement in the elastic regime. The differences between the proposed model and existing models are discussed.Keywords: computational mechanics, functionally graded interphase, long fibre composites, micromechanics
Procedia PDF Downloads 32028446 Maintenance Alternatives Related to Costs of Wind Turbines Using Finite State Markov Model
Authors: Boukelkoul Lahcen
Abstract:
The cumulative costs for O&M may represent as much as 65%-90% of the turbine's investment cost. Nowadays the cost effectiveness concept becomes a decision-making and technology evaluation metric. The cost of energy metric accounts for the effect replacement cost and unscheduled maintenance cost parameters. One key of the proposed approach is the idea of maintaining the WTs which can be captured via use of a finite state Markov chain. Such a model can be embedded within a probabilistic operation and maintenance simulation reflecting the action to be done. In this paper, an approach of estimating the cost of O&M is presented. The finite state Markov model is used for decision problems with number of determined periods (life cycle) to predict the cost according to various options of maintenance.Keywords: cost, finite state, Markov model, operation and maintenance
Procedia PDF Downloads 53428445 Seat Assignment Model for Student Admissions Process at Saudi Higher Education Institutions
Authors: Mohammed Salem Alzahrani
Abstract:
In this paper, student admission process is studied to optimize the assignment of vacant seats with three main objectives. Utilizing all vacant seats, satisfying all program of study admission requirements and maintaining fairness among all candidates are the three main objectives of the optimization model. Seat Assignment Method (SAM) is used to build the model and solve the optimization problem with help of Northwest Coroner Method and Least Cost Method. A closed formula is derived for applying the priority of assigning seat to candidate based on SAM.Keywords: admission process model, assignment problem, Hungarian Method, Least Cost Method, Northwest Corner Method, SAM
Procedia PDF Downloads 50028444 On the Mathematical Modelling of Aggregative Stability of Disperse Systems
Authors: Arnold M. Brener, Lesbek Tashimov, Ablakim S. Muratov
Abstract:
The paper deals with the special model for coagulation kernels which represents new control parameters in the Smoluchowski equation for binary aggregation. On the base of the model the new approach to evaluating aggregative stability of disperse systems has been submitted. With the help of this approach the simple estimates for aggregative stability of various types of hydrophilic nano-suspensions have been obtained.Keywords: aggregative stability, coagulation kernels, disperse systems, mathematical model
Procedia PDF Downloads 31128443 Strategies for the Optimization of Ground Resistance in Large Scale Foundations for Optimum Lightning Protection
Authors: Oibar Martinez, Clara Oliver, Jose Miguel Miranda
Abstract:
In this paper, we discuss the standard improvements which can be made to reduce the earth resistance in difficult terrains for optimum lightning protection, what are the practical limitations, and how the modeling can be refined for accurate diagnostics and ground resistance minimization. Ground resistance minimization can be made via three different approaches: burying vertical electrodes connected in parallel, burying horizontal conductive plates or meshes, or modifying the own terrain, either by changing the entire terrain material in a large volume or by adding earth-enhancing compounds. The use of vertical electrodes connected in parallel pose several practical limitations. In order to prevent loss of effectiveness, it is necessary to keep a minimum distance between each electrode, which is typically around five times larger than the electrode length. Otherwise, the overlapping of the local equipotential lines around each electrode reduces the efficiency of the configuration. The addition of parallel electrodes reduces the resistance and facilitates the measurement, but the basic parallel resistor formula of circuit theory will always underestimate the final resistance. Numerical simulation of equipotential lines around the electrodes overcomes this limitation. The resistance of a single electrode will always be proportional to the soil resistivity. The electrodes are usually installed with a backfilling material of high conductivity, which increases the effective diameter. However, the improvement is marginal, since the electrode diameter counts in the estimation of the ground resistance via a logarithmic function. Substances that are used for efficient chemical treatment must be environmentally friendly and must feature stability, high hygroscopicity, low corrosivity, and high electrical conductivity. A number of earth enhancement materials are commercially available. Many are comprised of carbon-based materials or clays like bentonite. These materials can also be used as backfilling materials to reduce the resistance of an electrode. Chemical treatment of soil has environmental issues. Some products contain copper sulfate or other copper-based compounds, which may not be environmentally friendly. Carbon-based compounds are relatively inexpensive and they do have very low resistivities, but they also feature corrosion issues. Typically, the carbon can corrode and destroy a copper electrode in around five years. These compounds also have potential environmental concerns. Some earthing enhancement materials contain cement, which, after installation acquire properties that are very close to concrete. This prevents the earthing enhancement material from leaching into the soil. After analyzing different configurations, we conclude that a buried conductive ring with vertical electrodes connected periodically should be the optimum baseline solution for the grounding of a large size structure installed on a large resistivity terrain. In order to show this, a practical example is explained here where we simulate the ground resistance of a conductive ring buried in a terrain with a resistivity in the range of 1 kOhm·m.Keywords: grounding improvements, large scale scientific instrument, lightning risk assessment, lightning standards
Procedia PDF Downloads 14028442 Analyzing Business Model Choices and Sustainable Value Capturing: A Multiple Case Study of Sharing Economy Business Models
Authors: Minttu Laukkanen, Janne Huiskonen
Abstract:
This study investigates the sharing economy business models as examples of the sustainable business models. The aim is to contribute to the limited literature on sharing economy in connection with sustainable business models by explaining sharing economy business models value capturing. Specifically, this research answers the following question: How business model choices affect captured sustainable value? A multiple case study approach is applied in this study. Twenty different successful sharing economy business models focusing on consumer business and covering four main areas, accommodation, mobility, food, and consumer goods, are selected for analysis. The secondary data available on companies’ websites, previous research, reports, and other public documents are used. All twenty cases are analyzed through the sharing economy business model framework and sustainable value analysis framework using qualitative data analysis. This study represents general sharing economy business model value attributes and their specifications, i.e. sustainable value propositions for different stakeholders, and further explains the sustainability impacts of different sharing economy business models through captured and uncaptured value. In conclusion, this study represents how business model choices affect sustainable value capturing through eight business model attributes identified in this study. This paper contributes to the research on sustainable business models and sharing economy by examining how business model choices affect captured sustainable value. This study highlights the importance of careful business model and sustainability impacts analyses including the triple bottom line, multiple stakeholders and value captured and uncaptured perspectives as well as sustainability trade-offs. It is not self-evident that sharing economy business models advance sustainability, and business model choices does matter.Keywords: sharing economy, sustainable business model innovation, sustainable value, value capturing
Procedia PDF Downloads 17528441 Rheological Characteristics of Ice Slurries Based on Propylene- and Ethylene-Glycol at High Ice Fractions
Authors: Senda Trabelsi, Sébastien Poncet, Michel Poirier
Abstract:
Ice slurries are considered as a promising phase-changing secondary fluids for air-conditioning, packaging or cooling industrial processes. An experimental study has been here carried out to measure the rheological characteristics of ice slurries. Ice slurries consist in a solid phase (flake ice crystals) and a liquid phase. The later is composed of a mixture of liquid water and an additive being here either (1) Propylene-Glycol (PG) or (2) Ethylene-Glycol (EG) used to lower the freezing point of water. Concentrations of 5%, 14% and 24% of both additives are investigated with ice mass fractions ranging from 5% to 85%. The rheological measurements are carried out using a Discovery HR-2 vane-concentric cylinder with four full-length blades. The experimental results show that the behavior of ice slurries is generally non-Newtonian with shear-thinning or shear-thickening behaviors depending on the experimental conditions. In order to determine the consistency and the flow index, the Herschel-Bulkley model is used to describe the behavior of ice slurries. The present results are finally validated against an experimental database found in the literature and the predictions of an Artificial Neural Network model.Keywords: ice slurry, propylene-glycol, ethylene-glycol, rheology
Procedia PDF Downloads 26428440 Creep Behaviour of Heterogeneous Timber-UHPFRC Beams Assembled by Bonding: Experimental and Analytical Investigation
Authors: K. Kong, E. Ferrier, L. Michel
Abstract:
The purpose of this research was to investigate the creep behaviour of the heterogeneous Timber-UHPFRC beams. New developments have been done to further improve the structural performance, such as strengthening of the timber (glulam) beam by bonding composite material combine with an ultra-high performance fibre reinforced concrete (UHPFRC) internally reinforced with or without carbon fibre reinforced polymer (CFRP) bars. However, in the design of wooden structures, in addition to the criteria of strengthening and stiffness, deformability due to the creep of wood, especially in horizontal elements, is also a design criterion. Glulam, UHPFRC and CFRP may be an interesting composite mix to respond to the issue of creep behaviour of composite structures made of different materials with different rheological properties. In this paper, we describe an experimental and analytical investigation of the creep performance of the glulam-UHPFRC-CFRP beams assembled by bonding. The experimental investigations creep behaviour was conducted for different environments: in- and outside under constant loading for approximately a year. The measured results are compared with numerical ones obtained by an analytical model. This model was developed to predict the creep response of the glulam-UHPFRC-CFRP beams based on the creep characteristics of the individual components. The results show that heterogeneous glulam-UHPFRC beams provide an improvement in both the strengthening and stiffness, and can also effectively reduce the creep deflection of wooden beams.Keywords: carbon fibre-reinforced polymer (CFRP) bars, creep behaviour, glulam, ultra-high performance fibre reinforced concrete (UHPFRC)
Procedia PDF Downloads 40628439 Causal Estimation for the Left-Truncation Adjusted Time-Varying Covariates under the Semiparametric Transformation Models of a Survival Time
Authors: Yemane Hailu Fissuh, Zhongzhan Zhang
Abstract:
In biomedical researches and randomized clinical trials, the most commonly interested outcomes are time-to-event so-called survival data. The importance of robust models in this context is to compare the effect of randomly controlled experimental groups that have a sense of causality. Causal estimation is the scientific concept of comparing the pragmatic effect of treatments conditional to the given covariates rather than assessing the simple association of response and predictors. Hence, the causal effect based semiparametric transformation model was proposed to estimate the effect of treatment with the presence of possibly time-varying covariates. Due to its high flexibility and robustness, the semiparametric transformation model which shall be applied in this paper has been given much more attention for estimation of a causal effect in modeling left-truncated and right censored survival data. Despite its wide applications and popularity in estimating unknown parameters, the maximum likelihood estimation technique is quite complex and burdensome in estimating unknown parameters and unspecified transformation function in the presence of possibly time-varying covariates. Thus, to ease the complexity we proposed the modified estimating equations. After intuitive estimation procedures, the consistency and asymptotic properties of the estimators were derived and the characteristics of the estimators in the finite sample performance of the proposed model were illustrated via simulation studies and Stanford heart transplant real data example. To sum up the study, the bias of covariates was adjusted via estimating the density function for truncation variable which was also incorporated in the model as a covariate in order to relax the independence assumption of failure time and truncation time. Moreover, the expectation-maximization (EM) algorithm was described for the estimation of iterative unknown parameters and unspecified transformation function. In addition, the causal effect was derived by the ratio of the cumulative hazard function of active and passive experiments after adjusting for bias raised in the model due to the truncation variable.Keywords: causal estimation, EM algorithm, semiparametric transformation models, time-to-event outcomes, time-varying covariate
Procedia PDF Downloads 12828438 Designing Elevations by Photocatalysis of Precast Concrete Materials, in Reducing Energy Consumption of Buildings: Case Study of Tabriz
Authors: Mahsa Faramarzi Asli, Mina Sarabi
Abstract:
The important issues that are addressed in most advanced industrial countries in recent decades, discussion of minimizing heat losses through the buildings. And the most influential parameters in the calculation of building energy consumption, is heat exchange, which takes place between the interior and outer space. One of the solutions to reduce heat loss is using materials with low thermal conductivity. The purpose of this article, is the effect of using some frontages with nano-concrete photo catalytic precast materials for reducing energy consumption in buildings. For this purpose, estimating the energy dissipation through the facade built with nano-concrete photo catalytic precast materials on a sample building in Tabriz city by BCS 19 software ( topic 19 simulation) is done and the results demonstrate reduce heat loss through the facade nano- concrete.Keywords: nano materials, optimize energy consumption, themal, stability
Procedia PDF Downloads 56528437 Modelling of Aerosols in Absorption Column
Authors: Hammad Majeed, Hanna Knuutila, Magne Hillestad, Hallvard F. Svendsen
Abstract:
Formation of aerosols can cause serious complications in industrial exhaust gas cleaning processes. Small mist droplets and fog formed can normally not be removed in conventional demisting equipment because their submicron size allows the particles or droplets to follow the gas flow. As a consequence of this, aerosol based emissions in the order of grams per Nm3 have been identified from PCCC plants. The model predicts the droplet size, the droplet internal variable profiles, and the mass transfer fluxes as function of position in the absorber. The Matlab model is based on a subclass method of weighted residuals for boundary value problems named, orthogonal collocation method. This paper presents results describing the basic simulation tool for the characterization of aerosols formed in CO2 absorption columns and describes how various entering droplets grow or shrink through an absorber and how their composition changes with respect to time. Below are given some preliminary simulation results for an aerosol droplet composition and temperature profiles.Keywords: absorption columns, aerosol formation, amine emissions, internal droplet profiles, monoethanolamine (MEA), post combustion CO2 capture, simulation
Procedia PDF Downloads 24528436 The Utilization of Recycled Construction and Demolition Waste Aggregate in Asphaltic Concrete
Authors: Inas Kamel, Noor Z. Habib
Abstract:
Utilizing construction and demolition wastes in hotmix asphalt (HMA) pavement construction can reduce the adverse environmental effect of its inadequate disposal and reduce the pressure of extracting and processing mineral aggregates (MA). This study aims to examine the viability of replacing MA by recycled construction and demolition waste aggregates (RCDWA) in the wearing course of asphaltic concrete (AC) pavements without compromising its loadbearing capacity. The Marshall Method was used to evaluate the performance of AC wearing course specimens by replacing MA by 10%, 20% and 30% RCDWA. Grade 60/70 bitumen was used in the range 3.0-5.5%, with 05% increments, to generate the optimum bitumen content (OBC). From the volumetric analysis and test property curves, the mixture containing 20% RCDWA was chosen as the preferred mix at 5.1% OBC. It possessed a 10% increase in Marshall Stability compared to the reference specimen, containing 100% MA, and a 6% increase in Marshall flow.Keywords: aggregate, asphaltic concrete, Marshall method, optimum bitumen content, recycled construction and demolition waste
Procedia PDF Downloads 15728435 Phosphate Use Efficiency in Plants: A GWAS Approach to Identify the Pathways Involved
Authors: Azizah M. Nahari, Peter Doerner
Abstract:
Phosphate (Pi) is one of the essential macronutrients in plant growth and development, and it plays a central role in metabolic processes in plants, particularly photosynthesis and respiration. Limitation of crop productivity by Pi is widespread and is likely to increase in the future. Applications of Pi fertilizers have improved soil Pi fertility and crop production; however, they have also caused environmental damage. Therefore, in order to reduce dependence on unsustainable Pi fertilizers, a better understanding of phosphate use efficiency (PUE) is required for engineering nutrient-efficient crop plants. Enhanced Pi efficiency can be achieved by improved productivity per unit Pi taken up. We aim to identify, by using association mapping, general features of the most important loci that contribute to increased PUE to allow us to delineate the physiological pathways involved in defining this trait in the model plant Arabidopsis. As PUE is in part determined by the efficiency of uptake, we designed a hydroponic system to avoid confounding effects due to differences in root system architecture leading to differences in Pi uptake. In this system, 18 parental lines and 217 lines of the MAGIC population (a Multiparent Advanced Generation Inter-Cross) grown in high and low Pi availability conditions. The results showed revealed a large variation of PUE in the parental lines, indicating that the MAGIC population was well suited to identify PUE loci and pathways. 2 of 18 parental lines had the highest PUE in low Pi while some lines responded strongly and increased PUE with increased Pi. Having examined the 217 MAGIC population, considerable variance in PUE was found. A general feature was the trend of most lines to exhibit higher PUE when grown in low Pi conditions. Association mapping is currently in progress, but initial observations indicate that a wide variety of physiological processes are involved in influencing PUE in Arabidopsis. The combination of hydroponic growth methods and genome-wide association mapping is a powerful tool to identify the physiological pathways underpinning complex quantitative traits in plants.Keywords: hydroponic system growth, phosphate use efficiency (PUE), Genome-wide association mapping, MAGIC population
Procedia PDF Downloads 32228434 Combining a Continuum of Hidden Regimes and a Heteroskedastic Three-Factor Model in Option Pricing
Authors: Rachid Belhachemi, Pierre Rostan, Alexandra Rostan
Abstract:
This paper develops a discrete-time option pricing model for index options. The model consists of two key ingredients. First, daily stock return innovations are driven by a continuous hidden threshold mixed skew-normal (HTSN) distribution which generates conditional non-normality that is needed to fit daily index return. The most important feature of the HTSN is the inclusion of a latent state variable with a continuum of states, unlike the traditional mixture distributions where the state variable is discrete with little number of states. The HTSN distribution belongs to the class of univariate probability distributions where parameters of the distribution capture the dependence between the variable of interest and the continuous latent state variable (the regime). The distribution has an interpretation in terms of a mixture distribution with time-varying mixing probabilities. It has been shown empirically that this distribution outperforms its main competitor, the mixed normal (MN) distribution, in terms of capturing the stylized facts known for stock returns, namely, volatility clustering, leverage effect, skewness, kurtosis and regime dependence. Second, heteroscedasticity in the model is captured by a threeexogenous-factor GARCH model (GARCHX), where the factors are taken from the principal components analysis of various world indices and presents an application to option pricing. The factors of the GARCHX model are extracted from a matrix of world indices applying principal component analysis (PCA). The empirically determined factors are uncorrelated and represent truly different common components driving the returns. Both factors and the eight parameters inherent to the HTSN distribution aim at capturing the impact of the state of the economy on price levels since distribution parameters have economic interpretations in terms of conditional volatilities and correlations of the returns with the hidden continuous state. The PCA identifies statistically independent factors affecting the random evolution of a given pool of assets -in our paper a pool of international stock indices- and sorting them by order of relative importance. The PCA computes a historical cross asset covariance matrix and identifies principal components representing independent factors. In our paper, factors are used to calibrate the HTSN-GARCHX model and are ultimately responsible for the nature of the distribution of random variables being generated. We benchmark our model to the MN-GARCHX model following the same PCA methodology and the standard Black-Scholes model. We show that our model outperforms the benchmark in terms of RMSE in dollar losses for put and call options, which in turn outperforms the analytical Black-Scholes by capturing the stylized facts known for index returns, namely, volatility clustering, leverage effect, skewness, kurtosis and regime dependence.Keywords: continuous hidden threshold, factor models, GARCHX models, option pricing, risk-premium
Procedia PDF Downloads 29828433 Combination of Modelling and Environmental Life Cycle Assessment Approach for Demand Driven Biogas Production
Authors: Juan A. Arzate, Funda C. Ertem, M. Nicolas Cruz-Bournazou, Peter Neubauer, Stefan Junne
Abstract:
— One of the biggest challenges the world faces today is global warming that is caused by greenhouse gases (GHGs) coming from the combustion of fossil fuels for energy generation. In order to mitigate climate change, the European Union has committed to reducing GHG emissions to 80–95% below the level of the 1990s by the year 2050. Renewable technologies are vital to diminish energy-related GHG emissions. Since water and biomass are limited resources, the largest contributions to renewable energy (RE) systems will have to come from wind and solar power. Nevertheless, high proportions of fluctuating RE will present a number of challenges, especially regarding the need to balance the variable energy demand with the weather dependent fluctuation of energy supply. Therefore, biogas plants in this content would play an important role, since they are easily adaptable. Feedstock availability varies locally or seasonally; however there is a lack of knowledge in how biogas plants should be operated in a stable manner by local feedstock. This problem may be prevented through suitable control strategies. Such strategies require the development of convenient mathematical models, which fairly describe the main processes. Modelling allows us to predict the system behavior of biogas plants when different feedstocks are used with different loading rates. Life cycle assessment (LCA) is a technique for analyzing several sides from evolution of a product till its disposal in an environmental point of view. It is highly recommend to use as a decision making tool. In order to achieve suitable strategies, the combination of a flexible energy generation provided by biogas plants, a secure production process and the maximization of the environmental benefits can be obtained by the combination of process modelling and LCA approaches. For this reason, this study focuses on the biogas plant which flexibly generates required energy from the co-digestion of maize, grass and cattle manure, while emitting the lowest amount of GHG´s. To achieve this goal AMOCO model was combined with LCA. The program was structured in Matlab to simulate any biogas process based on the AMOCO model and combined with the equations necessary to obtain climate change, acidification and eutrophication potentials of the whole production system based on ReCiPe midpoint v.1.06 methodology. Developed simulation was optimized based on real data from operating biogas plants and existing literature research. The results prove that AMOCO model can successfully imitate the system behavior of biogas plants and the necessary time required for the process to adapt in order to generate demanded energy from available feedstock. Combination with LCA approach provided opportunity to keep the resulting emissions from operation at the lowest possible level. This would allow for a prediction of the process, when the feedstock utilization supports the establishment of closed material circles within a smart bio-production grid – under the constraint of minimal drawbacks for the environment and maximal sustainability.Keywords: AMOCO model, GHG emissions, life cycle assessment, modelling
Procedia PDF Downloads 19028432 Value Generation of Construction and Demolition Waste Originated in the Building Rehabilitation to Improve Energy Efficiency; From Waste to Resources
Authors: Mercedes Del Rio Merino, Jaime Santacruz Astorqui, Paola Villoria Saez, Carmen Viñas Arrebola
Abstract:
The lack of treatment of the waste from construction and demolition waste (CDW) is a problem that must be solved immediately. It is estimated that in the world not to use CDW generates an increase in the use of new materials close to 20% of the total value of the materials used. The problem is even greater in case these wastes are considered hazardous because the final deposition of them may also generate significant contamination. Therefore, the possibility of including CDW in the manufacturing of building materials, represents an interesting alternative to ensure their use and to reduce their possible risk. In this context and in the last years, many researches are being carried out in order to analyze the viability of using CDW as a substitute for the traditional raw material of high environmental impact. Even though it is true, much remains to be done, because these works generally characterize materials but not specific applications that allow the agents of the construction to have the guarantees required by the projects. Therefore, it is necessary the involvement of all the actors included in the life cycle of these new construction materials, and also to promote its use for, for example, definition of standards, tax advantages or market intervention is necessary. This paper presents the main findings reached in "Waste to resources (W2R)" project since it began in October 2014. The main goal of the project is to develop new materials, elements and construction systems, manufactured from CDW, to be used in improving the energy efficiency of buildings. Other objectives of the project are: to quantify the CDW generated in the energy rehabilitation works, specifically wastes from the building envelope; and to study the traceability of CDW generated and promote CDW reuse and recycle in order to get close to the life cycle of buildings, generating zero waste and reducing the ecological footprint of the construction sector. This paper determines the most important aspects to consider during the design of new constructive solutions, which improve the energy efficiency of buildings and what materials made with CDW would be the most suitable for that. Also, a survey to select best practices for reducing "close to zero waste" in refurbishment was done. Finally, several pilot rehabilitation works conform the parameters analyzed in the project were selected, in order to apply the results and thus compare the theoretical with reality. Acknowledgements: This research was supported by the Spanish State Secretariat for Research, Development and Innovation of the Ministry of Economy and Competitiveness under "Waste 2 Resources" Project (BIA2013-43061-R).Keywords: building waste, construction and demolition waste, recycling, resources
Procedia PDF Downloads 25228431 Theorizing Income Inequality in the Face of Financial Globalization
Authors: Li Sheng
Abstract:
Based on an extended post-Keynesian model, we find that the association between the savings rate and income inequality is negative if savers’ funds are borrowed by spending households for consumption but positive if savings are channeled to investing firms for production. A negative association, such as the one that exists in the U.S., hinges on an income illusion created by an asset bubble and cheap credit. Thus, financial globalization leads consumption and income inequality to diverge, and the divergence is more extreme if lower-income groups have higher debt ratios. A positive association, such as the one that exists in China, relates to liquidity constraints faced by consumers such that consumption inequality closely follows income inequality. Our results imply that income inequality must be reduced in both types of countries to increase savings in deficit economies with negative associations and to reduce savings in surplus economies with positive associations.Keywords: savings rate, income inequality, financial globalization, global imbalances
Procedia PDF Downloads 47028430 Modeling of Global Solar Radiation on a Horizontal Surface Using Artificial Neural Network: A Case Study
Authors: Laidi Maamar, Hanini Salah
Abstract:
The present work investigates the potential of artificial neural network (ANN) model to predict the horizontal global solar radiation (HGSR). The ANN is developed and optimized using three years meteorological database from 2011 to 2013 available at the meteorological station of Blida (Blida 1 university, Algeria, Latitude 36.5°, Longitude 2.81° and 163 m above mean sea level). Optimal configuration of the ANN model has been determined by minimizing the Root Means Square Error (RMSE) and maximizing the correlation coefficient (R2) between observed and predicted data with the ANN model. To select the best ANN architecture, we have conducted several tests by using different combinations of parameters. A two-layer ANN model with six hidden neurons has been found as an optimal topology with (RMSE=4.036 W/m²) and (R²=0.999). A graphical user interface (GUI), was designed based on the best network structure and training algorithm, to enhance the users’ friendliness application of the model.Keywords: artificial neural network, global solar radiation, solar energy, prediction, Algeria
Procedia PDF Downloads 49928429 Heuristic Algorithms for Time Based Weapon-Target Assignment Problem
Authors: Hyun Seop Uhm, Yong Ho Choi, Ji Eun Kim, Young Hoon Lee
Abstract:
Weapon-target assignment (WTA) is a problem that assigns available launchers to appropriate targets in order to defend assets. Various algorithms for WTA have been developed over past years for both in the static and dynamic environment (denoted by SWTA and DWTA respectively). Due to the problem requirement to be solved in a relevant computational time, WTA has suffered from the solution efficiency. As a result, SWTA and DWTA problems have been solved in the limited situation of the battlefield. In this paper, the general situation under continuous time is considered by Time based Weapon Target Assignment (TWTA) problem. TWTA are studied using the mixed integer programming model, and three heuristic algorithms; decomposed opt-opt, decomposed opt-greedy, and greedy algorithms are suggested. Although the TWTA optimization model works inefficiently when it is characterized by a large size, the decomposed opt-opt algorithm based on the linearization and decomposition method extracted efficient solutions in a reasonable computation time. Because the computation time of the scheduling part is too long to solve by the optimization model, several algorithms based on greedy is proposed. The models show lower performance value than that of the decomposed opt-opt algorithm, but very short time is needed to compute. Hence, this paper proposes an improved method by applying decomposition to TWTA, and more practical and effectual methods can be developed for using TWTA on the battlefield.Keywords: air and missile defense, weapon target assignment, mixed integer programming, piecewise linearization, decomposition algorithm, military operations research
Procedia PDF Downloads 33728428 Seismic Response of Moment Resisting Steel Frame with Hysteresis Envelope Model of Joints
Authors: Krolo Paulina
Abstract:
The seismic response of moment-resisting steel frames depends on the behavior of the joints, especially when they are considered as ductile zones. The aim of this research is to provide a realistic assessment of the moment-resisting steel frame behavior under seismic loading using nonlinear static pushover analysis (N2 method). The hysteresis behavior of the joints in the frame model was described using a new hysteresis envelope model. The obtained seismic response was compared with the results of the seismic analysis obtained for the same steel frame that takes into account the monotonic model of the joints.Keywords: beam-to-column joints, hysteresis envelope model, moment-resisting frame, nonlinear static pushover analysis, N2 method
Procedia PDF Downloads 26828427 Modeling of Conjugate Heat Transfer including Radiation in a Kerosene/Air Certification Burner
Authors: Lancelot Boulet, Pierre Benard, Ghislain Lartigue, Vincent Moureau, Nicolas Chauvet, Sheddia Didorally
Abstract:
International aeronautic standards demand a fire certification for engines that demonstrate their resistance. This demonstration relies on tests performed with prototype engines in the late stages of the development. Hardest tests require to place a kerosene standardized flame in front of the engine casing during a given time with imposed temperature and heat flux. The purpose of this work is to provide a better characterization of a kerosene/air certification burner in order to minimize the risks of test failure. A first Large-Eddy Simulation (LES) study of the certification burner permitted to model and simulate this burner, including both adiabatic and Conjugate Heat Transfer (CHT) computations. Carried out on unstructured grids with 40 million tetrahedral cells, using the finite-volume YALES2 code, spray combustion, forced convection on walls and conduction in the solid parts of the burner were coupled to achieve a detailed description of heat transfer. It highlighted the fact that conduction inside the solid has a real impact on the flame topology and the combustion regime. However, in the absence of radiative heat transfer, unrealistic temperature of the equipment was obtained. The aim of the present study is to include the radiative heat transfer in order to reach the same temperature given by experimental measurements. First, various test-cases are conducted to validate the coupling between the different heat solvers. Then, adiabatic case, CHT case, as well as CHT including radiative transfer are studied and compared. The LES model is finally applied to investigate the heat transfer in a flame impaction configuration. The aim is to progress on fire test modeling so as to reach a good confidence level as far as success of the certification test is concerned.Keywords: conjugate heat transfer, fire resistance test, large-eddy simulation, radiative transfer, turbulent combustion
Procedia PDF Downloads 22428426 Automatic Target Recognition in SAR Images Based on Sparse Representation Technique
Authors: Ahmet Karagoz, Irfan Karagoz
Abstract:
Synthetic Aperture Radar (SAR) is a radar mechanism that can be integrated into manned and unmanned aerial vehicles to create high-resolution images in all weather conditions, regardless of day and night. In this study, SAR images of military vehicles with different azimuth and descent angles are pre-processed at the first stage. The main purpose here is to reduce the high speckle noise found in SAR images. For this, the Wiener adaptive filter, the mean filter, and the median filters are used to reduce the amount of speckle noise in the images without causing loss of data. During the image segmentation phase, pixel values are ordered so that the target vehicle region is separated from other regions containing unnecessary information. The target image is parsed with the brightest 20% pixel value of 255 and the other pixel values of 0. In addition, by using appropriate parameters of statistical region merging algorithm, segmentation comparison is performed. In the step of feature extraction, the feature vectors belonging to the vehicles are obtained by using Gabor filters with different orientation, frequency and angle values. A number of Gabor filters are created by changing the orientation, frequency and angle parameters of the Gabor filters to extract important features of the images that form the distinctive parts. Finally, images are classified by sparse representation method. In the study, l₁ norm analysis of sparse representation is used. A joint database of the feature vectors generated by the target images of military vehicle types is obtained side by side and this database is transformed into the matrix form. In order to classify the vehicles in a similar way, the test images of each vehicle is converted to the vector form and l₁ norm analysis of the sparse representation method is applied through the existing database matrix form. As a result, correct recognition has been performed by matching the target images of military vehicles with the test images by means of the sparse representation method. 97% classification success of SAR images of different military vehicle types is obtained.Keywords: automatic target recognition, sparse representation, image classification, SAR images
Procedia PDF Downloads 36828425 Cooperative Game Theory and Small Hold Farming: Towards A Conceptual Model
Authors: Abel Kahuni
Abstract:
Cooperative game theory (CGT) postulates that groups of players are crucial units of the decision-making and impose cooperative behaviour. Accordingly, cooperative games are regarded as competition between coalitions of players, rather than between individual players. However, the basic supposition in CGT is that the cooperative is formed by all players. One of the emerging questions in CGT is how to develop cooperatives and fairly allocate the payoff. Cooperative Game Theory (CGT) may provide a framework and insights into the ways small holder farmers in rural resettlements may develop competitive advantage through marketing cooperatives. This conceptual paper proposes a non-competition model for small holder farmers of homogenous agri-commodity under CGT conditions. This paper will also provide brief insights into to the theory of cooperative games in-order to generate an understanding of CGT, cooperative marketing gains and its application in small holder farming arrangements. Accordingly, the objective is to provide a basic introduction to this theory in connection with economic competitive theories in the context of small holder farmers. The key value proposition of CGT is the equitable and fair sharing of cooperative gains.Keywords: game theory, cooperative game theory, cooperatives, competition
Procedia PDF Downloads 7928424 XAI Implemented Prognostic Framework: Condition Monitoring and Alert System Based on RUL and Sensory Data
Authors: Faruk Ozdemir, Roy Kalawsky, Peter Hubbard
Abstract:
Accurate estimation of RUL provides a basis for effective predictive maintenance, reducing unexpected downtime for industrial equipment. However, while models such as the Random Forest have effective predictive capabilities, they are the so-called ‘black box’ models, where interpretability is at a threshold to make critical diagnostic decisions involved in industries related to aviation. The purpose of this work is to present a prognostic framework that embeds Explainable Artificial Intelligence (XAI) techniques in order to provide essential transparency in Machine Learning methods' decision-making mechanisms based on sensor data, with the objective of procuring actionable insights for the aviation industry. Sensor readings have been gathered from critical equipment such as turbofan jet engine and landing gear, and the prediction of the RUL is done by a Random Forest model. It involves steps such as data gathering, feature engineering, model training, and evaluation. These critical components’ datasets are independently trained and evaluated by the models. While suitable predictions are served, their performance metrics are reasonably good; such complex models, however obscure reasoning for the predictions made by them and may even undermine the confidence of the decision-maker or the maintenance teams. This is followed by global explanations using SHAP and local explanations using LIME in the second phase to bridge the gap in reliability within industrial contexts. These tools analyze model decisions, highlighting feature importance and explaining how each input variable affects the output. This dual approach offers a general comprehension of the overall model behavior and detailed insight into specific predictions. The proposed framework, in its third component, incorporates the techniques of causal analysis in the form of Granger causality tests in order to move beyond correlation toward causation. This will not only allow the model to predict failures but also present reasons, from the key sensor features linked to possible failure mechanisms to relevant personnel. The causality between sensor behaviors and equipment failures creates much value for maintenance teams due to better root cause identification and effective preventive measures. This step contributes to the system being more explainable. Surrogate Several simple models, including Decision Trees and Linear Models, can be used in yet another stage to approximately represent the complex Random Forest model. These simpler models act as backups, replicating important jobs of the original model's behavior. If the feature explanations obtained from the surrogate model are cross-validated with the primary model, the insights derived would be more reliable and provide an intuitive sense of how the input variables affect the predictions. We then create an iterative explainable feedback loop, where the knowledge learned from the explainability methods feeds back into the training of the models. This feeds into a cycle of continuous improvement both in model accuracy and interpretability over time. By systematically integrating new findings, the model is expected to adapt to changed conditions and further develop its prognosis capability. These components are then presented to the decision-makers through the development of a fully transparent condition monitoring and alert system. The system provides a holistic tool for maintenance operations by leveraging RUL predictions, feature importance scores, persistent sensor threshold values, and autonomous alert mechanisms. Since the system will provide explanations for the predictions given, along with active alerts, the maintenance personnel can make informed decisions on their end regarding correct interventions to extend the life of the critical machinery.Keywords: predictive maintenance, explainable artificial intelligence, prognostic, RUL, machine learning, turbofan engines, C-MAPSS dataset
Procedia PDF Downloads 928423 Models of Copyrights System
Authors: A. G. Matveev
Abstract:
The copyrights system is a combination of different elements. The number, content and the correlation of these elements are different for different legal orders. The models of copyrights systems display this system in terms of the interaction of economic and author's moral rights. Monistic and dualistic models are the most popular ones. The article deals with different points of view on the monism and dualism in copyright system. A specific model of the copyright in Switzerland in the XXth century is analyzed. The evolution of a French dualistic model of copyright is shown. The author believes that one should talk not about one, but rather about a number of dualism forms of copyright system.Keywords: copyright, exclusive copyright, economic rights, author's moral rights, rights of personality, monistic model, dualistic model
Procedia PDF Downloads 420