Search results for: stochastic matrices
208 Numerical Simulations of Acoustic Imaging in Hydrodynamic Tunnel with Model Adaptation and Boundary Layer Noise Reduction
Authors: Sylvain Amailland, Jean-Hugh Thomas, Charles Pézerat, Romuald Boucheron, Jean-Claude Pascal
Abstract:
The noise requirements for naval and research vessels have seen an increasing demand for quieter ships in order to fulfil current regulations and to reduce the effects on marine life. Hence, new methods dedicated to the characterization of propeller noise, which is the main source of noise in the far-field, are needed. The study of cavitating propellers in closed-section is interesting for analyzing hydrodynamic performance but could involve significant difficulties for hydroacoustic study, especially due to reverberation and boundary layer noise in the tunnel. The aim of this paper is to present a numerical methodology for the identification of hydroacoustic sources on marine propellers using hydrophone arrays in a large hydrodynamic tunnel. The main difficulties are linked to the reverberation of the tunnel and the boundary layer noise that strongly reduce the signal-to-noise ratio. In this paper it is proposed to estimate the reflection coefficients using an inverse method and some reference transfer functions measured in the tunnel. This approach allows to reduce the uncertainties of the propagation model used in the inverse problem. In order to reduce the boundary layer noise, a cleaning algorithm taking advantage of the low rank and sparse structure of the cross-spectrum matrices of the acoustic and the boundary layer noise is presented. This approach allows to recover the acoustic signal even well under the boundary layer noise. The improvement brought by this method is visible on acoustic maps resulting from beamforming and DAMAS algorithms.Keywords: acoustic imaging, boundary layer noise denoising, inverse problems, model adaptation
Procedia PDF Downloads 336207 Modified Silicates as Dissolved Oxygen Sensors in Water: Structural and Optical Properties
Authors: Andile Mkhohlakali, Tien-Chien Jen, James Tshilongo, Happy Mabowa
Abstract:
Among different parameters, oxygen is one of the most important analytes of interest, dissolved oxygen (DO) concentration is very crucial and significant for various areas of physical, chemical, and environmental monitoring. Herein we report oxygen-sensitive luminophores -based lanthanum(III) trifluoromethanesulfonate), [La]³⁺ was encapsulated into SiO₂-based xerogel matrix. The nanosensor is composed of organically modified silica nanoparticles, doped with the luminescent oxygen–sensitive lanthanum(III) trifluoromethanesulfonate complex. The precursor materials used for sensing film were triethyl ethoxy silane (TEOS) and (3-Mercaptopropyltriethoxysilane) (MPTMS- TEOS) used for SiO2-baed matrices. Brunauer–Emmett–Teller (BET), and BJH indicate that the SiO₂ transformed from microporous to mesoporous upon the addition of La³⁺ luminophore with increased surface area (SBET). The typical amorphous SiO₂ based xerogels were revealed with X-Ray diffraction (XRD) and Selected Area Electron Diffraction (SAED) analysis. Scanning electron microscope- (SEM) and transmission electron microscope (TEM) showed the porous morphology and reduced particle for SiO₂ and La-SiO₂ xerogels respectively. The existence of elements, siloxane networks, and thermal stability of xerogel was confirmed by energy dispersive spectroscopy (EDS), Fourier-transform infrared spectroscopy (FTIR), and Thermographic analysis (TGA). UV-Vis spectroscopy and photoluminescence (PL) have been used to characterize the optical properties of xerogels. La-SiO₂ demonstrates promising characteristic features of an active sensing film for dissolved oxygen in the water. Keywords: Sol-gel, ORMOSILs, encapsulation, Luminophores quenching, O₂-sensingKeywords: sol-gel, ORMOSILs, luminophores quenching, O₂-sensing
Procedia PDF Downloads 123206 Using Cyclic Structure to Improve Inference on Network Community Structure
Authors: Behnaz Moradijamei, Michael Higgins
Abstract:
Identifying community structure is a critical task in analyzing social media data sets often modeled by networks. Statistical models such as the stochastic block model have proven to explain the structure of communities in real-world network data. In this work, we develop a goodness-of-fit test to examine community structure's existence by using a distinguishing property in networks: cyclic structures are more prevalent within communities than across them. To better understand how communities are shaped by the cyclic structure of the network rather than just the number of edges, we introduce a novel method for deciding on the existence of communities. We utilize these structures by using renewal non-backtracking random walk (RNBRW) to the existing goodness-of-fit test. RNBRW is an important variant of random walk in which the walk is prohibited from returning back to a node in exactly two steps and terminates and restarts once it completes a cycle. We investigate the use of RNBRW to improve the performance of existing goodness-of-fit tests for community detection algorithms based on the spectral properties of the adjacency matrix. Our proposed test on community structure is based on the probability distribution of eigenvalues of the normalized retracing probability matrix derived by RNBRW. We attempt to make the best use of asymptotic results on such a distribution when there is no community structure, i.e., asymptotic distribution under the null hypothesis. Moreover, we provide a theoretical foundation for our statistic by obtaining the true mean and a tight lower bound for RNBRW edge weights variance.Keywords: hypothesis testing, RNBRW, network inference, community structure
Procedia PDF Downloads 152205 Extended Kalman Filter and Markov Chain Monte Carlo Method for Uncertainty Estimation: Application to X-Ray Fluorescence Machine Calibration and Metal Testing
Authors: S. Bouhouche, R. Drai, J. Bast
Abstract:
This paper is concerned with a method for uncertainty evaluation of steel sample content using X-Ray Fluorescence method. The considered method of analysis is a comparative technique based on the X-Ray Fluorescence; the calibration step assumes the adequate chemical composition of metallic analyzed sample. It is proposed in this work a new combined approach using the Kalman Filter and Markov Chain Monte Carlo (MCMC) for uncertainty estimation of steel content analysis. The Kalman filter algorithm is extended to the model identification of the chemical analysis process using the main factors affecting the analysis results; in this case, the estimated states are reduced to the model parameters. The MCMC is a stochastic method that computes the statistical properties of the considered states such as the probability distribution function (PDF) according to the initial state and the target distribution using Monte Carlo simulation algorithm. Conventional approach is based on the linear correlation, the uncertainty budget is established for steel Mn(wt%), Cr(wt%), Ni(wt%) and Mo(wt%) content respectively. A comparative study between the conventional procedure and the proposed method is given. This kind of approaches is applied for constructing an accurate computing procedure of uncertainty measurement.Keywords: Kalman filter, Markov chain Monte Carlo, x-ray fluorescence calibration and testing, steel content measurement, uncertainty measurement
Procedia PDF Downloads 285204 Replicating Brain’s Resting State Functional Connectivity Network Using a Multi-Factor Hub-Based Model
Authors: B. L. Ho, L. Shi, D. F. Wang, V. C. T. Mok
Abstract:
The brain’s functional connectivity while temporally non-stationary does express consistency at a macro spatial level. The study of stable resting state connectivity patterns hence provides opportunities for identification of diseases if such stability is severely perturbed. A mathematical model replicating the brain’s spatial connections will be useful for understanding brain’s representative geometry and complements the empirical model where it falls short. Empirical computations tend to involve large matrices and become infeasible with fine parcellation. However, the proposed analytical model has no such computational problems. To improve replicability, 92 subject data are obtained from two open sources. The proposed methodology, inspired by financial theory, uses multivariate regression to find relationships of every cortical region of interest (ROI) with some pre-identified hubs. These hubs acted as representatives for the entire cortical surface. A variance-covariance framework of all ROIs is then built based on these relationships to link up all the ROIs. The result is a high level of match between model and empirical correlations in the range of 0.59 to 0.66 after adjusting for sample size; an increase of almost forty percent. More significantly, the model framework provides an intuitive way to delineate between systemic drivers and idiosyncratic noise while reducing dimensions by more than 30 folds, hence, providing a way to conduct attribution analysis. Due to its analytical nature and simple structure, the model is useful as a standalone toolkit for network dependency analysis or as a module for other mathematical models.Keywords: functional magnetic resonance imaging, multivariate regression, network hubs, resting state functional connectivity
Procedia PDF Downloads 153203 The Strategic Entering Time of a Commerce Platform
Authors: Chia-li Wang
Abstract:
The surge of service and commerce platforms, such as e-commerce and internet-of-things, have rapidly changed our lives. How to avoid the congestion and get the job done in the platform is now a common problem that many people encounter every day. This requires platform users to make decisions about when to enter the platform. To that end, we investigate the strategic entering time of a simple platform containing random numbers of buyers and sellers of some item. Upon a trade, the buyer and the seller gain respective profits, yet they pay the cost of waiting in the platform. To maximize their expected payoffs from trading, both buyers and sellers can choose their entering times. This creates an interesting and practical framework of a game that is played among buyers, among sellers, and between them. That is, a strategy employed by a player is not only against players of its type but also a response to those of the other type, and, thus, a strategy profile is composed of strategies of buyers and sellers. The players' best response, the Nash equilibrium (NE) strategy profile, is derived by a pair of differential equations, which, in turn, are used to establish its existence and uniqueness. More importantly, its structure sheds valuable insights of how the entering strategy of one side (buyers or sellers) is affected by the entering behavior of the other side. These results provide a base for the study of dynamic pricing for stochastic demand-supply imbalances. Finally, comparisons between the social welfares (the sum of the payoffs incurred by individual participants) obtained by the optimal strategy and by the NE strategy are conducted for showing the efficiency loss relative to the socially optimal solution. That should help to manage the platform better.Keywords: double-sided queue, non-cooperative game, nash equilibrium, price of anarchy
Procedia PDF Downloads 87202 Overweight and Neurocognitive Functioning: Unraveling the Antagonistic Relationship in Adolescents
Authors: Swati Bajpai, S. P. K Jena
Abstract:
Background: There is dramatic increase in the prevalence and severity of overweight in adolescents, raising concerns about their psychosocial and cognitive consequences, thereby indicating the immediate need to understand the effects of increased weight on scholastic performance. Although the body of research is currently limited, available results have identified an inverse relationship between obesity and cognition in adolescents. Aim: to examine the association between increased Body Mass Index in adolescents and their neurocognitive functioning. Methods: A case –control study of 28 subjects in the age group of 11-17 years (14 Males and 14 females) was taken on the basis of main inclusion criteria (Body Mass Index). All of them were randomized to (experimental group: overweight) and (control group: normal weighted). A complete neurocognitive assessment was carried out using validated psychological scales namely, Color Progressive Matrices (to assess intelligence); Bender Visual Motor Gestalt Test (Perceptual motor functioning); PGI-Memory Scale for Children (memory functioning) and Malin’s Intelligence Scale Indian Children (verbal and performance ability). Results: statistical analysis of the results depicted that 57% of the experimental group lack in cognitive abilities, especially in general knowledge (99.1±12.0 vs. 102.8±6.7), working memory (91.5±8.4 vs. 93.1±8.7), concrete ability (82.3±11.5 vs. 92.6±1.7) and perceptual motor functioning (1.5±1.0 vs. 0.3±0.9) as compared to control group. Conclusion: Our investigations suggest that weight gain results, at least in part, from a neurological predisposition characterized by reduced executive function, and in turn obesity itself has a compounding negative impact on the brain. Though, larger sample is needed to make more affirmative claims.Keywords: adolescents, body mass index, neurocognition, obesity
Procedia PDF Downloads 487201 Electricity Load Modeling: An Application to Italian Market
Authors: Giovanni Masala, Stefania Marica
Abstract:
Forecasting electricity load plays a crucial role regards decision making and planning for economical purposes. Besides, in the light of the recent privatization and deregulation of the power industry, the forecasting of future electricity load turned out to be a very challenging problem. Empirical data about electricity load highlights a clear seasonal behavior (higher load during the winter season), which is partly due to climatic effects. We also emphasize the presence of load periodicity at a weekly basis (electricity load is usually lower on weekends or holidays) and at daily basis (electricity load is clearly influenced by the hour). Finally, a long-term trend may depend on the general economic situation (for example, industrial production affects electricity load). All these features must be captured by the model. The purpose of this paper is then to build an hourly electricity load model. The deterministic component of the model requires non-linear regression and Fourier series while we will investigate the stochastic component through econometrical tools. The calibration of the parameters’ model will be performed by using data coming from the Italian market in a 6 year period (2007- 2012). Then, we will perform a Monte Carlo simulation in order to compare the simulated data respect to the real data (both in-sample and out-of-sample inspection). The reliability of the model will be deduced thanks to standard tests which highlight a good fitting of the simulated values.Keywords: ARMA-GARCH process, electricity load, fitting tests, Fourier series, Monte Carlo simulation, non-linear regression
Procedia PDF Downloads 398200 Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark
Authors: B. Elshafei, X. Mao
Abstract:
The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.Keywords: data fusion, Gaussian process regression, signal denoise, temporal extrapolation
Procedia PDF Downloads 136199 Tractography Analysis and the Evolutionary Origin of Schizophrenia
Authors: Mouktafi Amine, Tahiri Asmaa
Abstract:
A substantial number of traditional medical research has been put forward to managing and treating mental disorders. At the present time, to our best knowledge, it is believed that a fundamental understanding of the underlying causes of the majority of psychological disorders needs to be explored further to inform early diagnosis, managing symptoms and treatment. The emerging field of evolutionary psychology is a promising prospect to address the origin of mental disorders, potentially leading to more effective treatments. Schizophrenia as a topical mental disorder has been linked to the evolutionary adaptation of the human brain represented in the brain connectivity and asymmetry directly linked to humans' higher brain cognition in contrast to other primates being our direct living representation of the structure and connectivity of our earliest common African ancestors. As proposed in the evolutionary psychology scientific literature, the pathophysiology of schizophrenia is expressed and directly linked to altered connectivity between the Hippocampal Formation (HF) and Dorsolateral Prefrontal Cortex (DLPFC). This research paper presents the results of the use of tractography analysis using multiple open access Diffusion Weighted Imaging (DWI) datasets of healthy subjects, schizophrenia-affected subjects and primates to illustrate the relevance of the aforementioned brain regions' connectivity and the underlying evolutionary changes in the human brain. Deterministic fiber tracking and streamline analysis were used to generate connectivity matrices from the DWI datasets overlaid to compute distances and highlight disconnectivity patterns in conjunction with other fiber tracking metrics: Fractional Anisotropy (FA), Mean Diffusivity (MD) and Radial Diffusivity (RD).Keywords: tractography, diffusion weighted imaging, schizophrenia, evolutionary psychology
Procedia PDF Downloads 51198 LiTa2PO8-based Composite Solid Polymer Electrolytes for High-Voltage Cathodes in Lithium-Metal Batteries
Authors: Kumlachew Zelalem Walle, Chun-Chen Yang
Abstract:
Solid-state Lithium metal batteries (SSLMBs) that contain polymer and ceramic solid electrolytes have received considerable attention as an alternative to substitute liquid electrolytes in lithium metal batteries (LMBs) for highly safe, excellent energy storage performance and stability under elevated temperature situations. Here, a novel fast Li-ion conducting material, LiTa₂PO₈ (LTPO), was synthesized and electrochemical performance of as-prepared powder and LTPO-incorporated composite solid polymer electrolyte (LTPO-CPE) membrane were investigated. The as-prepared LTPO powder was homogeneously dispersed in polymer matrices, and a hybrid solid electrolyte membrane was synthesized via a simple solution-casting method. The room temperature total ionic conductivity (σt) of the LTPO pellet and LTPO-CPE membrane were 0.14 and 0.57 mS cm-1, respectively. A coin battery with NCM811 cathode is cycled under 1C between 2.8 to 4.5 V at room temperature, achieving a Coulombic efficiency of 99.3% with capacity retention of 74.1% after 300 cycles. Similarly, the LFP cathode also delivered an excellent performance at 0.5C with an average Coulombic efficiency of 100% without virtually capacity loss (the maximum specific capacity is at 27th: 138 mAh g−1 and 500th: 131.3 mAh g−1). These results demonstrates the feasibility of a high Li-ion conductor LTPO as a filler, and the developed polymer/ceramic hybrid electrolyte has potential to be a high-performance electrolyte for high-voltage cathodes, which may provide a fresh platform for developing more advanced solid-state electrolytes.Keywords: li-ion conductor, lithium-metal batteries, composite solid electrolytes, liTa2PO8, high-voltage cathode
Procedia PDF Downloads 66197 Openness to Linguistic and Value Diversity as a Key Factor in the Development of a Learning Community
Authors: Caterina Calicchio, Talia Sbardella
Abstract:
The ability to move through geographical and symbolic spaces is key for building new nodes and social relationships. Especially in the framework of language learning, accepting and valuing diversity can help to create a constructive atmosphere of cooperation, innovation, and creativity. Thus, it is important to outline the stages of forming a learning community, focusing on the characteristics that can favor its development. It is known that elements like curiosity and motivation are significant for individual language learning; hence, the study attempts to investigate how factors like openness to diversity and cultural immersion could improve Italian learning and teaching. This paper aims to indicate the factors that could be significant for the development of a Learning Community by presenting a case study on a course on Italian as a second language for beginners: first, the theoretical matrices underlying social learning will be outlined. Secondly, a quantitative study will be described based on an adaptation of the openness to diversity and some insights psychometric scale questionnaire developed at the Umbra Institute. The questionnaire was delivered to 52 American college students with open-ended and closed-ended questions. Students were asked to specify their level of agreement to a set of statements on a six-point Likert scale ranging from (1) Strongly disagree to (6) Strongly agree. The data has been analyzed with a quantitative and qualitative method and has been represented in a pie chart and in a histogram. Moreover, mean and frequency have been calculated. The research findings demonstrate that openness to diversity and challenge enhances cross-cutting skills such as intercultural and communicative competence: through cultural immersion and the facility of speaking with locals, the participants have been able to develop their own Italian L2 language community. The goal is to share with the scientific community some insights to trace possible future lines of research.Keywords: Italian as second language, language learning, learning community, openness to diversity
Procedia PDF Downloads 74196 The Analysis of Emergency Shutdown Valves Torque Data in Terms of Its Use as a Health Indicator for System Prognostics
Authors: Ewa M. Laskowska, Jorn Vatn
Abstract:
Industry 4.0 focuses on digital optimization of industrial processes. The idea is to use extracted data in order to build a decision support model enabling use of those data for real time decision making. In terms of predictive maintenance, the desired decision support tool would be a model enabling prognostics of system's health based on the current condition of considered equipment. Within area of system prognostics and health management, a commonly used health indicator is Remaining Useful Lifetime (RUL) of a system. Because the RUL is a random variable, it has to be estimated based on available health indicators. Health indicators can be of different types and come from different sources. They can be process variables, equipment performance variables, data related to number of experienced failures, etc. The aim of this study is the analysis of performance variables of emergency shutdown valves (ESV) used in oil and gas industry. ESV is inspected periodically, and at each inspection torque and time of valve operation are registered. The data will be analyzed by means of machine learning or statistical analysis. The purpose is to investigate whether the available data could be used as a health indicator for a prognostic purpose. The second objective is to examine what is the most efficient way to incorporate the data into predictive model. The idea is to check whether the data can be applied in form of explanatory variables in Markov process or whether other stochastic processes would be a more convenient to build an RUL model based on the information coming from registered data.Keywords: emergency shutdown valves, health indicator, prognostics, remaining useful lifetime, RUL
Procedia PDF Downloads 91195 A TgCNN-Based Surrogate Model for Subsurface Oil-Water Phase Flow under Multi-Well Conditions
Authors: Jian Li
Abstract:
The uncertainty quantification and inversion problems of subsurface oil-water phase flow usually require extensive repeated forward calculations for new runs with changed conditions. To reduce the computational time, various forms of surrogate models have been built. Related research shows that deep learning has emerged as an effective surrogate model, while most surrogate models with deep learning are purely data-driven, which always leads to poor robustness and abnormal results. To guarantee the model more consistent with the physical laws, a coupled theory-guided convolutional neural network (TgCNN) based surrogate model is built to facilitate computation efficiency under the premise of satisfactory accuracy. The model is a convolutional neural network based on multi-well reservoir simulation. The core notion of this proposed method is to bridge two separate blocks on top of an overall network. They underlie the TgCNN model in a coupled form, which reflects the coupling nature of pressure and water saturation in the two-phase flow equation. The model is driven by not only labeled data but also scientific theories, including governing equations, stochastic parameterization, boundary, and initial conditions, well conditions, and expert knowledge. The results show that the TgCNN-based surrogate model exhibits satisfactory accuracy and efficiency in subsurface oil-water phase flow under multi-well conditions.Keywords: coupled theory-guided convolutional neural network, multi-well conditions, surrogate model, subsurface oil-water phase
Procedia PDF Downloads 86194 Solvent-Aided Dispersion of Tannic Acid to Enhance Flame Retardancy of Epoxy
Authors: Matthew Korey, Jeffrey Youngblood, John Howarter
Abstract:
Background and Significance: Tannic acid (TA) is a bio-based high molecular weight organic, aromatic molecule that has been found to increase thermal stability and flame retardancy of many polymer matrices when used as an additive. Although it is biologically sourced, TA is a pollutant in industrial wastewater streams, and there is a desire to find applications in which to downcycle this molecule after extraction from these streams. Additionally, epoxy thermosets have revolutionized many industries, but are too flammable to be used in many applications without additives which augment their flame retardancy (FR). Many flame retardants used in epoxy thermosets are synthesized from petroleum-based monomers leading to significant environmental impacts on the industrial scale. Many of these compounds also have significant impacts on human health. Various bio-based modifiers have been developed to improve the FR of the epoxy resin; however, increasing FR of the system without tradeoffs with other properties has proven challenging, especially for TA. Methodologies: In this work, TA was incorporated into the thermoset by use of solvent-exchange using methyl ethyl ketone, a co-solvent for TA, and epoxy resin. Samples were then characterized optically (UV-vis spectroscopy and optical microscopy), thermally (thermogravimetric analysis and differential scanning calorimetry), and for their flame retardancy (mass loss calorimetry). Major Findings: Compared to control samples, all samples were found to have increased thermal stability. Further, the addition of tannic acid to the polymer matrix by the use of solvent greatly increased the compatibility of the additive in epoxy thermosets. By using solvent-exchange, the highest loading level of TA found in literature was achieved in this work (40 wt%). Conclusions: The use of solvent-exchange shows promises for circumventing the limitations of TA in epoxy.Keywords: sustainable, flame retardant, epoxy, tannic acid
Procedia PDF Downloads 131193 Stochastic Optimization of a Vendor-Managed Inventory Problem in a Two-Echelon Supply Chain
Authors: Bita Payami-Shabestari, Dariush Eslami
Abstract:
The purpose of this paper is to develop a multi-product economic production quantity model under vendor management inventory policy and restrictions including limited warehouse space, budget, and number of orders, average shortage time and maximum permissible shortage. Since the “costs” cannot be predicted with certainty, it is assumed that data behave under uncertain environment. The problem is first formulated into the framework of a bi-objective of multi-product economic production quantity model. Then, the problem is solved with three multi-objective decision-making (MODM) methods. Then following this, three methods had been compared on information on the optimal value of the two objective functions and the central processing unit (CPU) time with the statistical analysis method and the multi-attribute decision-making (MADM). The results are compared with statistical analysis method and the MADM. The results of the study demonstrate that augmented-constraint in terms of optimal value of the two objective functions and the CPU time perform better than global criteria, and goal programming. Sensitivity analysis is done to illustrate the effect of parameter variations on the optimal solution. The contribution of this research is the use of random costs data in developing a multi-product economic production quantity model under vendor management inventory policy with several constraints.Keywords: economic production quantity, random cost, supply chain management, vendor-managed inventory
Procedia PDF Downloads 129192 Optimizing Groundwater Pumping for a Complex Groundwater/Surface Water System
Authors: Emery A. Coppola Jr., Suna Cinar, Ferenc Szidarovszky
Abstract:
Over-pumping of groundwater resources is a serious problem world-wide. In addition to depleting this valuable resource, hydraulically connected sensitive ecological resources like wetlands and surface water bodies are often impacted and even destroyed by over-pumping. Effectively managing groundwater in a way that satisfy human demand while preserving natural resources is a daunting challenge that will only worsen with growing human populations and climate change. As presented in this paper, a numerical flow model developed for a hypothetical but realistic groundwater/surface water system was combined with formal optimization. Response coefficients were used in an optimization management model to maximize groundwater pumping in a complex, multi-layered aquifer system while protecting against groundwater over-draft, streamflow depletion, and wetland impacts. Pumping optimization was performed for different constraint sets that reflect different resource protection preferences, yielding significantly different optimal pumping solutions. A sensitivity analysis on the optimal solutions was performed on select response coefficients to identify differences between wet and dry periods. Stochastic optimization was also performed, where uncertainty associated with changing irrigation demand due to changing weather conditions are accounted for. One of the strengths of this optimization approach is that it can efficiently and accurately identify superior management strategies that minimize risk and adverse environmental impacts associated with groundwater pumping under different hydrologic conditions.Keywords: numerical groundwater flow modeling, water management optimization, groundwater overdraft, streamflow depletion
Procedia PDF Downloads 233191 Portfolio Optimization with Reward-Risk Ratio Measure Based on the Mean Absolute Deviation
Authors: Wlodzimierz Ogryczak, Michal Przyluski, Tomasz Sliwinski
Abstract:
In problems of portfolio selection, the reward-risk ratio criterion is optimized to search for a risky portfolio with the maximum increase of the mean return in proportion to the risk measure increase when compared to the risk-free investments. In the classical model, following Markowitz, the risk is measured by the variance thus representing the Sharpe ratio optimization and leading to the quadratic optimization problems. Several Linear Programming (LP) computable risk measures have been introduced and applied in portfolio optimization. In particular, the Mean Absolute Deviation (MAD) measure has been widely recognized. The reward-risk ratio optimization with the MAD measure can be transformed into the LP formulation with the number of constraints proportional to the number of scenarios and the number of variables proportional to the total of the number of scenarios and the number of instruments. This may lead to the LP models with huge number of variables and constraints in the case of real-life financial decisions based on several thousands scenarios, thus decreasing their computational efficiency and making them hardly solvable by general LP tools. We show that the computational efficiency can be then dramatically improved by an alternative model based on the inverse risk-reward ratio minimization and by taking advantages of the LP duality. In the introduced LP model the number of structural constraints is proportional to the number of instruments thus not affecting seriously the simplex method efficiency by the number of scenarios and therefore guaranteeing easy solvability. Moreover, we show that under natural restriction on the target value the MAD risk-reward ratio optimization is consistent with the second order stochastic dominance rules.Keywords: portfolio optimization, reward-risk ratio, mean absolute deviation, linear programming
Procedia PDF Downloads 407190 [Keynote Talk]: Three Dimensional Finite Element Analysis of Functionally Graded Radiation Shielding Nanoengineered Sandwich Composites
Authors: Nasim Abuali Galehdari, Thomas J. Ryan, Ajit D. Kelkar
Abstract:
In recent years, nanotechnology has played an important role in the design of an efficient radiation shielding polymeric composites. It is well known that, high loading of nanomaterials with radiation absorption properties can enhance the radiation attenuation efficiency of shielding structures. However, due to difficulties in dispersion of nanomaterials into polymer matrices, there has been a limitation in higher loading percentages of nanoparticles in the polymer matrix. Therefore, the objective of the present work is to provide a methodology to fabricate and then to characterize the functionally graded radiation shielding structures, which can provide an efficient radiation absorption property along with good structural integrity. Sandwich structures composed of Ultra High Molecular Weight Polyethylene (UHMWPE) fabric as face sheets and functionally graded epoxy nanocomposite as core material were fabricated. A method to fabricate a functionally graded core panel with controllable gradient dispersion of nanoparticles is discussed. In order to optimize the design of functionally graded sandwich composites and to analyze the stress distribution throughout the sandwich composite thickness, a finite element method was used. The sandwich panels were discretized using 3-Dimensional 8 nodded brick elements. Classical laminate analysis in conjunction with simplified micromechanics equations were used to obtain the properties of the face sheets. The presented finite element model would provide insight into deformation and damage mechanics of the functionally graded sandwich composites from the structural point of view.Keywords: nanotechnology, functionally graded material, radiation shielding, sandwich composites, finite element method
Procedia PDF Downloads 469189 The Impact of Window Opening Occupant Behavior Models on Building Energy Performance
Authors: Habtamu Tkubet Ebuy
Abstract:
Purpose Conventional dynamic energy simulation tools go beyond the static dimension of simplified methods by providing better and more accurate prediction of building performance. However, their ability to forecast actual performance is undermined by a low representation of human interactions. The purpose of this study is to examine the potential benefits of incorporating information on occupant diversity into occupant behavior models used to simulate building performance. The co-simulation of the stochastic behavior of the occupants substantially increases the accuracy of the simulation. Design/methodology/approach In this article, probabilistic models of the "opening and closing" behavior of the window of inhabitants have been developed in a separate multi-agent platform, SimOcc, and implemented in the building simulation, TRNSYS, in such a way that the behavior of the window with the interconnectivity can be reflected in the simulation analysis of the building. Findings The results of the study prove that the application of complex behaviors is important to research in predicting actual building performance. The results aid in the identification of the gap between reality and existing simulation methods. We hope this study and its results will serve as a guide for researchers interested in investigating occupant behavior in the future. Research limitations/implications Further case studies involving multi-user behavior for complex commercial buildings need to more understand the impact of the occupant behavior on building performance. Originality/value This study is considered as a good opportunity to achieve the national strategy by showing a suitable tool to help stakeholders in the design phase of new or retrofitted buildings to improve the performance of office buildings.Keywords: occupant behavior, co-simulation, energy consumption, thermal comfort
Procedia PDF Downloads 105188 In Search for the 'Bilingual Advantage' in Immersion Education
Authors: M. E. Joret, F. Germeys, P. Van de Craen
Abstract:
Background: Previous studies have shown that ‘full’ bilingualism seems to enhance the executive functions in children, young adults and elderly people. Executive functions refer to a complex cognitive system responsible for self-controlled and planned behavior and seem to predict academic achievement. The present study aimed at investigating whether similar effects could be found in children learning their second language at school in immersion education programs. Methods: In this study, 44 children involved in immersion education for 4 to 5 years were compared to 48 children in traditional schools. All children were between 9 and 11 years old. To assess executive functions, the Simon task was used, a neuropsychological measure assessing executive functions with reaction times and accuracy on congruent and incongruent trials. To control for background measures, all children underwent the Raven’s coloured progressive matrices, to measure non-verbal intelligence and the Echelle de Vocabulaire en Images Peabody (EVIP), assessing verbal intelligence. In addition, a questionnaire was given to the parents to control for other confounding variables, such as socio-economic status (SES), home language, developmental disorders, etc. Results: There were no differences between groups concerning non-verbal intelligence and verbal intelligence. Furthermore, the immersion learners showed overall faster reaction times on both congruent and incongruent trials compared to the traditional learners, but only after 5 years of training, not before. Conclusion: These results show that the cognitive benefits found in ‘full’ bilinguals also appear in children involved in immersion education, but only after a sufficient exposure to the second language. Our results suggest that the amount of second language training needs to be sufficient before these cognitive effects may emerge.Keywords: bilingualism, executive functions, immersion education, Simon task
Procedia PDF Downloads 442187 A Closed-Loop Design Model for Sustainable Manufacturing by Integrating Forward Design and Reverse Design
Authors: Yuan-Jye Tseng, Yi-Shiuan Chen
Abstract:
In this paper, a new concept of closed-loop design model is presented. The closed-loop design model is developed by integrating forward design and reverse design. Based on this new concept, a closed-loop design model for sustainable manufacturing by integrated evaluation of forward design, reverse design, and green manufacturing using a fuzzy analytic network process is developed. In the design stage of a product, with a given product requirement and objective, there can be different ways to design the detailed components and specifications. Therefore, there can be different design cases to achieve the same product requirement and objective. Thus, in the design evaluation stage, it is required to analyze and evaluate the different design cases. The purpose of this research is to develop a model for evaluating the design cases by integrated evaluation of forward design, reverse design, and green manufacturing models. A fuzzy analytic network process model is presented for integrated evaluation of the criteria in the three models. The comparison matrices for evaluating the criteria in the three groups are established. The total relational values among the three groups represent the total relational effects. In application, a super matrix can be created and the total relational values can be used to evaluate the design cases for decision-making to select the final design case. An example product is demonstrated in this presentation. It shows that the model is useful for integrated evaluation of forward design, reverse design, and green manufacturing to achieve a closed-loop design for sustainable manufacturing objective.Keywords: design evaluation, forward design, reverse design, closed-loop design, supply chain management, closed-loop supply chain, fuzzy analytic network process
Procedia PDF Downloads 676186 Adding Business Value in Enterprise Applications through Quality Matrices Using Agile
Authors: Afshan Saad, Muhammad Saad, Shah Muhammad Emaduddin
Abstract:
Nowadays the business condition is so quick paced that enhancing ourselves consistently has turned into a huge factor for the presence of an undertaking. We can check this for structural building and significantly more so in the quick-paced universe of data innovation and programming designing. The lithe philosophies, similar to Scrum, have a devoted advance in the process that objectives the enhancement of the improvement procedure and programming items. Pivotal to process enhancement is to pick up data that grants you to assess the condition of the procedure and its items. From the status data, you can design activities for the upgrade and furthermore assess the accomplishment of those activities. This investigation builds a model that measures the product nature of the improvement procedure. The product quality is dependent on the useful and auxiliary nature of the product items, besides the nature of the advancement procedure is likewise vital to enhance programming quality. Utilitarian quality covers the adherence to client prerequisites, while the auxiliary quality tends to the structure of the product item's source code with reference to its practicality. The procedure quality is identified with the consistency and expectedness of the improvement procedure. The product quality model is connected in a business setting by social occasion the information for the product measurements in the model. To assess the product quality model, we investigate the information and present it to the general population engaged with the light-footed programming improvement process. The outcomes from the application and the client input recommend that the model empowers a reasonable evaluation of the product quality and that it very well may be utilized to help the persistent enhancement of the advancement procedure and programming items.Keywords: Agile SDLC Tools, Agile Software development, business value, enterprise applications, IBM, IBM Rational Team Concert, RTC, software quality, software metrics
Procedia PDF Downloads 175185 Modeling Flow and Deposition Characteristics of Solid CO2 during Choked Flow of CO2 Pipeline in CCS
Authors: Teng lin, Li Yuxing, Han Hui, Zhao Pengfei, Zhang Datong
Abstract:
With the development of carbon capture and storage (CCS), the flow assurance of CO2 transportation becomes more important, particularly for supercritical CO2 pipelines. The relieving system using the choke valve is applied to control the pressure in CO2 pipeline. However, the temperature of fluid would drop rapidly because of Joule-Thomson cooling (JTC), which may cause solid CO2 form and block the pipe. In this paper, a Computational Fluid Dynamic (CFD) model, using the modified Lagrangian method, Reynold's Stress Transport model (RSM) for turbulence and stochastic tracking model (STM) for particle trajectory, was developed to predict the deposition characteristic of solid carbon dioxide. The model predictions were in good agreement with the experiment data published in the literature. It can be observed that the particle distribution affected the deposition behavior. In the region of the sudden expansion, the smaller particles accumulated tightly on the wall were dominant for pipe blockage. On the contrary, the size of solid CO2 particles deposited near the outlet usually was bigger and the stacked structure was looser. According to the calculation results, the movement of the particles can be regarded as the main four types: turbulent motion close to the sudden expansion structure, balanced motion at sudden expansion-middle region, inertial motion near the outlet and the escape. Furthermore the particle deposits accumulated primarily in the sudden expansion region, reattachment region and outlet region because of the four type of motion. Also the Stokes number had an effect on the deposition ratio and it is recommended for Stokes number to avoid 3-8St.Keywords: carbon capture and storage, carbon dioxide pipeline, gas-particle flow, deposition
Procedia PDF Downloads 370184 Time-Dependent Reliability Analysis of Corrosion Affected Cast Iron Pipes with Mixed Mode Fracture
Authors: Chun-Qing Li, Guoyang Fu, Wei Yang
Abstract:
A significant portion of current water networks is made of cast iron pipes. Due to aging and deterioration with corrosion being the most predominant mechanism, the failure rate of cast iron pipes is very high. Although considerable research has been carried out in the past few decades, most are on the effect of corrosion on the structural capacity of pipes using strength theory as the failure criterion. This paper presents a reliability-based methodology for the assessment of corrosion affected cast iron pipe cracking failures. A nonlinear limit state function taking into account all three fracture modes is proposed for brittle metal pipes with mixed mode fracture. A stochastic model of the load effect is developed, and time-dependent reliability method is employed to quantify the probability of failure and predict the remaining service life. A case study is carried out using the proposed methodology, followed by sensitivity analysis to investigate the effects of the random variables on the probability of failure. It has been found that the larger the inclination angle or the Mode I fracture toughness is, the smaller the probability of pipe failure is. It has also been found that the multiplying and exponential coefficients k and n in the power law corrosion model and the internal pressure have the most influence on the probability of failure for cast iron pipes. The methodology presented in this paper can assist pipe engineers and asset managers in developing a risk-informed and cost-effective strategy for better management of corrosion-affected pipelines.Keywords: corrosion, inclined surface cracks, pressurized cast iron pipes, stress intensity
Procedia PDF Downloads 322183 Revenue Management of Perishable Products Considering Freshness and Price Sensitive Customers
Authors: Onur Kaya, Halit Bayer
Abstract:
Global grocery and supermarket sales are among the largest markets in the world and perishable products such as fresh produce, dairy and meat constitute the biggest section of these markets. Due to their deterioration over time, the demand for these products depends highly on their freshness. They become totally obsolete after a certain amount of time causing a high amount of wastage and decreases in grocery profits. In addition, customers are asking for higher product variety in perishable product categories, leading to less predictable demand per product and to more out-dating. Effective management of these perishable products is an important issue since it is observed that billions of dollars’ worth of food is expired and wasted every month. We consider coordinated inventory and pricing decisions for perishable products with a time and price dependent random demand function. We use stochastic dynamic programming to model this system for both periodically-reviewed and continuously-reviewed inventory systems and prove certain structural characteristics of the optimal solution. We prove that the optimal ordering decision scenario has a monotone structure and the optimal price value decreases by time. However, the optimal price changes in a non-monotonic structure with respect to inventory size. We also analyze the effect of 1 different parameters on the optimal solution through numerical experiments. In addition, we analyze simple-to-implement heuristics, investigate their effectiveness and extract managerial insights. This study gives valuable insights about the management of perishable products in order to decrease wastage and increase profits.Keywords: age-dependent demand, dynamic programming, perishable inventory, pricing
Procedia PDF Downloads 247182 The Effect of Mixing and Degassing Conditions on the Properties of Epoxy/Anhydride Resin System
Authors: Latha Krishnan, Andrew Cobley
Abstract:
Epoxy resin is most widely used as matrices for composites of aerospace, automotive and electronic applications due to its outstanding mechanical properties. These properties are chiefly predetermined by the chemical structure of the prepolymer and type of hardener but can also be varied by the processing conditions such as prepolymer and hardener mixing, degassing and curing conditions. In this research, the effect of degassing on the curing behaviour and the void occurrence is experimentally evaluated for epoxy /anhydride resin system. The epoxy prepolymer was mixed with an anhydride hardener and accelerator in an appropriate quantity. In order to investigate the effect of degassing on the curing behaviour and void content of the resin, the uncured resin samples were prepared using three different methods: 1) no degassing 2) degassing on prepolymer and 3) degassing on mixed solution of prepolymer and hardener with an accelerator. The uncured resins were tested in differential scanning calorimeter (DSC) to observe the changes in curing behaviour of the above three resin samples by analysing factors such as gel temperature, peak cure temperature and heat of reaction/heat flow in curing. Additionally, the completely cured samples were tested in DSC to identify the changes in the glass transition temperature (Tg) between the three samples. In order to evaluate the effect of degassing on the void content and morphology changes in the cured epoxy resin, the fractured surfaces of cured epoxy resin were examined under the scanning electron microscope (SEM). Also, the changes in the mechanical properties of the cured resin were studied by three-point bending test. It was found that degassing at different stages of resin mixing had significant effects on properties such as glass transition temperature, the void content and void size of the epoxy/anhydride resin system. For example, degassing (vacuum applied on the mixed resin) has shown higher glass transition temperature (Tg) with lower void content.Keywords: anhydride epoxy, curing behaviour, degassing, void occurrence
Procedia PDF Downloads 349181 Advanced Bio-Composite Materials Based on Biopolymer Blends and Cellulose Nanocrystals
Authors: Zineb Kassab, Nassima El Miri, A. Aboulkas, Abdellatif Barakat, Mounir El Achaby
Abstract:
Recently, more attention has been given to biopolymers with a focus on sustainable development and environmental preservation. Following this tendency, the attempt has been made to replace polymers derived from petroleum with superior biodegradable polymers (biopolymers). In this context, biopolymers are considered potential replacements for conventional plastic materials. However, some of their properties must be improved for better competitiveness, especially regarding their mechanical, thermal and barrier properties. Bio-nanocomposite technology using nanofillers has already been proven as an effective way to produce new materials with specific properties and high performances. With the emergence of nanostructured bio-composite materials, incorporating elongated rod-like cellulose nanocrystals (CNC) has attracted more and more attention in the field of nanotechnology. This study is aimed to develop bio-composite films of biopolymer matrices [Carboxymethyle cellulose (CMC), Starch (ST), Chitosan (CS) and Polyvinyl alcohol (PVA)] reinforced with cellulose nanocrystals (CNC) using the solution casting method. The CNC were extracted at a nanometric scale from lignocellulosic fibers via sulfuric acid hydrolysis and then characterized using X-ray diffraction (XRD), thermogravimetric analysis (TGA), confocal microscopy, infrared spectroscopy (IR), atomic force and transmission electron microscopies (AFM and TEM) techniques. The as extracted CNC were used as a reinforcing phase to produce a variety of bio-composite films at different CNC loading (0.5-10 wt %) with specific properties. The rheological properties of film-forming solutions (FFS) of bio-composites were studied, and their relation to the casting process was evaluated. Then, the structural, optical transparency, water vapor permeability, thermal stability and mechanical properties of all prepared bio-composite films were evaluated and studied in this report. The high performances of these bio-composite films are expected to have potential in biomaterials or packaging applications.Keywords: biopolymer composites, cellulose nanocrystals, food packaging, lignocellulosic fibers
Procedia PDF Downloads 240180 Aggregation of Electric Vehicles for Emergency Frequency Regulation of Two-Area Interconnected Grid
Authors: S. Agheb, G. Ledwich, G.Walker, Z.Tong
Abstract:
Frequency control has become more of concern for reliable operation of interconnected power systems due to the integration of low inertia renewable energy sources to the grid and their volatility. Also, in case of a sudden fault, the system has less time to recover before widespread blackouts. Electric Vehicles (EV)s have the potential to cooperate in the Emergency Frequency Regulation (EFR) by a nonlinear control of the power system in case of large disturbances. The time is not adequate to communicate with each individual EV on emergency cases, and thus, an aggregate model is necessary for a quick response to prevent from much frequency deviation and the occurrence of any blackout. In this work, an aggregate of EVs is modelled as a big virtual battery in each area considering various aspects of uncertainty such as the number of connected EVs and their initial State of Charge (SOC) as stochastic variables. A control law was proposed and applied to the aggregate model using Lyapunov energy function to maximize the rate of reduction of total kinetic energy in a two-area network after the occurrence of a fault. The control methods are primarily based on the charging/ discharging control of available EVs as shunt capacity in the distribution system. Three different cases were studied considering the locational aspect of the model with the virtual EV either in the center of the two areas or in the corners. The simulation results showed that EVs could help the generator lose its kinetic energy in a short time after a contingency. Earlier estimation of possible contributions of EVs can help the supervisory control level to transmit a prompt control signal to the subsystems such as the aggregator agents and the grid. Thus, the percentage of EVs contribution for EFR will be characterized in the future as the goal of this study.Keywords: emergency frequency regulation, electric vehicle, EV, aggregation, Lyapunov energy function
Procedia PDF Downloads 100179 Consequence of Multi-Templating of Closely Related Structural Analogues on a Chitosan-Methacryllic Acid Molecularly Imprinted Polymer Matrix-Thermal and Chromatographic Traits
Authors: O.Ofoegbu, S. Roongnapa, A.N. Eboatu
Abstract:
Most polluted environments, most challengingly, aerosol types, contain a cocktail of different toxicants. Multi-templating of matrices have been the recent target by researchers in a bid to solving complex mixed-toxicant challenges using single or common remediation systems. This investigation looks at the effect of such multi-templated system vis-a-vis the synthesis by non-covalent interaction, of a molecularly imprinted polymer architecture using nicotine and its structural analogue Phenylalanine amide individually and, in the blend, (50:50), as template materials in a Chitosan-Methacrylic acid functional monomer matrix. The temperature for polymerization is 60OC and time for polymerization, 12hrs (water bath heating), 4mins for (microwave heating). The characteristic thermal properties of the molecularly imprinted materials are investigated using Simultaneous Thermal Analysis (STA) profiling, while the absorption and separation efficiencies based on the relative retention times and peak areas of templates were studied amongst other properties. Transmission Electron Microscopy (TEM) results obtained, show the creation of heterogeneous nanocavities, regardless, the introduction of Caffeine a close structural analogue presented near-zero perfusion. This confirms the selectivity and specificity of the templated polymers despite its dual-templated nature. The STA results presented the materials as having decomposition temperatures above 250OC and a relative loss in mass of less than19% over a period within 50mins of heating. Consequent to this outcome, multi-templated systems can be fabricated to sequester specifically and selectively targeted toxicants in a mixed toxicant populated system effectively.Keywords: chitosan, dual-templated, methacrylic acid, mixed-toxicants, molecularly-imprinted-polymer
Procedia PDF Downloads 117