Search results for: watershed models
4639 A New Nonlinear State-Space Model and Its Application
Authors: Abdullah Eqal Al Mazrooei
Abstract:
In this work, a new nonlinear model will be introduced. The model is in the state-space form. The nonlinearity of this model is in the state equation where the state vector is multiplied by its self. This technique makes our model generalizes many famous models as Lotka-Volterra model and Lorenz model which have many applications in the real life. We will apply our new model to estimate the wind speed by using a new nonlinear estimator which suitable to work with our model.Keywords: nonlinear systems, state-space model, Kronecker product, nonlinear estimator
Procedia PDF Downloads 6914638 Anticipating the Change: Visions and Perspectives towards a Post-Car World
Authors: Farzaneh Bahrami
Abstract:
Different indicators, such as modal shares in mobility practices or car ownership, may suggest that the century of car dominance - at least in Europe and North America - is already behind us. If the emergence of the car had radical spatial and social consequences, what would be the implications of its gradual disappearance - which could be expected in the context of ecological consciousness, economic and energetic constraints as a result of both urban policies as well as lifestyle choices? To what extend shall urban experts account for this limited but visible transition from car-dominated systems towards alternative models of mobility in which the individual-motorized mobility (car) is not central; what models of urbanity could be imagined to support such a transformation? We have examined a selection of projects at different scales and within different contexts - new planned cities, dense urban areas or territories of dispersion – whose visions involve a significant shift from the current car system. We have been looking into their tools, strategies and different measures of car reduction, as well as their varied approaches to public space as an inevitable corollary to this change. The car’s dominance was formerly questioned by advocates of public space, rather than through interests in ecological urban design or other urban planning concerns. In the 60s already a universal longing for the qualities of traditional urban space led to a critique of the proliferation of fast roads, and thus the car’s colonization of everyday life. Reclamation of public space as the city’s quintessential social territory reappears today in contemporary discourses and reinforces the shift-provoking trends towards a new urbanity freed from car dominance. In a hypothetical process of the progressive phasing-out of the car, we shall expect fundamental transformations in spatial practices of the city, accompanied by the physical configuration of its public spaces. What will be the main characteristics of the new emerging spaces of sociability and where shall we encounter them? This contribution is an ongoing research within the framework of Post-Car World, an interdisciplinary project that explores the future of mobility through the role of the car.Keywords: mobility, urbanity, future visions, public space
Procedia PDF Downloads 3704637 Assimilating Multi-Mission Satellites Data into a Hydrological Model
Authors: Mehdi Khaki, Ehsan Forootan, Joseph Awange, Michael Kuhn
Abstract:
Terrestrial water storage, as a source of freshwater, plays an important role in human lives. Hydrological models offer important tools for simulating and predicting water storages at global and regional scales. However, their comparisons with 'reality' are imperfect mainly due to a high level of uncertainty in input data and limitations in accounting for all complex water cycle processes, uncertainties of (unknown) empirical model parameters, as well as the absence of high resolution (both spatially and temporally) data. Data assimilation can mitigate this drawback by incorporating new sets of observations into models. In this effort, we use multi-mission satellite-derived remotely sensed observations to improve the performance of World-Wide Water Resources Assessment system (W3RA) hydrological model for estimating terrestrial water storages. For this purpose, we assimilate total water storage (TWS) data from the Gravity Recovery And Climate Experiment (GRACE) and surface soil moisture data from the Advanced Microwave Scanning Radiometer for the Earth Observing System (AMSR-E) into W3RA. This is done to (i) improve model estimations of water stored in ground and soil moisture, and (ii) assess the impacts of each satellite of data (from GRACE and AMSR-E) and their combination on the final terrestrial water storage estimations. These data are assimilated into W3RA using the Ensemble Square-Root Filter (EnSRF) filtering technique over Mississippi Basin (the United States) and Murray-Darling Basin (Australia) between 2002 and 2013. In order to evaluate the results, independent ground-based groundwater and soil moisture measurements within each basin are used.Keywords: data assimilation, GRACE, AMSR-E, hydrological model, EnSRF
Procedia PDF Downloads 2894636 The Effect of Metformin in Combination with Dexamethasone on the CXCR4 Level in Multiple Myeloma Cell Line
Authors: Seyede Sanaz Seyedebrahimi, Shima Rahimi, Shohreh Fakhari, Ali Jalili
Abstract:
Background: CXCR4, as a chemokine receptor, plays well-known roles in various types of cancers. Several studies have been conducted to overcome CXCR4 axis acts in multiple myeloma (MM) pathogenesis and progression. Dexamethasone, a standard treatment for multiple myeloma, has been shown to increase CXCR4 levels in multiple myeloma cell lines. Herein, we focused on the effects of metformin and dexamethasone on CXCR4 at the cellular level and the migration rate of cell lines after exposure to a combination compared to single-agent models. Materials and Method: Multiple myeloma cell lines (U266 and RPMI8226) were cultured with different metformin and dexamethasone concentrations in single-agent and combination models. The simultaneous combination doses were calculated by CompuSyn software. Cell surface and mRNA expression of CXCR4 were determined using flow cytometry and the quantitative reverse transcription-polymerase chain reaction (qRT-PCR) assay, respectively. The Transwell cell migration assay evaluated the migration ability. Results: In concurred with previous studies, our results showed a dexamethasone up-regulation effect on CXCR4 in a dose-dependent manner. Although, the metformin single-agent model could reduce CXCR4 expression of U266 and RPMI8226 in cell surface and mRNA expression level. Moreover, the administration of metformin and dexamethasone simultaneously exerted a higher suppression effect on CXCR4 expression than the metformin single-agent model. The migration rate through the combination model's matrigel membrane was remarkably lower than the metformin and dexamethasone single-agent model. Discussion: According to our findings, the combination of metformin and dexamethasone effectively inhibited dexamethasone-induced CXCR4 expression in multiple myeloma cell lines. As a result, metformin may be counted as an alternative medicine combined with other chemotherapies to combat multiple myeloma. However, more research is required.Keywords: CXCR4, dexamethasone, metformin, migration, multiple myeloma
Procedia PDF Downloads 1564635 Determination of the Axial-Vector from an Extended Linear Sigma Model
Authors: Tarek Sayed Taha Ali
Abstract:
The dependence of the axial-vector coupling constant gA on the quark masses has been investigated in the frame work of the extended linear sigma model. The field equations have been solved in the mean-field approximation. Our study shows a better fitting to the experimental data compared with the existing models.Keywords: extended linear sigma model, nucleon properties, axial coupling constant, physic
Procedia PDF Downloads 4454634 Experimental Study Damage in a Composite Structure by Vibration Analysis- Glass / Polyester
Authors: R. Abdeldjebar, B. Labbaci, L. Missoum, B. Moudden, M. Djermane
Abstract:
The basic components of a composite material made him very sensitive to damage, which requires techniques for detecting damage reliable and efficient. This work focuses on the detection of damage by vibration analysis, whose main objective is to exploit the dynamic response of a structure to detect understand the damage. The experimental results are compared with those predicted by numerical models to confirm the effectiveness of the approach.Keywords: experimental, composite, vibration analysis, damage
Procedia PDF Downloads 6744633 Pomegranate Attenuated Levodopa-Induced Dyskinesia and Dopaminergic Degeneration in MPTP Mice Models of Parkinson’s Disease
Authors: Mahsa Hadipour Jahromy, Sara Rezaii
Abstract:
Parkinson’s disease (PD) results primarily from the death of dopaminergic neurons in the substantia nigra. Soon after the discovery of levodopa and its beneficial effects in chronic administration, debilitating involuntary movements observed, termed levodopa-induced dyskinesia (LID) with poorly understood pathogenesis. Polyphenol-rich compounds, like pomegranate, provided neuroprotection in several animal models of brain diseases. In the present work, we investigated whether pomegranate has preventive effects following 4-phenyl-1,2,3,6-tetrahydropyridine (MPTP)-induced dopaminergic degenerations and the potential to diminish LID in mice. Mice model of PD was induced by MPTP (30 mg/kg daily for five consecutive days). To induce a mice model of LID, valid PD mice were treated with levodopa (50 mg/kg, i.p) for 15 days. Then the effects of chronic co-administration of pomegranate juice (20 ml/kg) with levodopa and continuing for 10 days, evaluated. Behavioural tests were performed in all groups, every other day including: Abnormal involuntary movements (AIMS), forelimb adjusting steps, cylinder, and catatonia tests. Finally, brain tissue sections were prepared to study substantia nigra changes and dopamine neuron density after treatments. With this MPTP regimen, significant movement disorders revealed in AIMS tests and there was a reduction in dopamine striatal density. Levodopa attenuates their loss caused by MPTP, however, in chronic administration, dyskinesia observed in forelimb adjusting step and cylinder tests. Besides, catatonia observed in some cases. Chronic pomegranate co-administration significantly improved LID in both tests and reduced dopaminergic loss in substantia nigra. These data indicate that pomegranate might be a good adjunct for preserving dopaminergic neurons in the substantia nigra and reducing LID in mice.Keywords: levodopa-induced dyskinesia, MPTP, Parkinson’s disease, pomegranate
Procedia PDF Downloads 4924632 D-Wave Quantum Computing Ising Model: A Case Study for Forecasting of Heat Waves
Authors: Dmytro Zubov, Francesco Volponi
Abstract:
In this paper, D-Wave quantum computing Ising model is used for the forecasting of positive extremes of daily mean air temperature. Forecast models are designed with two to five qubits, which represent 2-, 3-, 4-, and 5-day historical data respectively. Ising model’s real-valued weights and dimensionless coefficients are calculated using daily mean air temperatures from 119 places around the world, as well as sea level (Aburatsu, Japan). In comparison with current methods, this approach is better suited to predict heat wave values because it does not require the estimation of a probability distribution from scarce observations. Proposed forecast quantum computing algorithm is simulated based on traditional computer architecture and combinatorial optimization of Ising model parameters for the Ronald Reagan Washington National Airport dataset with 1-day lead-time on learning sample (1975-2010 yr). Analysis of the forecast accuracy (ratio of successful predictions to total number of predictions) on the validation sample (2011-2014 yr) shows that Ising model with three qubits has 100 % accuracy, which is quite significant as compared to other methods. However, number of identified heat waves is small (only one out of nineteen in this case). Other models with 2, 4, and 5 qubits have 20 %, 3.8 %, and 3.8 % accuracy respectively. Presented three-qubit forecast model is applied for prediction of heat waves at other five locations: Aurel Vlaicu, Romania – accuracy is 28.6 %; Bratislava, Slovakia – accuracy is 21.7 %; Brussels, Belgium – accuracy is 33.3 %; Sofia, Bulgaria – accuracy is 50 %; Akhisar, Turkey – accuracy is 21.4 %. These predictions are not ideal, but not zeros. They can be used independently or together with other predictions generated by different method(s). The loss of human life, as well as environmental, economic, and material damage, from extreme air temperatures could be reduced if some of heat waves are predicted. Even a small success rate implies a large socio-economic benefit.Keywords: heat wave, D-wave, forecast, Ising model, quantum computing
Procedia PDF Downloads 5004631 Buy-and-Hold versus Alternative Strategies: A Comparison of Market-Timing Techniques
Authors: Jonathan J. Burson
Abstract:
With the rise of virtually costless, mobile-based trading platforms, stock market trading activity has increased significantly over the past decade, particularly for the millennial generation. This increased stock market attention, combined with the recent market turmoil due to the economic upset caused by COVID-19, make the topics of market-timing and forecasting particularly relevant. While the overall stock market saw an unprecedented, historically-long bull market from March 2009 to February 2020, the end of that bull market reignited a search by investors for a way to reduce risk and increase return. Similar searches for outperformance occurred in the early, and late 2000’s as the Dotcom bubble burst and the Great Recession led to years of negative returns for mean-variance, index investors. Extensive research has been conducted on fundamental analysis, technical analysis, macroeconomic indicators, microeconomic indicators, and other techniques—all using different methodologies and investment periods—in pursuit of higher returns with lower risk. The enormous variety of timeframes, data, and methodologies used by the diverse forecasting methods makes it difficult to compare the outcome of each method directly to other methods. This paper establishes a process to evaluate the market-timing methods in an apples-to-apples manner based on simplicity, performance, and feasibility. Preliminary findings show that certain technical analysis models provide a higher return with lower risk when compared to the buy-and-hold method and to other market-timing strategies. Furthermore, technical analysis models tend to be easier for individual investors both in terms of acquiring the data and in analyzing it, making technical analysis-based market-timing methods the preferred choice for retail investors.Keywords: buy-and-hold, forecast, market-timing, probit, technical analysis
Procedia PDF Downloads 974630 A Study on the Construction Process and Sustainable Renewal Development of High-Rise Residential Areas in Chongqing (1978-2023)
Authors: Xiaoting Jing, Ling Huang
Abstract:
After the reform and opening up, Chongqing has formed far more high-rise residential areas than other cities in its more than 40 years of urban construction. High-rise residential areas have become one of the main modern living models in Chongqing and an important carrier reflecting the city's high quality of life. Reviewing the construction process and renewal work helps understand the characteristics of high-rise residential areas in Chongqing at different stages, clarify current development demands, and look forward to the focus of future renewal work. Based on socio-economic development and policy background, the article sorts the construction process of high-rise residential areas in Chongqing into four stages: the early experimental construction period of high-rise residential areas (1978-1996), the rapid start-up period of high-rise commodity housing construction (1997-2006), the large-scale construction period of high-rise commodity housing and public rental housing (2007-2014), and the period of renewal and renovation of high-rise residential areas and step-by-step construction of quality commodity housing (2015-present). Based on the construction demands and main construction types of each stage, the article summarizes that the construction of high-rise residential areas in Chongqing features large scale, high speed, and high density. It points out that a large number of high-rise residential areas built after 2000 will become important objects of renewal and renovation in the future. Based on existing renewal work experience, it is urgent to explore a path for sustainable renewal and development in terms of policy mechanisms, digital supervision, and renewal and renovation models, leading the high-rise living in Chongqing toward high-quality development.Keywords: high-rise residential areas, construction process, renewal and renovation, Chongqing
Procedia PDF Downloads 674629 GIS Based Spatial Modeling for Selecting New Hospital Sites Using APH, Entropy-MAUT and CRITIC-MAUT: A Study in Rural West Bengal, India
Authors: Alokananda Ghosh, Shraban Sarkar
Abstract:
The study aims to identify suitable sites for new hospitals with critical obstetric care facilities in Birbhum, one of the vulnerable and underserved districts of Eastern India, considering six main and 14 sub-criteria, using GIS-based Analytic Hierarchy Process (AHP) and Multi-Attribute Utility Theory (MAUT) approach. The criteria were identified through field surveys and previous literature. After collecting expert decisions, a pairwise comparison matrix was prepared using the Saaty scale to calculate the weights through AHP. On the contrary, objective weighting methods, i.e., Entropy and Criteria Importance through Interaction Correlation (CRITIC), were used to perform the MAUT. Finally, suitability maps were prepared by weighted sum analysis. Sensitivity analyses of AHP were performed to explore the effect of dominant criteria. Results from AHP reveal that ‘maternal death in transit’ followed by ‘accessibility and connectivity’, ‘maternal health care service (MHCS) coverage gap’ were three important criteria with comparatively higher weighted values. Whereas ‘accessibility and connectivity’ and ‘maternal death in transit’ were observed to have more imprint in entropy and CRITIC, respectively. While comparing the predictive suitable classes of these three models with the layer of existing hospitals, except Entropy-MAUT, the other two are pointing towards the left-over underserved areas of existing facilities. Only 43%-67% of existing hospitals were in the moderate to lower suitable class. Therefore, the results of the predictive models might bring valuable input in future planning.Keywords: hospital site suitability, analytic hierarchy process, multi-attribute utility theory, entropy, criteria importance through interaction correlation, multi-criteria decision analysis
Procedia PDF Downloads 674628 A New Model for Production Forecasting in ERP
Authors: S. F. Wong, W. I. Ho, B. Lin, Q. Huang
Abstract:
ERP has been used in many enterprises for management, the accuracy of the production forecasting module is vital to the decision making of the enterprise, and the profit is affected directly. Therefore, enhancing the accuracy of the production forecasting module can also increase the efficiency and profitability. To deal with a lot of data, a suitable, reliable and accurate statistics model is necessary. LSSVM and Grey System are two main models to be studied in this paper, and a case study is used to demonstrate how the combination model is effective to the result of forecasting.Keywords: ERP, grey system, LSSVM, production forecasting
Procedia PDF Downloads 4634627 A Hybrid Model of Structural Equation Modelling-Artificial Neural Networks: Prediction of Influential Factors on Eating Behaviors
Authors: Maryam Kheirollahpour, Mahmoud Danaee, Amir Faisal Merican, Asma Ahmad Shariff
Abstract:
Background: The presence of nonlinearity among the risk factors of eating behavior causes a bias in the prediction models. The accuracy of estimation of eating behaviors risk factors in the primary prevention of obesity has been established. Objective: The aim of this study was to explore the potential of a hybrid model of structural equation modeling (SEM) and Artificial Neural Networks (ANN) to predict eating behaviors. Methods: The Partial Least Square-SEM (PLS-SEM) and a hybrid model (SEM-Artificial Neural Networks (SEM-ANN)) were applied to evaluate the factors affecting eating behavior patterns among university students. 340 university students participated in this study. The PLS-SEM analysis was used to check the effect of emotional eating scale (EES), body shape concern (BSC), and body appreciation scale (BAS) on different categories of eating behavior patterns (EBP). Then, the hybrid model was conducted using multilayer perceptron (MLP) with feedforward network topology. Moreover, Levenberg-Marquardt, which is a supervised learning model, was applied as a learning method for MLP training. The Tangent/sigmoid function was used for the input layer while the linear function applied for the output layer. The coefficient of determination (R²) and mean square error (MSE) was calculated. Results: It was proved that the hybrid model was superior to PLS-SEM methods. Using hybrid model, the optimal network happened at MPLP 3-17-8, while the R² of the model was increased by 27%, while, the MSE was decreased by 9.6%. Moreover, it was found that which one of these factors have significantly affected on healthy and unhealthy eating behavior patterns. The p-value was reported to be less than 0.01 for most of the paths. Conclusion/Importance: Thus, a hybrid approach could be suggested as a significant methodological contribution from a statistical standpoint, and it can be implemented as software to be able to predict models with the highest accuracy.Keywords: hybrid model, structural equation modeling, artificial neural networks, eating behavior patterns
Procedia PDF Downloads 1564626 Arabic Light Word Analyser: Roles with Deep Learning Approach
Authors: Mohammed Abu Shquier
Abstract:
This paper introduces a word segmentation method using the novel BP-LSTM-CRF architecture for processing semantic output training. The objective of web morphological analysis tools is to link a formal morpho-syntactic description to a lemma, along with morpho-syntactic information, a vocalized form, a vocalized analysis with morpho-syntactic information, and a list of paradigms. A key objective is to continuously enhance the proposed system through an inductive learning approach that considers semantic influences. The system is currently under construction and development based on data-driven learning. To evaluate the tool, an experiment on homograph analysis was conducted. The tool also encompasses the assumption of deep binary segmentation hypotheses, the arbitrary choice of trigram or n-gram continuation probabilities, language limitations, and morphology for both Modern Standard Arabic (MSA) and Dialectal Arabic (DA), which provide justification for updating this system. Most Arabic word analysis systems are based on the phonotactic morpho-syntactic analysis of a word transmitted using lexical rules, which are mainly used in MENA language technology tools, without taking into account contextual or semantic morphological implications. Therefore, it is necessary to have an automatic analysis tool taking into account the word sense and not only the morpho-syntactic category. Moreover, they are also based on statistical/stochastic models. These stochastic models, such as HMMs, have shown their effectiveness in different NLP applications: part-of-speech tagging, machine translation, speech recognition, etc. As an extension, we focus on language modeling using Recurrent Neural Network (RNN); given that morphological analysis coverage was very low in dialectal Arabic, it is significantly important to investigate deeply how the dialect data influence the accuracy of these approaches by developing dialectal morphological processing tools to show that dialectal variability can support to improve analysis.Keywords: NLP, DL, ML, analyser, MSA, RNN, CNN
Procedia PDF Downloads 424625 An Overview of Bioinformatics Methods to Detect Novel Riboswitches Highlighting the Importance of Structure Consideration
Authors: Danny Barash
Abstract:
Riboswitches are RNA genetic control elements that were originally discovered in bacteria and provide a unique mechanism of gene regulation. They work without the participation of proteins and are believed to represent ancient regulatory systems in the evolutionary timescale. One of the biggest challenges in riboswitch research is that many are found in prokaryotes but only a small percentage of known riboswitches have been found in certain eukaryotic organisms. The few examples of eukaryotic riboswitches were identified using sequence-based bioinformatics search methods that include some slight structural considerations. These pattern-matching methods were the first ones to be applied for the purpose of riboswitch detection and they can also be programmed very efficiently using a data structure called affix arrays, making them suitable for genome-wide searches of riboswitch patterns. However, they are limited by their ability to detect harder to find riboswitches that deviate from the known patterns. Several methods have been developed since then to tackle this problem. The most commonly used by practitioners is Infernal that relies on Hidden Markov Models (HMMs) and Covariance Models (CMs). Profile Hidden Markov Models were also carried out in the pHMM Riboswitch Scanner web application, independently from Infernal. Other computational approaches that have been developed include RMDetect by the use of 3D structural modules and RNAbor that utilizes Boltzmann probability of structural neighbors. We have tried to incorporate more sophisticated secondary structure considerations based on RNA folding prediction using several strategies. The first idea was to utilize window-based methods in conjunction with folding predictions by energy minimization. The moving window approach is heavily geared towards secondary structure consideration relative to sequence that is treated as a constraint. However, the method cannot be used genome-wide due to its high cost because each folding prediction by energy minimization in the moving window is computationally expensive, enabling to scan only at the vicinity of genes of interest. The second idea was to remedy the inefficiency of the previous approach by constructing a pipeline that consists of inverse RNA folding considering RNA secondary structure, followed by a BLAST search that is sequence-based and highly efficient. This approach, which relies on inverse RNA folding in general and our own in-house fragment-based inverse RNA folding program called RNAfbinv in particular, shows capability to find attractive candidates that are missed by Infernal and other standard methods being used for riboswitch detection. We demonstrate attractive candidates found by both the moving-window approach and the inverse RNA folding approach performed together with BLAST. We conclude that structure-based methods like the two strategies outlined above hold considerable promise in detecting riboswitches and other conserved RNAs of functional importance in a variety of organisms.Keywords: riboswitches, RNA folding prediction, RNA structure, structure-based methods
Procedia PDF Downloads 2344624 Smart Cities, Morphology of the Uncertain: A Study on Development Processes Applied by Amazonian Cities in Ecuador
Authors: Leonardo Coloma
Abstract:
The world changes constantly, every second its properties vary due either natural factors or human intervention. As the most intelligent creatures on the planet, human beings have transformed the environment and paradoxically –have allowed ‘mother nature’ to lose species, accelerate the processes of climate change, the deterioration of the ozone layer, among others. The rapid population growth, the procurement, administration and distribution of resources, waste management, and technological advances are some of the factors that boost urban sprawl whose gray stain extends over the territory, facing challenges such as pollution, overpopulation and scarcity of resources. In Ecuador, these problems are added to the social, cultural, economic and political anomalies that have historically affected it. This fact can represent a greater delay when trying to solve global problems, without having paid attention to local inconveniences –smaller ones, but ones that could be the key to project smart solutions on bigger ones. This research aims to highlight the main characteristics of the development models adopted by two Amazonian cities, and analyze the impact of such urban growth on society; to finally define the parameters that would allow the development of an intelligent city in Ecuador, prepared for the challenges of the XXI Century. Contrasts in the climate, temperature, and landscape of Ecuadorian cities are fused with the cultural diversity of its people, generating a multiplicity of nuances of an indecipherable wealth. However, we strive to apply development models that do not recognize that wealth, not understanding them and ignoring that their proposals will vary according to where they are applied. Urban plans seem to take a bit of each of the new theories and proposals of development, which, in the encounter with the informal growth of cities, with those excluded and ‘isolated’ societies, generate absurd morphologies - where the uncertain becomes tangible. The desire to project smart cities is ever growing, but it is important to consider that this concept does not only have to do with the use of information and communication technologies. Its success is achieved when advances in science and technology allow the establishment of a better relationship between people and their context (natural and built). As a research methodology, urban analysis through mappings, diagrams and geographical studies, as well as the identification of sensorial elements when living the city, will make evident the shortcomings of the urban models adopted by certain populations of the Ecuadorian Amazon. Following the vision of previous investigations started since 2014 as part of ‘Centro de Acciones Urbanas,’ the results of this study will encourage the dialogue between the city (as a physical fact) and those who ‘make the city’ (people as its main actors). This research will allow the development of workshops and meetings with different professionals, organizations and individuals in general.Keywords: Latin American cities, smart cities, urban development, urban morphology, urban sprawl
Procedia PDF Downloads 1574623 Geometric Design to Improve the Temperature
Authors: H. Ghodbane, A. A. Taleb, O. Kraa
Abstract:
This paper presents geometric design of induction heating system. The objective of this design is to improve the temperature distribution in the load. The study of such a device requires the use of models or modeling representation, physical, mathematical, and numerical. This modeling is the basis of the understanding, the design, and optimization of these systems. The optimization technique is to find values of variables that maximize or minimize the objective function.Keywords: optimization, modeling, geometric design system, temperature increase
Procedia PDF Downloads 5304622 Seasonal Assessment of Snow Cover Dynamics Based on Aerospace Multispectral Data on Livingston Island, South Shetland Islands in Antarctica and on Svalbard in Arctic
Authors: Temenuzhka Spasova, Nadya Yanakieva
Abstract:
Snow modulates the hydrological cycle and influences the functioning of ecosystems and is a significant resource for many populations whose water is harvested from cold regions. Snow observations are important for validating climate models. The accumulation and rapid melt of snow are two of the most dynamical seasonal environmental changes on the Earth’s surface. The actuality of this research is related to the modern tendencies of the remote sensing application in the solution of problems of different nature in the ecological monitoring of the environment. The subject of the study is the dynamic during the different seasons on Livingstone Island, South Shetland Islands in Antarctica and on Svalbard in Arctic. The objects were analyzed and mapped according to the Еuropean Space Agency data (ESA), acquired by sensors Sentinel-1 SAR (Synthetic Aperture Radar), Sentinel 2 MSI and GIS. Results have been obtained for changes in snow coverage during the summer-winter transition and its dynamics in the two hemispheres. The data used is of high time-spatial resolution, which is an advantage when looking at the snow cover. The MSI images are with different spatial resolution at the Earth surface range. The changes of the environmental objects are shown with the SAR images and different processing approaches. The results clearly show that snow and snow melting can be best registered by using SAR data via hh- horizontal polarization. The effect of the researcher on aerospace data and technology enables us to obtain different digital models, structuring and analyzing results excluding the subjective factor. Because of the large extent of terrestrial snow coverage and the difficulties in obtaining ground measurements over cold regions, remote sensing and GIS represent an important tool for studying snow areas and properties from regional to global scales.Keywords: climate changes, GIS, remote sensing, SAR images, snow coverage
Procedia PDF Downloads 2194621 Centrifuge Modelling Approach on Sysmic Loading Analysis of Clay: A Geotechnical Study
Authors: Anthony Quansah, Tresor Ntaryamira, Shula Mushota
Abstract:
Models for geotechnical centrifuge testing are usually made from re-formed soil, allowing for comparisons with naturally occurring soil deposits. However, there is a fundamental omission in this process because the natural soil is deposited in layers creating a unique structure. Nonlinear dynamics of clay material deposit is an essential part of changing the attributes of ground movements when subjected to solid seismic loading, particularly when diverse intensification conduct of speeding up and relocation are considered. The paper portrays a review of axis shaking table tests and numerical recreations to explore the offshore clay deposits subjected to seismic loadings. These perceptions are accurately reenacted by DEEPSOIL with appropriate soil models and parameters reviewed from noteworthy centrifuge modeling researches. At that point, precise 1-D site reaction investigations are performed on both time and recurrence spaces. The outcomes uncover that for profound delicate clay is subjected to expansive quakes, noteworthy increasing speed lessening may happen close to the highest point of store because of soil nonlinearity and even neighborhood shear disappointment; nonetheless, huge enhancement of removal at low frequencies are normal in any case the forces of base movements, which proposes that for dislodging touchy seaward establishments and structures, such intensified low-recurrence relocation reaction will assume an essential part in seismic outline. This research shows centrifuge as a tool for creating a layered sample important for modelling true soil behaviour (such as permeability) which is not identical in all directions. Currently, there are limited methods for creating layered soil samples.Keywords: seismic analysis, layered modeling, terotechnology, finite element modeling
Procedia PDF Downloads 1554620 Innovations for Freight Transport Systems
Authors: M. Lu
Abstract:
The paper presents part of the results of EU-funded projects: SoCool@EU (Sustainable Organisation between Clusters Of Optimized Logistics @ Europe), DG-RTD (Research and Innovation), Regions of Knowledge Programme (FP7-REGIONS-2011-1). It will provide an in-depth review of emerging technologies for further improving urban mobility and freight transport systems, such as (information and physical) infrastructure, ICT-based Intelligent Transport Systems (ITS), vehicles, advanced logistics, and services. Furthermore, the paper will provide an analysis of the barriers and will review business models for the market uptake of innovations. From a perspective of science and technology, the challenges of urbanization could be mainly handled through adequate (human-oriented) solutions for urban planning, sustainable energy, the water system, building design and construction, the urban transport system (both physical and information aspects), and advanced logistics and services. Implementation of solutions for these domains should be follow a highly integrated and balanced approach, a silo approach should be avoided. To develop a sustainable urban transport system (for people and goods), including inter-hubs and intra-hubs, a holistic view is needed. To achieve a sustainable transport system for people and goods (in terms of cost-effectiveness, efficiency, environment-friendliness and fulfillment of the mobility, transport and logistics needs of the society), a proper network and information infrastructure, advanced transport systems and operations, as well as ad hoc and seamless services are required. In addition, a road map for an enhanced urban transport system until 2050 will be presented. This road map aims to address the challenges of urban transport, and to provide best practices in inter-city and intra-city environments from various perspectives, including policy, traveler behaviour, economy, liability, business models, and technology.Keywords: synchromodality, multimodal transport, logistics, Intelligent Transport Systems (ITS)
Procedia PDF Downloads 3164619 Inverse Matrix in the Theory of Dynamical Systems
Authors: Renata Masarova, Bohuslava Juhasova, Martin Juhas, Zuzana Sutova
Abstract:
In dynamic system theory a mathematical model is often used to describe their properties. In order to find a transfer matrix of a dynamic system we need to calculate an inverse matrix. The paper contains the fusion of the classical theory and the procedures used in the theory of automated control for calculating the inverse matrix. The final part of the paper models the given problem by the Matlab.Keywords: dynamic system, transfer matrix, inverse matrix, modeling
Procedia PDF Downloads 5164618 Scenario Analysis to Assess the Competitiveness of Hydrogen in Securing the Italian Energy System
Authors: Gianvito Colucci, Valeria Di Cosmo, Matteo Nicoli, Orsola Maria Robasto, Laura Savoldi
Abstract:
The hydrogen value chain deployment is likely to be boosted in the near term by the energy security measures planned by European countries to face the recent energy crisis. In this context, some countries are recognized to have a crucial role in the geopolitics of hydrogen as importers, consumers and exporters. According to the European Hydrogen Backbone Initiative, Italy would be part of one of the 5 corridors that will shape the European hydrogen market. However, the set targets are very ambitious and require large investments to rapidly develop effective hydrogen policies: in this regard, scenario analysis is becoming increasingly important to support energy planning, and energy system optimization models appear to be suitable tools to quantitively carry on that kind of analysis. The work aims to assess the competitiveness of hydrogen in contributing to the Italian energy security in the coming years, under different price and import conditions, using the energy system model TEMOA-Italy. A wide spectrum of hydrogen technologies is included in the analysis, covering the production, storage, delivery, and end-uses stages. National production from fossil fuels with and without CCS, as well as electrolysis and import of low-carbon hydrogen from North Africa, are the supply solutions that would compete with other ones, such as natural gas, biomethane and electricity value chains, to satisfy sectoral energy needs (transport, industry, buildings, agriculture). Scenario analysis is then used to study the competition under different price and import conditions. The use of TEMOA-Italy allows the work to catch the interaction between the economy and technological detail, which is much needed in the energy policies assessment, while the transparency of the analysis and of the results is ensured by the full accessibility of the TEMOA open-source modeling framework.Keywords: energy security, energy system optimization models, hydrogen, natural gas, open-source modeling, scenario analysis, TEMOA
Procedia PDF Downloads 1164617 Predicting Loss of Containment in Surface Pipeline using Computational Fluid Dynamics and Supervised Machine Learning Model to Improve Process Safety in Oil and Gas Operations
Authors: Muhammmad Riandhy Anindika Yudhy, Harry Patria, Ramadhani Santoso
Abstract:
Loss of containment is the primary hazard that process safety management is concerned within the oil and gas industry. Escalation to more serious consequences all begins with the loss of containment, starting with oil and gas release from leakage or spillage from primary containment resulting in pool fire, jet fire and even explosion when reacted with various ignition sources in the operations. Therefore, the heart of process safety management is avoiding loss of containment and mitigating its impact through the implementation of safeguards. The most effective safeguard for the case is an early detection system to alert Operations to take action prior to a potential case of loss of containment. The detection system value increases when applied to a long surface pipeline that is naturally difficult to monitor at all times and is exposed to multiple causes of loss of containment, from natural corrosion to illegal tapping. Based on prior researches and studies, detecting loss of containment accurately in the surface pipeline is difficult. The trade-off between cost-effectiveness and high accuracy has been the main issue when selecting the traditional detection method. The current best-performing method, Real-Time Transient Model (RTTM), requires analysis of closely positioned pressure, flow and temperature (PVT) points in the pipeline to be accurate. Having multiple adjacent PVT sensors along the pipeline is expensive, hence generally not a viable alternative from an economic standpoint.A conceptual approach to combine mathematical modeling using computational fluid dynamics and a supervised machine learning model has shown promising results to predict leakage in the pipeline. Mathematical modeling is used to generate simulation data where this data is used to train the leak detection and localization models. Mathematical models and simulation software have also been shown to provide comparable results with experimental data with very high levels of accuracy. While the supervised machine learning model requires a large training dataset for the development of accurate models, mathematical modeling has been shown to be able to generate the required datasets to justify the application of data analytics for the development of model-based leak detection systems for petroleum pipelines. This paper presents a review of key leak detection strategies for oil and gas pipelines, with a specific focus on crude oil applications, and presents the opportunities for the use of data analytics tools and mathematical modeling for the development of robust real-time leak detection and localization system for surface pipelines. A case study is also presented.Keywords: pipeline, leakage, detection, AI
Procedia PDF Downloads 1914616 Enhancement of Aircraft Longitudinal Stability Using Tubercles
Authors: Muhammad Umer, Aishwariya Giri, Umaiyma Rakha
Abstract:
Mimicked from the humpback whale flippers, the application of tubercle technology is seen to be particularly advantageous at high angles of attack. This particular advantage is of paramount importance when it comes to structures producing lift at high angles of attack. This characteristic of the technology makes it ideal for horizontal stabilizers and selecting the same as the subject of study to identify and exploit the advantage highlighted by researchers on airfoils, this project aims in establishing a foundation for the application of the bio-mimicked technology on an existing aircraft. Using a baseline and 2 tubercle configuration integrated models, the project targets to achieve the twin aim of highlighting the possibility and merits over the base model and also choosing the right configuration in providing the best characteristic suitable for high angles of attack. To facilitate this study, the required models are generated using Solidworks followed by trials in a virtual aerodynamic environment using Fluent in Ansys for resolving the project objectives. Following a structured plan, the aim is to initially identify the advantages mathematically and then selecting the optimal configuration, simulate the end configuration at angles mimicking the actual operation envelope for the particular structure. Upon simulating the baseline configuration at various angles of attack, the stall angle was determined to be 22 degrees. Thus, the tubercle configurations will be simulated and compared at 4 different angles of attacks: 0, 10, 20, and 24. Further, after providing the optimum configuration of horizontal stabilizers, this study aims at the integration of aircraft structure so that the results better imply the end deliverables of real life application. This draws the project scope closer at this point into longitudinal static stability considerations and improvements in the manoeuvrability characteristics. The objective of the study is to achieve a complete overview ready for real life application with marked benefits obtainable from bio morphing of the tubercle technology.Keywords: flow simulation, horizontal stabilizer, stability enhancement, tubercle
Procedia PDF Downloads 3204615 The Use of the TRIGRS Model and Geophysics Methodologies to Identify Landslides Susceptible Areas: Case Study of Campos do Jordao-SP, Brazil
Authors: Tehrrie Konig, Cassiano Bortolozo, Daniel Metodiev, Rodolfo Mendes, Marcio Andrade, Marcio Moraes
Abstract:
Gravitational mass movements are recurrent events in Brazil, usually triggered by intense rainfall. When these events occur in urban areas, they end up becoming disasters due to the economic damage, social impact, and loss of human life. To identify the landslide-susceptible areas, it is important to know the geotechnical parameters of the soil, such as cohesion, internal friction angle, unit weight, hydraulic conductivity, and hydraulic diffusivity. The measurement of these parameters is made by collecting soil samples to analyze in the laboratory and by using geophysical methodologies, such as Vertical Electrical Survey (VES). The geophysical surveys analyze the soil properties with minimal impact in its initial structure. Statistical analysis and mathematical models of physical basis are used to model and calculate the Factor of Safety for steep slope areas. In general, such mathematical models work from the combination of slope stability models and hydrological models. One example is the mathematical model TRIGRS (Transient Rainfall Infiltration and Grid-based Regional Slope- Stability Model) which calculates the variation of the Factor of Safety of a determined study area. The model relies on changes in pore-pressure and soil moisture during a rainfall event. TRIGRS was written in the Fortran programming language and associates the hydrological model, which is based on the Richards Equation, with the stability model based on the principle of equilibrium limit. Therefore, the aims of this work are modeling the slope stability of Campos do Jordão with TRIGRS, using geotechnical and geophysical methodologies to acquire the soil properties. The study area is located at southern-east of Sao Paulo State in the Mantiqueira Mountains and has a historic landslide register. During the fieldwork, soil samples were collected, and the VES method applied. These procedures provide the soil properties, which were used as input data in the TRIGRS model. The hydrological data (infiltration rate and initial water table height) and rainfall duration and intensity, were acquired from the eight rain gauges installed by Cemaden in the study area. A very high spatial resolution digital terrain model was used to identify the slopes declivity. The analyzed period is from March 6th to March 8th of 2017. As results, the TRIGRS model calculates the variation of the Factor of Safety within a 72-hour period in which two heavy rainfall events stroke the area and six landslides were registered. After each rainfall, the Factor of Safety declined, as expected. The landslides happened in areas identified by the model with low values of Factor of Safety, proving its efficiency on the identification of landslides susceptible areas. This study presents a critical threshold for landslides, in which an accumulated rainfall higher than 80mm/m² in 72 hours might trigger landslides in urban and natural slopes. The geotechnical and geophysics methods are shown to be very useful to identify the soil properties and provide the geological characteristics of the area. Therefore, the combine geotechnical and geophysical methods for soil characterization and the modeling of landslides susceptible areas with TRIGRS are useful for urban planning. Furthermore, early warning systems can be developed by combining the TRIGRS model and weather forecast, to prevent disasters in urban slopes.Keywords: landslides, susceptibility, TRIGRS, vertical electrical survey
Procedia PDF Downloads 1734614 Removal of Cr (VI) from Water through Adsorption Process Using GO/PVA as Nanosorbent
Authors: Syed Hadi Hasan, Devendra Kumar Singh, Viyaj Kumar
Abstract:
Cr (VI) is a known toxic heavy metal and has been considered as a priority pollutant in water. The effluent of various industries including electroplating, anodizing baths, leather tanning, steel industries and chromium based catalyst are the major source of Cr (VI) contamination in the aquatic environment. Cr (VI) show high mobility in the environment and can easily penetrate cell membrane of the living tissues to exert noxious effects. The Cr (VI) contamination in drinking water causes various hazardous health effects to the human health such as cancer, skin and stomach irritation or ulceration, dermatitis, damage to liver, kidney circulation and nerve tissue damage. Herein, an attempt has been done to develop an efficient adsorbent for the removal of Cr (VI) from water. For this purpose nanosorbent composed of polyvinyl alcohol functionalized graphene oxide (GO/PVA) was prepared. Thus, obtained GO/PVA was characterized through FTIR, XRD, SEM, and Raman Spectroscopy. As prepared nanosorbent of GO/PVA was utilized for the removal Cr (VI) in batch mode experiment. The process variables such as contact time, initial Cr (VI) concentration, pH, and temperature were optimized. The maximum 99.8 % removal of Cr (VI) was achieved at initial Cr (VI) concentration 60 mg/L, pH 2, temperature 35 °C and equilibrium was achieved within 50 min. The two widely used isotherm models viz. Langmuir and Freundlich were analyzed using linear correlation coefficient (R2) and it was found that Langmuir model gives best fit with high value of R2 for the data of present adsorption system which indicate the monolayer adsorption of Cr (VI) on the GO/PVA. Kinetic studies were also conducted using pseudo-first order and pseudo-second order models and it was observed that chemosorptive pseudo-second order model described the kinetics of current adsorption system in better way with high value of correlation coefficient. Thermodynamic studies were also conducted and results showed that the adsorption was spontaneous and endothermic in nature.Keywords: adsorption, GO/PVA, isotherm, kinetics, nanosorbent, thermodynamics
Procedia PDF Downloads 3894613 The Structural Behavior of Fiber Reinforced Lightweight Concrete Beams: An Analytical Approach
Authors: Jubee Varghese, Pouria Hafiz
Abstract:
Increased use of lightweight concrete in the construction industry is mainly due to its reduction in the weight of the structural elements, which in turn reduces the cost of production, transportation, and the overall project cost. However, the structural application of these lightweight concrete structures is limited due to its reduced density. Hence, further investigations are in progress to study the effect of fiber inclusion in improving the mechanical properties of lightweight concrete. Incorporating structural steel fibers, in general, enhances the performance of concrete and increases its durability by minimizing its potential to cracking and providing crack arresting mechanism. In this research, Geometric and Materially Non-linear Analysis (GMNA) was conducted for Finite Element Modelling using a software known as ABAQUS, to investigate the structural behavior of lightweight concrete with and without the addition of steel fibers and shear reinforcement. 21 finite element models of beams were created to study the effect of steel fibers based on three main parameters; fiber volume fraction (Vf = 0, 0.5 and 0.75%), shear span to depth ratio (a/d of 2, 3 and 4) and ratio of area of shear stirrups to spacing (As/s of 0.7, 1 and 1.6). The models created were validated with the previous experiment conducted by H.K. Kang et al. in 2011. It was seen that the lightweight fiber reinforcement can replace the use of fiber reinforced normal weight concrete as structural elements. The effect of an increase in steel fiber volume fraction is dominant for beams with higher shear span to depth ratio than for lower ratios. The effect of stirrups in the presence of fibers was very negligible; however; it provided extra confinement to the cracks by reducing the crack propagation and extra shear resistance than when compared to beams with no stirrups.Keywords: ABAQUS, beams, fiber-reinforced concrete, finite element, light weight, shear span-depth ratio, steel fibers, steel-fiber volume fraction
Procedia PDF Downloads 1074612 A Parallel Cellular Automaton Model of Tumor Growth for Multicore and GPU Programming
Authors: Manuel I. Capel, Antonio Tomeu, Alberto Salguero
Abstract:
Tumor growth from a transformed cancer-cell up to a clinically apparent mass spans through a range of spatial and temporal magnitudes. Through computer simulations, Cellular Automata (CA) can accurately describe the complexity of the development of tumors. Tumor development prognosis can now be made -without making patients undergo through annoying medical examinations or painful invasive procedures- if we develop appropriate CA-based software tools. In silico testing mainly refers to Computational Biology research studies of application to clinical actions in Medicine. To establish sound computer-based models of cellular behavior, certainly reduces costs and saves precious time with respect to carrying out experiments in vitro at labs or in vivo with living cells and organisms. These aim to produce scientifically relevant results compared to traditional in vitro testing, which is slow, expensive, and does not generally have acceptable reproducibility under the same conditions. For speeding up computer simulations of cellular models, specific literature shows recent proposals based on the CA approach that include advanced techniques, such the clever use of supporting efficient data structures when modeling with deterministic stochastic cellular automata. Multiparadigm and multiscale simulation of tumor dynamics is just beginning to be developed by the concerned research community. The use of stochastic cellular automata (SCA), whose parallel programming implementations are open to yield a high computational performance, are of much interest to be explored up to their computational limits. There have been some approaches based on optimizations to advance in multiparadigm models of tumor growth, which mainly pursuit to improve performance of these models through efficient memory accesses guarantee, or considering the dynamic evolution of the memory space (grids, trees,…) that holds crucial data in simulations. In our opinion, the different optimizations mentioned above are not decisive enough to achieve the high performance computing power that cell-behavior simulation programs actually need. The possibility of using multicore and GPU parallelism as a promising multiplatform and framework to develop new programming techniques to speed-up the computation time of simulations is just starting to be explored in the few last years. This paper presents a model that incorporates parallel processing, identifying the synchronization necessary for speeding up tumor growth simulations implemented in Java and C++ programming environments. The speed up improvement that specific parallel syntactic constructs, such as executors (thread pools) in Java, are studied. The new tumor growth parallel model is proved using implementations with Java and C++ languages on two different platforms: chipset Intel core i-X and a HPC cluster of processors at our university. The parallelization of Polesczuk and Enderling model (normally used by researchers in mathematical oncology) proposed here is analyzed with respect to performance gain. We intend to apply the model and overall parallelization technique presented here to solid tumors of specific affiliation such as prostate, breast, or colon. Our final objective is to set up a multiparadigm model capable of modelling angiogenesis, or the growth inhibition induced by chemotaxis, as well as the effect of therapies based on the presence of cytotoxic/cytostatic drugs.Keywords: cellular automaton, tumor growth model, simulation, multicore and manycore programming, parallel programming, high performance computing, speed up
Procedia PDF Downloads 2444611 Optimizing The Residential Design Process Using Automated Technologies
Authors: Martin Georgiev, Milena Nanova, Damyan Damov
Abstract:
Architects, engineers, and developers need to analyse and implement a wide spectrum of data in different formats, if they want to produce viable residential developments. Usually, this data comes from a number of different sources and is not well structured. The main objective of this research project is to provide parametric tools working with real geodesic data that can generate residential solutions. Various codes, regulations and design constraints are described by variables and prioritized. In this way, we establish a common workflow for architects, geodesists, and other professionals involved in the building and investment process. This collaborative medium ensures that the generated design variants conform to various requirements, contributing to a more streamlined and informed decision-making process. The quantification of distinctive characteristics inherent to typical residential structures allows a systematic evaluation of the generated variants, focusing on factors crucial to designers, such as daylight simulation, circulation analysis, space utilization, view orientation, etc. Integrating real geodesic data offers a holistic view of the built environment, enhancing the accuracy and relevance of the design solutions. The use of generative algorithms and parametric models offers high productivity and flexibility of the design variants. It can be implemented in more conventional CAD and BIM workflow. Experts from different specialties can join their efforts, sharing a common digital workspace. In conclusion, our research demonstrates that a generative parametric approach based on real geodesic data and collaborative decision-making could be introduced in the early phases of the design process. This gives the designers powerful tools to explore diverse design possibilities, significantly improving the qualities of the building investment during its entire lifecycle.Keywords: architectural design, residential buildings, urban development, geodesic data, generative design, parametric models, workflow optimization
Procedia PDF Downloads 524610 Interfacial Reactions between Aromatic Polyamide Fibers and Epoxy Matrix
Authors: Khodzhaberdi Allaberdiev
Abstract:
In order to understand the interactions on the interface polyamide fibers and epoxy matrix in fiber- reinforced composites were investigated industrial aramid fibers: armos, svm, terlon using individual epoxy matrix components, epoxies: diglycidyl ether of bisphenol A (DGEBA), three- and diglycidyl derivatives of m, p-amino-, m, p-oxy-, o, m,p-carboxybenzoic acids, the models: curing agent, aniline and the compound, that depict of the structure the primary addition reaction the amine to the epoxy resin, N-di (oxyethylphenoxy) aniline. The chemical structure of the surface of untreated and treated polyamide fibers analyzed using Fourier transform infrared spectroscopy (FTIR). The impregnation of fibers with epoxy matrix components and N-di (oxyethylphenoxy) aniline has been carried out by heating 150˚C (6h). The optimum fiber loading is at 65%.The result a thermal treatment is the covalent bonds formation , derived from a combined of homopolymerization and crosslinking mechanisms in the interfacial region between the epoxy resin and the surface of fibers. The reactivity of epoxy resins on interface in microcomposites (MC) also depends from processing aids treated on surface of fiber and the absorbance moisture. The influences these factors as evidenced by the conversion of epoxy groups values in impregnated with DGEBA of the terlons: industrial, dried (in vacuum) and purified samples: 5.20 %, 4.65% and 14.10%, respectively. The same tendency for svm and armos fibers is observed. The changes in surface composition of these MC were monitored by X-ray photoelectron spectroscopy (XPS). In the case of the purified fibers, functional groups of fibers act as well as a catalyst and curing agent of epoxy resin. It is found that the value of the epoxy groups conversion for reinforced formulations depends on aromatic polyamides nature and decreases in the order: armos >svm> terlon. This difference is due of the structural characteristics of fibers. The interfacial interactions also examined between polyglycidyl esters substituted benzoic acids and polyamide fibers in the MC. It is found that on interfacial interactions these systems influences as well as the structure and the isomerism of epoxides. The IR-spectrum impregnated fibers with aniline showed that the polyamide fibers appreciably with aniline do not react. FTIR results of treated fibers with N-di (oxyethylphenoxy) aniline fibers revealed dramatically changes IR-characteristic of the OH groups of the amino alcohol. These observations indicated hydrogen bondings and covalent interactions between amino alcohol and functional groups of fibers. This result also confirms appearance of the exo peak on Differential Scanning Calorimetry (DSC) curve of the MC. Finally, the theoretical evaluation non-covalent interactions between individual epoxy matrix components and fibers has been performed using the benzanilide and its derivative contaning the benzimidazole moiety as a models of terlon and svm,armos, respectively. Quantum-topological analysis also demonstrated the existence hydrogen bond between amide group of models and epoxy matrix components.All the results indicated that on the interface polyamide fibers and epoxy matrix exist not only covalent, but and non-covalent the interactions during the preparation of MC.Keywords: epoxies, interface, modeling, polyamide fibers
Procedia PDF Downloads 266