Search results for: evolution algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5280

Search results for: evolution algorithm

1440 Triassic Magmatism in Southern Beishan Orogen, Northwest China: Zircon U–Pb Geochronology, Petrogenesis and Tectonic Implications

Authors: Zengda Li

Abstract:

The tectonic evolution of the Beishan orogen, which forms part of the Central Asian Orogenic Belt, remains debated. This study reports the identification of three Triassic granitic plutons representing two distinct stages of magmatism in southern Beishan orogen. Zircon U–Pb dating constrains the early stage as 238–237 Ma and the late stage as 229–227 Ma. The granitoids belong to high-K calc-alkaline and shoshonitic series and exhibit alkalic-calcic and calc-alkalic features, and are weakly peraluminous rocks. Most of these granitoids are highly fractionated I-type and A-type granites. They have relatively high Isr values (0.7049–0.7086) and weak negative εNd(t) values of −1.5 to −2.1, with young Nd model ages of 1.04–0.91 Ga, indicating a crustal contribution. They also show markedly positive zircon εHf(t) values (+3.4 to +11.8) and two-stage Hf model ages of 1.06–0.69 Ga, indicating a mixture of mantle and crustal components. The lithospheric mantle beneath this region incorporating older subducted materials was metasomatized by fluids or melts. Partial melting of the metasomatized lithospheric mantle resulted in underplated magmas, which provided the heat and material input to generate the granitoids. The Middle Triassic granitic plutons show moderate negative Eu anomalies, enrichment of LILEs and depletion in Nb, Ta, and Ti suggesting partial melting of crustal components in response to the underplated mantle-derived magmas, probably linked to lithospheric delamination and asthenospheric upwelling. The Late Triassic granitic plutons show characteristics of post-orogenic granite with strong negative anomalies of Eu, Ba, Nb, Sr, P, and Ti, indicating fractional crystallization and crustal contamination during the emplacement process.

Keywords: Triassic, magmatism, geochronology, petrogenesis, Beishan orogen

Procedia PDF Downloads 143
1439 Multidisciplinary Approach for a Tsunami Reconstruction Plan in Coquimbo, Chile

Authors: Ileen Van den Berg, Reinier J. Daals, Chris E. M. Heuberger, Sven P. Hildering, Bob E. Van Maris, Carla M. Smulders, Rafael Aránguiz

Abstract:

Chile is located along the subduction zone of the Nazca plate beneath the South American plate, where large earthquakes and tsunamis have taken place throughout history. The last significant earthquake (Mw 8.2) occurred in September 2015 and generated a destructive tsunami, which mainly affected the city of Coquimbo (71.33°W, 29.96°S). The inundation area consisted of a beach, damaged seawall, damaged railway, wetland and old neighborhood; therefore, local authorities started a reconstruction process immediately after the event. Moreover, a seismic gap has been identified in the same area, and another large event could take place in the near future. The present work proposed an integrated tsunami reconstruction plan for the city of Coquimbo that considered several variables such as safety, nature & recreation, neighborhood welfare, visual obstruction, infrastructure, construction process, and durability & maintenance. Possible future tsunami scenarios are simulated by means of the Non-hydrostatic Evolution of Ocean WAVEs (NEOWAVE) model with 5 nested grids and a higher grid resolution of ~10 m. Based on the score from a multi-criteria analysis, the costs of the alternatives and a preference for a multifunctional solution, the alternative that includes an elevated coastal road with floodgates to reduce tsunami overtopping and control the return flow of a tsunami was selected as the best solution. It was also observed that the wetlands are significantly restored to their former configuration; moreover, the dynamic behavior of the wetlands is stimulated. The numerical simulation showed that the new coastal protection decreases damage and the probability of loss of life by delaying tsunami arrival time. In addition, new evacuation routes and a smaller inundation zone in the city increase safety for the area.

Keywords: tsunami, Coquimbo, Chile, reconstruction, numerical simulation

Procedia PDF Downloads 229
1438 Simulation of Flow through Dam Foundation by FEM and ANN Methods Case Study: Shahid Abbaspour Dam

Authors: Mehrdad Shahrbanozadeh, Gholam Abbas Barani, Saeed Shojaee

Abstract:

In this study, a finite element (Seep3D model) and an artificial neural network (ANN) model were developed to simulate flow through dam foundation. Seep3D model is capable of simulating three-dimensional flow through a heterogeneous and anisotropic, saturated and unsaturated porous media. Flow through the Shahid Abbaspour dam foundation has been used as a case study. The FEM with 24960 triangular elements and 28707 nodes applied to model flow through foundation of this dam. The FEM being made denser in the neighborhood of the curtain screen. The ANN model developed for Shahid Abbaspour dam is a feedforward four layer network employing the sigmoid function as an activator and the back-propagation algorithm for the network learning. The water level elevations of the upstream and downstream of the dam have been used as input variables and the piezometric heads as the target outputs in the ANN model. The two models are calibrated and verified using the Shahid Abbaspour’s dam piezometric data. Results of the models were compared with those measured by the piezometers which are in good agreement. The model results also revealed that the ANN model performed as good as and in some cases better than the FEM.

Keywords: seepage, dam foundation, finite element method, neural network, seep 3D model

Procedia PDF Downloads 457
1437 A Bio-Inspired Approach for Self-Managing Wireless Sensor and Actor Networks

Authors: Lyamine Guezouli, Kamel Barka, Zineb Seghir

Abstract:

Wireless sensor and actor networks (WSANs) present a research challenge for different practice areas. Researchers are trying to optimize the use of such networks through their research work. This optimization is done on certain criteria, such as improving energy efficiency, exploiting node heterogeneity, self-adaptability and self-configuration. In this article, we present our proposal for BIFSA (Biologically-Inspired Framework for Wireless Sensor and Actor networks). Indeed, BIFSA is a middleware that addresses the key issues of wireless sensor and actor networks. BIFSA consists of two types of agents: sensor agents (SA) that operate at the sensor level to collect and transport data to actors and actor agents (AA) that operate at the actor level to transport data to base stations. Once the sensor agent arrives at the actor, it becomes an actor agent, which can exploit the resources of the actors and vice versa. BIFSA allows agents to evolve their genetic structures and adapt to the current network conditions. The simulation results show that BIFSA allows the agents to make better use of all the resources available in each type of node, which improves the performance of the network.

Keywords: wireless sensor and actor networks, self-management, genetic algorithm, agent.

Procedia PDF Downloads 72
1436 Multi-Criteria Optimization of High-Temperature Reversed Starter-Generator

Authors: Flur R. Ismagilov, Irek Kh. Khayrullin, Vyacheslav E. Vavilov, Ruslan D. Karimov, Anton S. Gorbunov, Danis R. Farrakhov

Abstract:

The paper presents another structural scheme of high-temperature starter-generator with external rotor to be installed on High Pressure Shaft (HPS) of aircraft engines (AE) to implement More Electrical Engine concept. The basic materials to make this starter-generator (SG) were selected and justified. Multi-criteria optimization of the developed structural scheme was performed using a genetic algorithm and Pareto method. The optimum (in Pareto terms) active length and thickness of permanent magnets of SG were selected as a result of the optimization. Using the dimensions obtained, allowed to reduce the weight of the designed SG by 10 kg relative to a base option at constant thermal loads. Multidisciplinary computer simulation was performed on the basis of the optimum geometric dimensions, which proved performance efficiency of the design. We further plan to make a full-scale sample of SG of HPS and publish the results of its experimental research.

Keywords: high-temperature starter-generator, more electrical engine, multi-criteria optimization, permanent magnet

Procedia PDF Downloads 348
1435 Green Accounting and Firm Performance: A Bibliometric Literature Review

Authors: Francesca di Donato, Sara Trucco

Abstract:

Green accounting is a growing topic of interest. Indeed, nowadays, most firms affect the environment; therefore, companies are seeking the best way to disclose environmental information. Furthermore, companies are increasingly committed to improving the environment, and the topic is gaining more importance to the public, governments, and policymakers. Green accounting is a type of accounting that considers environmental costs and their impact on the financial performance of firms. Thus, the motivation of the current research is to investigate the state-of-the-art literature on the relationship between green accounting and firm performance since the birth of the topic of green accounting and to investigate gaps in the literature that represent fruitful terrain for future research. In doing so, this study provides a bibliometric literature review of existing evidence related to the link between green accounting and firm performance since 2000. The search, based on the most relevant databases for scientific journals (which are Scopus, Emerald, Web of Science, Google Scholar, and Econlit), returned 1917 scientific articles. The articles were manually reviewed in order to identify only the relevant studies in the field by excluding articles with titles and abstracts out of scope. The final sample was composed of 107 articles. A content analysis was carried out on the final sample of articles; in doing so, a classification system has been proposed. Findings show the most relevant environmental costs and issues considered in previous studies and how green accounting may be linked to the financial and non-financial performance of a firm. The study also offers suggestions for future research in this domain. This study has several practical implications. Indeed, the topic of green accounting may be applied to different sectors and different types of companies. Therefore, this study may help managers to better understand the most relevant environmental information to disclose and how environmental issues may be managed to improve the performance of the firms. Moreover, the bibliometric literature review may be of interest to those stakeholders who are interested in the historical evolution of the topic.

Keywords: bibliometric literature review, firm performance, green accounting, literature review

Procedia PDF Downloads 48
1434 Neural Network Approach to Classifying Truck Traffic

Authors: Ren Moses

Abstract:

The process of classifying vehicles on a highway is hereby viewed as a pattern recognition problem in which connectionist techniques such as artificial neural networks (ANN) can be used to assign vehicles to their correct classes and hence to establish optimum axle spacing thresholds. In the United States, vehicles are typically classified into 13 classes using a methodology commonly referred to as “Scheme F”. In this research, the ANN model was developed, trained, and applied to field data of vehicles. The data comprised of three vehicular features—axle spacing, number of axles per vehicle, and overall vehicle weight. The ANN reduced the classification error rate from 9.5 percent to 6.2 percent when compared to an existing classification algorithm that is not ANN-based and which uses two vehicular features for classification, that is, axle spacing and number of axles. The inclusion of overall vehicle weight as a third classification variable further reduced the error rate from 6.2 percent to only 3.0 percent. The promising results from the neural networks were used to set up new thresholds that reduce classification error rate.

Keywords: artificial neural networks, vehicle classification, traffic flow, traffic analysis, and highway opera-tions

Procedia PDF Downloads 297
1433 Comparative Analysis of Predictive Models for Customer Churn Prediction in the Telecommunication Industry

Authors: Deepika Christopher, Garima Anand

Abstract:

To determine the best model for churn prediction in the telecom industry, this paper compares 11 machine learning algorithms, namely Logistic Regression, Support Vector Machine, Random Forest, Decision Tree, XGBoost, LightGBM, Cat Boost, AdaBoost, Extra Trees, Deep Neural Network, and Hybrid Model (MLPClassifier). It also aims to pinpoint the top three factors that lead to customer churn and conducts customer segmentation to identify vulnerable groups. According to the data, the Logistic Regression model performs the best, with an F1 score of 0.6215, 81.76% accuracy, 68.95% precision, and 56.57% recall. The top three attributes that cause churn are found to be tenure, Internet Service Fiber optic, and Internet Service DSL; conversely, the top three models in this article that perform the best are Logistic Regression, Deep Neural Network, and AdaBoost. The K means algorithm is applied to establish and analyze four different customer clusters. This study has effectively identified customers that are at risk of churn and may be utilized to develop and execute strategies that lower customer attrition.

Keywords: attrition, retention, predictive modeling, customer segmentation, telecommunications

Procedia PDF Downloads 41
1432 A Predictive MOC Solver for Water Hammer Waves Distribution in Network

Authors: A. Bayle, F. Plouraboué

Abstract:

Water Distribution Network (WDN) still suffers from a lack of knowledge about fast pressure transient events prediction, although the latter may considerably impact their durability. Accidental or planned operating activities indeed give rise to complex pressure interactions and may drastically modified the local pressure value generating leaks and, in rare cases, pipe’s break. In this context, a numerical predictive analysis is conducted to prevent such event and optimize network management. A couple of Python/FORTRAN 90, home-made software, has been developed using Method Of Characteristic (MOC) solving for water-hammer equations. The solver is validated by direct comparison with theoretical and experimental measurement in simple configurations whilst afterward extended to network analysis. The algorithm's most costly steps are designed for parallel computation. A various set of boundary conditions and energetic losses models are considered for the network simulations. The results are analyzed in both real and frequencies domain and provide crucial information on the pressure distribution behavior within the network.

Keywords: energetic losses models, method of characteristic, numerical predictive analysis, water distribution network, water hammer

Procedia PDF Downloads 211
1431 Harmonization of Financial Information Systems in Latin America in Light of International Public Sector Accounting Standards Using the Herfindahl-Hirschman Index

Authors: Laura Sour

Abstract:

Government accounting is an essential instrument of transparency and accountability in public administration, which allows connecting internal management with the implementation of policies and their evaluation by third parties through the construction of indicators on the cost of government. Several countries have adopted the International Public Sector Accounting Standards (IPSAS) as part of their modernization strategy. This document will evaluate the quantity and harmonization of the financial information published in the financial statements of 12 Latin American countries based on what is established in IPSAS 1, 2 and 17. For this, seven types of financial statements are analyzed. published during the period from 2015 to 2019. Based on this information, it will be possible to describe the evolution in the government financial publication to carry out a detailed analysis of the items that have been most transparent in these countries. Finally, the level of harmonization of the financial statements will be studied using the Herfindahl-Hirschman index (IHH) to determine the degree of comparability of the information. To date, the results indicate that the public sector has increased the quantity and harmonization of the financial information published during the study period, but in a heterogeneous way: From the data collected, it has been found that the financial statement published with greater frequency and quantity is the Income Statement (classification of expenses by nature). On the other hand, the most complete reports were published by Costa Rica (2017 to 2019) and Mexico (2016 to 2018), periods during which these countries complied with 92.9 percent of the items analyzed. Although 2017 and 2018 are the years in which the most financial statements were reported, it is important to mention that Mexico is the country that has published the most financial information throughout the entire study period. The use of the IHH is expected to provide accurate information on the quality with which countries have adopted IPSAS within their government accounting systems to promote transparency and accountability in the continent.

Keywords: accounting and auditing, government policy and regulation, harmonization, public sector accounting and audits IPSAS

Procedia PDF Downloads 76
1430 Methodology of Automation and Supervisory Control and Data Acquisition for Restructuring Industrial Systems

Authors: Lakhoua Najeh

Abstract:

Introduction: In most situations, an industrial system already existing, conditioned by its history, its culture and its context are in difficulty facing the necessity to restructure itself in an organizational and technological environment in perpetual evolution. This is why all operations of restructuring first of all require a diagnosis based on a functional analysis. After a presentation of the functionality of a supervisory system for complex processes, we present the concepts of industrial automation and supervisory control and data acquisition (SCADA). Methods: This global analysis exploits the various available documents on the one hand and takes on the other hand in consideration the various testimonies through investigations, the interviews or the collective workshops; otherwise, it also takes observations through visits as a basis and even of the specific operations. The exploitation of this diagnosis enables us to elaborate the project of restructuring thereafter. Leaving from the system analysis for the restructuring of industrial systems, and after a technical diagnosis based on visits, an analysis of the various technical documents and management as well as on targeted interviews, a focusing retailing the various levels of analysis has been done according a general methodology. Results: The methodology adopted in order to contribute to the restructuring of industrial systems by its participative and systemic character and leaning on a large consultation a lot of human resources that of the documentary resources, various innovating actions has been proposed. These actions appear in the setting of the TQM gait requiring applicable parameter quantification and a treatment valorising some information. The new management environment will enable us to institute an information and communication system possibility of migration toward an ERP system. Conclusion: Technological advancements in process monitoring, control and industrial automation over the past decades have contributed greatly to improve the productivity of virtually all industrial systems throughout the world. This paper tries to identify the principles characteristics of a process monitoring, control and industrial automation in order to provide tools to help in the decision-making process.

Keywords: automation, supervision, SCADA, TQM

Procedia PDF Downloads 162
1429 Multimedia Data Fusion for Event Detection in Twitter by Using Dempster-Shafer Evidence Theory

Authors: Samar M. Alqhtani, Suhuai Luo, Brian Regan

Abstract:

Data fusion technology can be the best way to extract useful information from multiple sources of data. It has been widely applied in various applications. This paper presents a data fusion approach in multimedia data for event detection in twitter by using Dempster-Shafer evidence theory. The methodology applies a mining algorithm to detect the event. There are two types of data in the fusion. The first is features extracted from text by using the bag-ofwords method which is calculated using the term frequency-inverse document frequency (TF-IDF). The second is the visual features extracted by applying scale-invariant feature transform (SIFT). The Dempster - Shafer theory of evidence is applied in order to fuse the information from these two sources. Our experiments have indicated that comparing to the approaches using individual data source, the proposed data fusion approach can increase the prediction accuracy for event detection. The experimental result showed that the proposed method achieved a high accuracy of 0.97, comparing with 0.93 with texts only, and 0.86 with images only.

Keywords: data fusion, Dempster-Shafer theory, data mining, event detection

Procedia PDF Downloads 395
1428 Improving Fingerprinting-Based Localization System Using Generative Artificial Intelligence

Authors: Getaneh Berie Tarekegn

Abstract:

A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 39 cm, and more than 90% of the errors are less than 82 cm. That is, numerical results proved that, in comparison to traditional methods, the proposed SRCLoc method can significantly improve positioning performance and reduce radio map construction costs.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 54
1427 Enhancement of Critical Current Density of Liquid Infiltration Processed Y-Ba-Cu-O Bulk Superconductors Used for Flywheel Energy Storage System

Authors: Asif Mahmood, Yousef Alzeghayer

Abstract:

The size effects of a precursor Y2BaCuO5 (Y211) powder on the microstructure and critical current density (Jc) of liquid infiltration growth (LIG)-processed YBa2Cu3O7-y (Y123) bulk superconductors were investigated in terms of milling time (t). YBCO bulk samples having high Jc values have been selected for the flywheel energy storage system. Y211 powders were attrition-milled for 0-10 h in 2 h increments at a fixed rotation speed of 400 RPM. Y211 pre-forms were made by pelletizing the milled Y211 powders followed by subsequent sintering, after which an LIG process with top seeding was applied to the Y211/Ba3Cu5O8 (Y035) pre-forms. Spherical pores were observed in all LIG-processed Y123 samples, and the pore density gradually decreased as t increased from 0 h to 8 h. In addition to the reduced pore density, the Y211 particle size in the final Y123 products also decreased with increasing t. As t increased further to 10 h, unexpected Y211 coarsening and large pore evolutions were observed. The magnetic susceptibility-temperature curves showed that the onset superconducting transition temperature (Tc, onset) of all samples was the same (91.5 K), but the transition width became greater as t increased. The Jc of the Y123 bulk superconductors fabricated in this study was observed to correlate well with t of the Y211 precursor powder. The maximum Jc of 1.0×105 A cm-2 (at 77 K, 0 T) was achieved at t = 8 h, which is attributed to the reduction in pore density and Y211 particle size. The prolonged milling time of t = 10 h decreased the Jc of the LIG-processed Y123 superconductor owing to the evolution of large pores and exaggerated Y211 growth. YBCO bulk samples having high Jc (samples prepared using 8 h milled powders) have been used for the energy storage system in flywheel energy storage system.

Keywords: critical current, bulk superconductor, liquid infiltration, bioinformatics

Procedia PDF Downloads 198
1426 Smart Container Farming: Innovative Urban Strawberry Farming Model from Japan to the World

Authors: Nishantha Giguruwa

Abstract:

This research investigates the transformative potential of smart container farming, building upon the successful cultivation of Japanese mushrooms at Sakai Farms in Aichi Prefecture, Japan, under the strategic collaboration with the Daikei Group. Inspired by this success, the study focuses on establishing an advanced urban strawberry farming laboratory with the aim of understanding strawberry farming technologies, fostering collaboration, and strategizing marketing approaches for both local and global markets. Positioned within the business framework of Sakai Farms and the Daikei Group, the study underscores the sustainability and forward-looking solutions offered by smart container farming in agriculture. The global significance of strawberries is emphasized, acknowledging their economic and cultural importance. The detailed examination of strawberry farming intricacies informs the technological framework developed for smart containers, implemented at Sakai Farms. Integral to this research is the incorporation of controlled bee pollination, a groundbreaking addition to the smart container farming model. The study anticipates future trends, outlining avenues for continuing exploration, stakeholder collaborations, policy considerations, and expansion strategies. Notably, the author expresses a strategic intent to approach the global market, leveraging the foreign student/faculty base at Ritsumeikan Asia Pacific University, where the author is affiliated. This unique approach aims to disseminate the research findings globally, contributing to the broader landscape of agricultural innovation. The integration of controlled bee pollination within this innovative framework not only enhances sustainability but also marks a significant stride in the evolution of urban agriculture, aligning with global agricultural trends.

Keywords: smart container farming, urban agriculture, strawberry farming technologies, controlled bee pollination, agricultural innovation

Procedia PDF Downloads 39
1425 A Local Invariant Generalized Hough Transform Method for Integrated Circuit Visual Positioning

Authors: Wei Feilong

Abstract:

In this study, an local invariant generalized Houghtransform (LI-GHT) method is proposed for integrated circuit (IC) visual positioning. The original generalized Hough transform (GHT) is robust to external noise; however, it is not suitable for visual positioning of IC chips due to the four-dimensionality (4D) of parameter space which leads to the substantial storage requirement and high computational complexity. The proposed LI-GHT method can reduce the dimensionality of parameter space to 2D thanks to the rotational invariance of local invariant geometric feature and it can estimate the accuracy position and rotation angle of IC chips in real-time under noise and blur influence. The experiment results show that the proposed LI-GHT can estimate position and rotation angle of IC chips with high accuracy and fast speed. The proposed LI-GHT algorithm was implemented in IC visual positioning system of radio frequency identification (RFID) packaging equipment.

Keywords: Integrated Circuit Visual Positioning, Generalized Hough Transform, Local invariant Generalized Hough Transform, ICpacking equipment

Procedia PDF Downloads 255
1424 MPPT Control with (P&O) and (FLC) Algorithms of Solar Electric Generator

Authors: Dib Djalel, Mordjaoui Mourad

Abstract:

The current trend towards the exploitation of various renewable energy resources has become indispensable, so it is important to improve the efficiency and reliability of the GPV photovoltaic systems. Maximum Power Point Tracking (MPPT) plays an important role in photovoltaic power systems because it maximize the power output from a PV system for a given set of conditions. This paper presents a new fuzzy logic control based MPPT algorithm for solar panel. The solar panel is modeled and analyzed in Matlab/Simulink. The Solar panel can produce maximum power at a particular operating point called Maximum Power Point(MPP). To produce maximum power and to get maximum efficiency, the entire photovoltaic panel must operate at this particular point. Maximum power point of PV panel keeps on changing with changing environmental conditions such as solar irradiance and cell temperature. Thus, to extract maximum available power from a PV module, MPPT algorithms are implemented and Perturb and Observe (P&O) MPPT and fuzzy logic control FLC, MPPT are developed and compared. Simulation results show the effectiveness of the fuzzy control technique to produce a more stable power.

Keywords: MPPT, photovoltaic panel, fuzzy logic control, modeling, solar power

Procedia PDF Downloads 468
1423 3D Numerical Modelling of a Pulsed Pumping Process of a Large Dense Non-Aqueous Phase Liquid Pool: In situ Pilot-Scale Case Study of Hexachlorobutadiene in a Keyed Enclosure

Authors: Q. Giraud, J. Gonçalvès, B. Paris

Abstract:

Remediation of dense non-aqueous phase liquids (DNAPLs) represents a challenging issue because of their persistent behaviour in the environment. This pilot-scale study investigates, by means of in situ experiments and numerical modelling, the feasibility of the pulsed pumping process of a large amount of a DNAPL in an alluvial aquifer. The main compound of the DNAPL is hexachlorobutadiene, an emerging organic pollutant. A low-permeability keyed enclosure was built at the location of the DNAPL source zone in order to isolate a finite undisturbed volume of soil, and a 3-month pulsed pumping process was applied inside the enclosure to exclusively extract the DNAPL. The water/DNAPL interface elevation at both the pumping and observation wells and the cumulated pumped volume of DNAPL were also recorded. A total volume of about 20m³ of purely DNAPL was recovered since no water was extracted during the process. The three-dimensional and multiphase flow simulator TMVOC was used, and a conceptual model was elaborated and generated with the pre/post-processing tool mView. Numerical model consisted of 10 layers of variable thickness and 5060 grid cells. Numerical simulations reproduce the pulsed pumping process and show an excellent match between simulated, and field data of DNAPL cumulated pumped volume and a reasonable agreement between modelled and observed data for the evolution of the water/DNAPL interface elevations at the two wells. This study offers a new perspective in remediation since DNAPL pumping system optimisation may be performed where a large amount of DNAPL is encountered.

Keywords: dense non-aqueous phase liquid (DNAPL), hexachlorobutadiene, in situ pulsed pumping, multiphase flow, numerical modelling, porous media

Procedia PDF Downloads 167
1422 Modeling and Dynamics Analysis for Intelligent Skid-Steering Vehicle Based on Trucksim-Simulink

Authors: Yansong Zhang, Xueyuan Li, Junjie Zhou, Xufeng Yin, Shihua Yuan, Shuxian Liu

Abstract:

Aiming at the verification of control algorithms for skid-steering vehicles, a vehicle simulation model of 6×6 electric skid-steering unmanned vehicle was established based on Trucksim and Simulink. The original transmission and steering mechanism of Trucksim are removed, and the electric skid-steering model and a closed-loop controller for the vehicle speed and yaw rate are built in Simulink. The simulation results are compared with the ones got by theoretical formulas. The results show that the predicted tire mechanics and vehicle kinematics of Trucksim-Simulink simulation model are closed to the theoretical results. Therefore, it can be used as an effective approach to study the dynamic performance and control algorithm of skid-steering vehicle. In this paper, a method of motion control based on feed forward control is also designed. The simulation results show that the feed forward control strategy can make the vehicle follow the target yaw rate more quickly and accurately, which makes the vehicle have more maneuverability.

Keywords: skid-steering, Trucksim-Simulink, feedforward control, dynamics

Procedia PDF Downloads 313
1421 Transient Phenomena in a 100 W Hall Thrusters: Experimental Measurements of Discharge Current and Plasma Parameter Evolution

Authors: Clémence Royer, Stéphane Mazouffre

Abstract:

Nowadays, electric propulsion systems play a crucial role in space exploration missions due to their high specific impulse and long operational life. The Hall thrusters are one of the most mature EP technologies. It is a gridless ion thruster that has proved reliable and high-performance for decades in various space missions. Operation of HT relies on electron emissions through a cathode placed outside a hollow dielectric channel that includes an anode at the back. Negatively charged particles are trapped in a magnetic field and efficiently slow down. By collisions, the electron cloud ionizes xenon atoms. A large electric field is generated in the axial direction due to the low electron transverse mobility in the region of a strong magnetic field. Positive particles are pulled out of the chamber at high velocity and are neutralized directly at the exhaust area. This phenomenon leads to the acceleration of the spacecraft system at a high specific impulse. While HT’s architecture and operating principle are relatively simple, the physics behind thrust is complex and still partly unknown. Current and voltage oscillations, as well as electron properties, have been captured over a 30 mn time period after ignition. The observed low-frequency oscillations exhibited specific frequency ranges, amplitudes, and stability patterns. Correlations between the oscillations and plasma characteristics we analyzed. The impact of these instabilities on thruster performance, including thrust efficiency, has been evaluated as well. Moreover, strategies for mitigating and controlling these instabilities have been developed, such as filtering. In this contribution, in addition to presenting a summary of the results obtained in the transient regime, we will present and discuss recent advances in Hall thruster plasma discharge filtering and control.

Keywords: electric propulsion, Hall Thruster, plasma diagnostics, low-frequency oscillations

Procedia PDF Downloads 74
1420 Influence of the Paint Coating Thickness in Digital Image Correlation Experiments

Authors: Jesús A. Pérez, Sam Coppieters, Dimitri Debruyne

Abstract:

In the past decade, the use of digital image correlation (DIC) techniques has increased significantly in the area of experimental mechanics, especially for materials behavior characterization. This non-contact tool enables full field displacement and strain measurements over a complete region of interest. The DIC algorithm requires a random contrast pattern on the surface of the specimen in order to perform properly. To create this pattern, the specimen is usually first coated using a white matt paint. Next, a black random speckle pattern is applied using any suitable method. If the applied paint coating is too thick, its top surface may not be able to exactly follow the deformation of the specimen, and consequently, the strain measurement might be underestimated. In the present article, a study of the influence of the paint thickness on the strain underestimation is performed for different strain levels. The results are then compared to typical paint coating thicknesses applied by experienced DIC users. A slight strain underestimation was observed for paint coatings thicker than about 30μm. On the other hand, this value was found to be uncommonly high compared to coating thicknesses applied by DIC users.

Keywords: digital image correlation, paint coating thickness, strain

Procedia PDF Downloads 505
1419 Optimization of Marine Waste Collection Considering Dynamic Transport and Ship’s Wake Impact

Authors: Guillaume Richard, Sarra Zaied

Abstract:

Marine waste quantities increase more and more, 5 million tons of plastic waste enter the ocean every year. Their spatiotemporal distribution is never homogeneous and depends mainly on the hydrodynamic characteristics of the environment, as well as the size and location of the waste. As part of optimizing collect of marine plastic wastes, it is important to measure and monitor their evolution over time. In this context, diverse studies have been dedicated to describing waste behavior in order to identify its accumulation in ocean areas. None of the existing tools which track objects at sea had the objective of tracking down a slick of waste. Moreover, the applications related to marine waste are in the minority compared to rescue applications or oil slicks tracking applications. These approaches are able to accurately simulate an object's behavior over time but not during the collection mission of a waste sheet. This paper presents numerical modeling of a boat’s wake impact on the floating marine waste behavior during a collection mission. The aim is to predict the trajectory of a marine waste slick to optimize its collection using meteorological data of ocean currents, wind, and possibly waves. We have made the choice to use Ocean Parcels which is a Python library suitable for trajectoring particles in the ocean. The modeling results showed the important role of advection and diffusion processes in the spatiotemporal distribution of floating plastic litter. The performance of the proposed method was evaluated on real data collected from the Copernicus Marine Environment Monitoring Service (CMEMS). The results of the evaluation in Cape of Good Hope (South Africa) prove that the proposed approach can effectively predict the position and velocity of marine litter during collection, which allowed for optimizing time and more than $90\%$ of the amount of collected waste.

Keywords: marine litter, advection-diffusion equation, sea current, numerical model

Procedia PDF Downloads 77
1418 Argumentation Frameworks and Theories of Judging

Authors: Sonia Anand Knowlton

Abstract:

With the rise of artificial intelligence, computer science is becoming increasingly integrated in virtually every area of life. Of course, the law is no exception. Through argumentation frameworks (AFs), computer scientists have used abstract algebra to structure the legal reasoning process in a way that allows conclusions to be drawn from a formalized system of arguments. In AFs, arguments compete against each other for logical success and are related to one another through the binary operation of the attack. The prevailing arguments make up the preferred extension of the given argumentation framework, telling us what set of arguments must be accepted from a logical standpoint. There have been several developments of AFs since its original conception in the early 90’s in efforts to make them more aligned with the human reasoning process. Generally, these developments have sought to add nuance to the factors that influence the logical success of competing arguments (e.g., giving an argument more logical strength based on the underlying value it promotes). The most cogent development was that of the Extended Argumentation Framework (EAF), in which attacks can themselves be attacked by other arguments, and the promotion of different competing values can be formalized within the system. This article applies the logical structure of EAFs to current theoretical understandings of judicial reasoning to contribute to theories of judging and to the evolution of AFs simultaneously. The argument is that the main limitation of EAFs, when applied to judicial reasoning, is that they require judges to themselves assign values to different arguments and then lexically order these values to determine the given framework’s preferred extension. Drawing on John Rawls’ Theory of Justice, the examination that follows is whether values are lexical and commensurable to this extent. The analysis that follows then suggests a potential extension of the EAF system with an approach that formalizes different “planes of attack” for competing arguments that promote lexically ordered values. This article concludes with a summary of how these insights contribute to theories of judging and of legal reasoning more broadly, specifically in indeterminate cases where judges must turn to value-based approaches.

Keywords: computer science, mathematics, law, legal theory, judging

Procedia PDF Downloads 47
1417 Analysis Of Non-uniform Characteristics Of Small Underwater Targets Based On Clustering

Authors: Tianyang Xu

Abstract:

Small underwater targets generally have a non-centrosymmetric geometry, and the acoustic scattering field of the target has spatial inhomogeneity under active sonar detection conditions. In view of the above problems, this paper takes the hemispherical cylindrical shell as the research object, and considers the angle continuity implied in the echo characteristics, and proposes a cluster-driven research method for the non-uniform characteristics of target echo angle. First, the target echo features are extracted, and feature vectors are constructed. Secondly, the t-SNE algorithm is used to improve the internal connection of the feature vector in the low-dimensional feature space and to construct the visual feature space. Finally, the implicit angular relationship between echo features is extracted under unsupervised condition by cluster analysis. The reconstruction results of the local geometric structure of the target corresponding to different categories show that the method can effectively divide the angle interval of the local structure of the target according to the natural acoustic scattering characteristics of the target.

Keywords: underwater target;, non-uniform characteristics;, cluster-driven method;, acoustic scattering characteristics

Procedia PDF Downloads 106
1416 Median-Based Nonparametric Estimation of Returns in Mean-Downside Risk Portfolio Frontier

Authors: H. Ben Salah, A. Gannoun, C. de Peretti, A. Trabelsi

Abstract:

The Downside Risk (DSR) model for portfolio optimisation allows to overcome the drawbacks of the classical mean-variance model concerning the asymetry of returns and the risk perception of investors. This model optimization deals with a positive definite matrix that is endogenous with respect to portfolio weights. This aspect makes the problem far more difficult to handle. For this purpose, Athayde (2001) developped a new recurcive minimization procedure that ensures the convergence to the solution. However, when a finite number of observations is available, the portfolio frontier presents an appearance which is not very smooth. In order to overcome that, Athayde (2003) proposed a mean kernel estimation of the returns, so as to create a smoother portfolio frontier. This technique provides an effect similar to the case in which we had continuous observations. In this paper, taking advantage on the the robustness of the median, we replace the mean estimator in Athayde's model by a nonparametric median estimator of the returns. Then, we give a new version of the former algorithm (of Athayde (2001, 2003)). We eventually analyse the properties of this improved portfolio frontier and apply this new method on real examples.

Keywords: Downside Risk, Kernel Method, Median, Nonparametric Estimation, Semivariance

Procedia PDF Downloads 473
1415 Optimization of Reinforced Concrete Buildings According to the Algerian Seismic Code

Authors: Nesreddine Djafar Henni, Nassim Djedoui, Rachid Chebili

Abstract:

Recent decades have witnessed significant efforts being made to optimize different types of structures and components. The concept of cost optimization in reinforced concrete structures, which aims at minimizing financial resources while ensuring maximum building safety, comprises multiple materials, and the objective function for their optimal design is derived from the construction cost of the steel as well as concrete that significantly contribute to the overall weight of reinforced concrete (RC) structures. To achieve this objective, this work has been devoted to optimizing the structural design of 3D RC frame buildings which integrates, for the first time, the Algerian regulations. Three different test examples were investigated to assess the efficiency of our work in optimizing RC frame buildings. The hybrid GWOPSO algorithm is used, and 30000 generations are made. The cost of the building is reduced by iteration each time. Concrete and reinforcement bars are used in the building cost. As a result, the cost of a reinforced concrete structure is reduced by 30% compared with the initial design. This result means that the 3D cost-design optimization of the framed structure is successfully achieved.

Keywords: optimization, automation, API, Malab, RC structures

Procedia PDF Downloads 32
1414 The Effectiveness of a Hybrid Diffie-Hellman-RSA-Advanced Encryption Standard Model

Authors: Abdellahi Cheikh

Abstract:

With the emergence of quantum computers with very powerful capabilities, the security of the exchange of shared keys between two interlocutors poses a big problem in terms of the rapid development of technologies such as computing power and computing speed. Therefore, the Diffie-Hellmann (DH) algorithm is more vulnerable than ever. No mechanism guarantees the security of the key exchange, so if an intermediary manages to intercept it, it is easy to intercept. In this regard, several studies have been conducted to improve the security of key exchange between two interlocutors, which has led to interesting results. The modification made on our model Diffie-Hellman-RSA-AES (DRA), which encrypts the information exchanged between two users using the three-encryption algorithms DH, RSA and AES, by using stenographic photos to hide the contents of the p, g and ClesAES values that are sent in an unencrypted state at the level of DRA model to calculate each user's public key. This work includes a comparative study between the DRA model and all existing solutions, as well as the modification made to this model, with an emphasis on the aspect of reliability in terms of security. This study presents a simulation to demonstrate the effectiveness of the modification made to the DRA model. The obtained results show that our model has a security advantage over the existing solution, so we made these changes to reinforce the security of the DRA model.

Keywords: Diffie-Hellmann, DRA, RSA, advanced encryption standard

Procedia PDF Downloads 76
1413 A Historical Analysis of The Concept of Equivalence from Different Theoretical Perspectives in Translation Studies

Authors: Amenador Kate Benedicta, Wang Zhiwei

Abstract:

Since the later parts of the 20th century, the notion of equivalence continues to be a central and critical concept in the development of translation theory. After decades of arguments over word-for-word and free translations methods, scholars attempting to develop more systematic and efficient translation theories began to focus on fundamental translation concepts such as equivalence. Although the concept of equivalence has piqued the interest of many scholars, its definition, scope, and applicability have sparked contentious arguments within the discipline. As a result, several distinct theories and explanations on the concept of equivalence have been put forward over the last half-century. Thus, this study explores and discusses the evolution of the critical concept of equivalence in translation studies through a bibliometric method of investigation of manual and digital books and articles by analyzing different scholars' key contributions and limitations on equivalence from various theoretical perspectives. While analyzing them, emphasis is placed on the innovations that each theory has brought to the comprehension of equivalence. In order to achieve the aim of the study, the article began by discussing the contributions of linguistically motivated theories to the notion of equivalence in translation, followed by functionalist-oriented contributions, before moving on to more recent advancements in translation studies on the concept. Because equivalence is such a broad notion, it is impossible to discuss each researcher in depth. As a result, the most well-known names and their equivalent theories are compared and contrasted in this research. The study emphasizes the developmental progression in our comprehension of the equivalence concept and equivalent effect. It concluded that the various theoretical perspective's contributions to the notion of equivalence rather complement and make up for the limitations of each other. The study also highlighted how troublesome the equivalent concept might become in terms of identifying the nature of translation and how central and unavoidable the concept is in every translation action, despite its limitations. The significance of the study lies in its synthesis of the different contributions and limitations of the various theories offered by scholars on the notion of equivalence, lending literature to both student and scholars in the field, and providing insight on future theoretical development

Keywords: equivalence, functionalist translation theories, linguistic translation approaches, translation theories, Skopos

Procedia PDF Downloads 101
1412 East West Discourse: An Esoteric Comparison of the Western Philosophy and the Eastern Vedanta

Authors: Chandrabati Chakraborty

Abstract:

The progressive emergence, in the course of the evolution of life, mind and personality, requires us to assume a creative Principle operating timeless Reality in the temporal. The difference between Western philosophy and that of India, concerns the origin and the purpose of the philosophical enquiry. While the former wonders at the external world, the latter is awareness of perennial suffering associated with human existence. The present world suffers from a basic form of rootlessness,reflecting many psychological, philosophical studies. Alienation,a major theme of human condition in the contemporary epoch has emerged as natural consequences of existential predicament. As Edmund Fuller also observes that individuals suffer not only from famine, ruin or even war but also from devastating inner problems, which are that of estrangement, hopelessness and utter despair. This existentialism is thus considered by Jean Wahl as the “Philosophies of existence”.The post world war scenario well analyses the chaos,annihilation,frustration and anguished estrangement. In such conditions when the West cries out , “What is there?I know first of all that I am.But who am I?.....I am sepeated.What I am seperated from I cannot name it. But I am seperated.”(Dostoevsky:The Confession), Vedantic philosophy looks upon the Pilgrim’s Progress of Humanity as being essentially one,operationg squarely within the bounds of reality, reflecting a basic human experience, outbraving indecorous dictims that have failed to give due honour to human beings,echoing for centuries the Sanskrit slokas with ultimate certitude: II Esa Atma samaha plusina samo masakena samo nagena sama ebhis tribhir lokaih.....sama nena sarvena II (The Atman (Divine Soul) is the same in the ant, the same in the gnat, the same in the elephant, the same in these three worlds....the same in the whole Universe). The present paper aims at a comparative study of cultural and philosophical expression taking into view extensive illustrations from Western Philosophers and The Vedantic,Upanishadic lores of Indian philosophy.

Keywords: existentialism, Vedanta, philosophy, absurdism

Procedia PDF Downloads 74
1411 Parameter Tuning of Complex Systems Modeled in Agent Based Modeling and Simulation

Authors: Rabia Korkmaz Tan, Şebnem Bora

Abstract:

The major problem encountered when modeling complex systems with agent-based modeling and simulation techniques is the existence of large parameter spaces. A complex system model cannot be expected to reflect the whole of the real system, but by specifying the most appropriate parameters, the actual system can be represented by the model under certain conditions. When the studies conducted in recent years were reviewed, it has been observed that there are few studies for parameter tuning problem in agent based simulations, and these studies have focused on tuning parameters of a single model. In this study, an approach of parameter tuning is proposed by using metaheuristic algorithms such as Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Artificial Bee Colonies (ABC), Firefly (FA) algorithms. With this hybrid structured study, the parameter tuning problems of the models in the different fields were solved. The new approach offered was tested in two different models, and its achievements in different problems were compared. The simulations and the results reveal that this proposed study is better than the existing parameter tuning studies.

Keywords: parameter tuning, agent based modeling and simulation, metaheuristic algorithms, complex systems

Procedia PDF Downloads 212