Search results for: Radial Basis Functions (RBF) neural networks
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9470

Search results for: Radial Basis Functions (RBF) neural networks

7430 Applying Renowned Energy Simulation Engines to Neural Control System of Double Skin Façade

Authors: Zdravko Eškinja, Lovre Miljanić, Ognjen Kuljača

Abstract:

This paper is an overview of simulation tools used to model specific thermal dynamics that occurs while controlling double skin façade. Research has been conducted on simplified construction with single zone where one side is glazed. Heat flow and temperature responses are simulated in three different simulation tools: IDA-ICE, EnergyPlus and HAMBASE. The excitation of observed system, used in all simulations, was a temperature step of exterior environment. Air infiltration, insulation and other disturbances are excluded from this research. Although such isolated behaviour is not possible in reality, experiments are carried out to gain novel information about heat flow transients which are not observable under regular conditions. Results revealed new possibilities for adapting the parameters of the neural network regulator. Along numerical simulations, the same set-up has been also tested in a real-time experiment with a 1:18 scaled model and thermal chamber. The comparison analysis brings out interesting conclusion about simulation accuracy in this particular case.

Keywords: double skin façade, experimental tests, heat control, heat flow, simulated tests, simulation tools

Procedia PDF Downloads 231
7429 Ontology Expansion via Synthetic Dataset Generation and Transformer-Based Concept Extraction

Authors: Andrey Khalov

Abstract:

The rapid proliferation of unstructured data in IT infrastructure management demands innovative approaches for extracting actionable knowledge. This paper presents a framework for ontology-based knowledge extraction that combines relational graph neural networks (R-GNN) with large language models (LLMs). The proposed method leverages the DOLCE framework as the foundational ontology, extending it with concepts from ITSMO for domain-specific applications in IT service management and outsourcing. A key component of this research is the use of transformer-based models, such as DeBERTa-v3-large, for automatic entity and relationship extraction from unstructured texts. Furthermore, the paper explores how transfer learning techniques can be applied to fine-tune large language models (LLaMA) for using to generate synthetic datasets to improve precision in BERT-based entity recognition and ontology alignment. The resulting IT Ontology (ITO) serves as a comprehensive knowledge base that integrates domain-specific insights from ITIL processes, enabling more efficient decision-making. Experimental results demonstrate significant improvements in knowledge extraction and relationship mapping, offering a cutting-edge solution for enhancing cognitive computing in IT service environments.

Keywords: ontology expansion, synthetic dataset, transformer fine-tuning, concept extraction, DOLCE, BERT, taxonomy, LLM, NER

Procedia PDF Downloads 14
7428 Spatiotemporal Neural Network for Video-Based Pose Estimation

Authors: Bin Ji, Kai Xu, Shunyu Yao, Jingjing Liu, Ye Pan

Abstract:

Human pose estimation is a popular research area in computer vision for its important application in human-machine interface. In recent years, 2D human pose estimation based on convolution neural network has got great progress and development. However, in more and more practical applications, people often need to deal with tasks based on video. It’s not far-fetched for us to consider how to combine the spatial and temporal information together to achieve a balance between computing cost and accuracy. To address this issue, this study proposes a new spatiotemporal model, namely Spatiotemporal Net (STNet) to combine both temporal and spatial information more rationally. As a result, the predicted keypoints heatmap is potentially more accurate and spatially more precise. Under the condition of ensuring the recognition accuracy, the algorithm deal with spatiotemporal series in a decoupled way, which greatly reduces the computation of the model, thus reducing the resource consumption. This study demonstrate the effectiveness of our network over the Penn Action Dataset, and the results indicate superior performance of our network over the existing methods.

Keywords: convolutional long short-term memory, deep learning, human pose estimation, spatiotemporal series

Procedia PDF Downloads 148
7427 Clubhouse: A Minor Rebellion against the Algorithmic Tyranny of the Majority

Authors: Vahid Asadzadeh, Amin Ataee

Abstract:

Since the advent of social media, there has been a wave of optimism among researchers and civic activists about the influence of virtual networks on the democratization process, which has gradually waned. One of the lesser-known concerns is how to increase the possibility of hearing the voices of different minorities. According to the theory of media logic, the media, using their technological capabilities, act as a structure through which events and ideas are interpreted. Social media, through the use of the learning machine and the use of algorithms, has formed a kind of structure in which the voices of minorities and less popular topics are lost among the commotion of the trends. In fact, the recommended systems and algorithms used in social media are designed to help promote trends and make popular content more popular, and content that belongs to minorities is constantly marginalized. As social networks gradually play a more active role in politics, the possibility of freely participating in the reproduction and reinterpretation of structures in general and political structures in particular (as Laclau‎ and Mouffe had in mind‎) can be considered as criteria to democracy in action. The point is that the media logic of virtual networks is shaped by the rule and even the tyranny of the majority, and this logic does not make it possible to design a self-foundation and self-revolutionary model of democracy. In other words, today's social networks, though seemingly full of variety But they are governed by the logic of homogeneity, and they do not have the possibility of multiplicity as is the case in immanent radical democracies (influenced by Gilles Deleuze). However, with the emergence and increasing popularity of Clubhouse as a new social media, there seems to be a shift in the social media space, and that is the diminishing role of algorithms and systems reconditioners as content delivery interfaces. This has led to the fact that in the Clubhouse, the voices of minorities are better heard, and the diversity of political tendencies manifests itself better. The purpose of this article is to show, first, how social networks serve the elimination of minorities in general, and second, to argue that the media logic of social networks must adapt to new interpretations of democracy that give more space to minorities and human rights. Finally, this article will show how the Clubhouse serves the new interpretations of democracy at least in a minimal way. To achieve the mentioned goals, in this article by a descriptive-analytical method, first, the relation between media logic and postmodern democracy will be inquired. The political economy popularity in social media and its conflict with democracy will be discussed. Finally, it will be explored how the Clubhouse provides a new horizon for the concepts embodied in radical democracy, a horizon that more effectively serves the rights of minorities and human rights in general.

Keywords: algorithmic tyranny, Clubhouse, minority rights, radical democracy, social media

Procedia PDF Downloads 146
7426 Effects of Dietary E on Semen, Hormonal Profile and Testicular Biometry in Teddy Goat Bucks

Authors: Muhammad Zubair, Maqbool Ahmad, Al-Hafizah Shafia Tehseen Gul, Shujait Ali

Abstract:

The use of vitamins has significant effects on the reproductive system of mammals. The present study was conducted to investigate the useful effects of vitamin E on reproductive functions of Teddy bucks. For this purpose, 8 adult Teddy bucks were randomly divided into two treatment groups viz; A (control) and B (vitamin E with dose of 200 mg/kg BW/day). These treatments continued for 12 weeks. Semen quality parameters (volume, motility, sperm morphology and sperm DNA integrity) of experimental bucks of each group was evaluated on weekly basis, while testicular measurements (length, scrotal circumference and weights) were recorded at 0 and 12th week of experiment. Serum concentrations of male sex hormones (testosterone, LH, FSH) and cortisol were recorded fortnightly. Similarly, body weights of bucks were also measured fortnightly until completion of the study. The data were subjected to two-way analysis of variance, followed by Duncan test for multiple mean comparisons. Supplementation of vitamin E improved significantly (P<0.05) the semen quality parameter, body weights, testicular measurements and serum levels of sex hormones. However, there was no effect on serum cortisol. It was concluded from the present study that dietary supplementation of vitamin E has beneficial effects on the semen and hormones in male reproductive system.

Keywords: hormones, semen, teddy bucks, testicular measurements

Procedia PDF Downloads 293
7425 Traffic Congestion Analysis and Modeling for Urban Roads of Srinagar City

Authors: Adinarayana Badveeti, Mohammad Shafi Mir

Abstract:

In Srinagar City, in India, traffic congestion is a condition on transport networks that occurs as use increases and is characterized by slower speeds, longer trip times, and increased vehicular queuing. Traffic congestion is conventionally measured using indicators such as roadway level-of-service, the Travel Time Index and their variants. Several measures have been taken in order to counteract congestion like road pricing, car pooling, improved traffic management, etc. While new road construction can temporarily relieve congestion in the longer term, it simply encourages further growth in car traffic through increased travel and a switch away from public transport. The full paper report, on which this abstract is based, aims to provide policymakers and technical staff with the real-time data, conceptual framework and guidance on some of the engineering tools necessary to manage congestion in such a way as to reduce its overall impact on individuals, families, communities, and societies dynamic, affordable, liveable and attractive urban regions will never be free of congestion. Road transport policies, however, should seek to manage congestion on a cost-effective basis with the aim of reducing the burden that excessive congestion imposes upon travellers and urban dwellers throughout the urban road network.

Keywords: traffic congestion, modeling, traffic management, travel time index

Procedia PDF Downloads 319
7424 A Modified Decoupled Semi-Analytical Approach Based On SBFEM for Solving 2D Elastodynamic Problems

Authors: M. Fakharian, M. I. Khodakarami

Abstract:

In this paper, a new trend for improvement in semi-analytical method based on scale boundaries in order to solve the 2D elastodynamic problems is provided. In this regard, only the boundaries of the problem domain discretization are by specific sub-parametric elements. Mapping functions are uses as a class of higher-order Lagrange polynomials, special shape functions, Gauss-Lobatto -Legendre numerical integration, and the integral form of the weighted residual method, the matrix is diagonal coefficients in the equations of elastodynamic issues. Differences between study conducted and prior research in this paper is in geometry production procedure of the interpolation function and integration of the different is selected. Validity and accuracy of the present method are fully demonstrated through two benchmark problems which are successfully modeled using a few numbers of DOFs. The numerical results agree very well with the analytical solutions and the results from other numerical methods.

Keywords: 2D elastodynamic problems, lagrange polynomials, G-L-Lquadrature, decoupled SBFEM

Procedia PDF Downloads 444
7423 Bridging the Gap between M and E, and KM: Towards the Integration of Evidence-Based Information and Policy Decision-Making

Authors: Xueqing Ivy Chen, Christo De Coning

Abstract:

It is clear from practice that a gap exists between Result-Based Monitoring and Evaluation (RBME) as a discipline, and Knowledge Management (KM) on the other hand. Whereas various government departments have institutionalised these functions, KM and M&E has functioned in isolation from each other in a practical sense in the public sector. It’s therefore necessary to explore the relationship between KM and M&E and the necessity for integration, so that a convergence of these disciplines can be established. An integration of KM and M&E will lead to integration and improvement of evidence-based information and policy decision-making. M&E and KM process models are available but the complementarity between specific process steps of these process models are not exploited. A need exists to clarify the relationships between these functions in order to ensure evidence based information and policy decision-making. This paper will depart from the well-known policy process models, such as the generic model and consider recent on the interface between policy, M&E and KM.

Keywords: result-based monitoring and evaluation, RBME, knowledge management, KM, evident based decision making, public policy, information systems, institutional arrangement

Procedia PDF Downloads 152
7422 Finite Element Simulation of Four Point Bending of Laminated Veneer Lumber (LVL) Arch

Authors: Eliska Smidova, Petr Kabele

Abstract:

This paper describes non-linear finite element simulation of laminated veneer lumber (LVL) under tensile and shear loads that induce cracking along fibers. For this purpose, we use 2D homogeneous orthotropic constitutive model of tensile and shear fracture in timber that has been recently developed and implemented into ATENA® finite element software by the authors. The model captures (i) material orthotropy for small deformations in both linear and non-linear range, (ii) elastic behavior until anisotropic failure criterion is fulfilled, (iii) inelastic behavior after failure criterion is satisfied, (iv) different post-failure response for cracks along and across the grain, (v) unloading/reloading behavior. The post-cracking response is treated by fixed smeared crack model where Reinhardt-Hordijk function is used. The model requires in total 14 input parameters that can be obtained from standard tests, off-axis test results and iterative numerical simulation of compact tension (CT) or compact tension-shear (CTS) test. New engineered timber composites, such as laminated veneer lumber (LVL), offer improved structural parameters compared to sawn timber. LVL is manufactured by laminating 3 mm thick wood veneers aligned in one direction using water-resistant adhesives (e.g. polyurethane). Thus, 3 main grain directions, namely longitudinal (L), tangential (T), and radial (R), are observed within the layered LVL product. The core of this work consists in 3 numerical simulations of experiments where Radiata Pine LVL and Yellow Poplar LVL were involved. The first analysis deals with calibration and validation of the proposed model through off-axis tensile test (at a load-grain angle of 0°, 10°, 45°, and 90°) and CTS test (at a load-grain angle of 30°, 60°, and 90°), both of which were conducted for Radiata Pine LVL. The second finite element simulation reproduces load-CMOD curve of compact tension (CT) test of Yellow Poplar with the aim of obtaining cohesive law parameters to be used as an input in the third finite element analysis. That is four point bending test of small-size arch of 780 mm span that is made of Yellow Poplar LVL. The arch is designed with a through crack between two middle layers in the crown. Curved laminated beams are exposed to high radial tensile stress compared to timber strength in radial tension in the crown area. Let us note that in this case the latter parameter stands for tensile strength in perpendicular direction with respect to the grain. Standard tests deliver most of the relevant input data whereas traction-separation law for crack along the grain can be obtained partly by inverse analysis of compact tension (CT) test or compact tension-shear test (CTS). The initial crack was modeled as a narrow gap separating two layers in the middle the arch crown. Calculated load-deflection curve is in good agreement with the experimental ones. Furthermore, crack pattern given by numerical simulation coincides with the most important observed crack paths.

Keywords: compact tension (CT) test, compact tension shear (CTS) test, fixed smeared crack model, four point bending test, laminated arch, laminated veneer lumber LVL, off-axis test, orthotropic elasticity, orthotropic fracture criterion, Radiata Pine LVL, traction-separation law, yellow poplar LVL, 2D constitutive model

Procedia PDF Downloads 290
7421 Increasing of Resiliency by Using Gas Storage in Iranian Gas Network

Authors: Mohsen Dourandish

Abstract:

Iran has a huge pipeline network in every state of country which is the longest and vastest pipeline network after Russia and USA (360,000 Km high pressure pipelines and 250,000 Km distribution networks). Furthermore in recent years National Iranian Gas Company is planning to develop natural gas network to cover all cities and villages above 20 families, in a way that 97 percent of Iran population will be gas consumer by 2020. In this condition, network resiliency will be the first priority of NIGC and due to that several planning for increasing resiliency of gas network is under construction. The most important strategy of NIGC is converting tree form pattern network to loop gas networks and developing underground gas storage near main gas consuming centers. In this regard NIGC is planning for construction of over 3500 km high-pressure pipeline and also 10 TCM gas storage capacities in UGSs.

Keywords: Iranian gas network, peak shaving, resiliency, underground gas storage

Procedia PDF Downloads 325
7420 'I Mean' in Teacher Questioning Sequences in Post-Task Discussions: A Conversation Analytic Study

Authors: Derya Duran, Christine Jacknick

Abstract:

Despite a growing body of research on classroom, especially language classroom interactions, much more is yet to be discovered on how interaction is organized in higher education settings. This study investigates how the discourse marker 'I mean' in teacher questioning turns functions as a resource to promote student participation as well as to enhance collective understanding in whole-class discussions. This paper takes a conversation analytic perspective, drawing on 30-hour video recordings of classroom interaction in an English as a medium of instruction university in Turkey. Two content classrooms (i.e., Guidance) were observed during an academic term. The course was offered to 4th year students (n=78) in the Faculty of Education; students were majoring in different subjects (i.e., Early Childhood Education, Foreign Language Education, Mathematics Education). Results of the study demonstrate the multi-functionality of discourse marker 'I mean' in teacher questioning turns. In the context of English as a medium of instruction classrooms where possible sources of confusion may occur, we found that 'I mean' is primarily used to indicate upcoming adjustments. More specifically, it is employed for a variety of interactional purposes such as elaboration, clarification, specification, reformulation, and reference to the instructional activity. The study sheds light on the multiplicity of functions of the discourse marker in academic interactions and it uncovers how certain linguistic resources serve functions to the organization of repair such as the maintenance of understanding in classroom interaction. In doing so, it also shows the ways in which participation is routinely enacted in shared interactional events through linguistic resources.

Keywords: conversation analysis, discourse marker, English as a medium of instruction, repair

Procedia PDF Downloads 161
7419 Hybrid Weighted Multiple Attribute Decision Making Handover Method for Heterogeneous Networks

Authors: Mohanad Alhabo, Li Zhang, Naveed Nawaz

Abstract:

Small cell deployment in 5G networks is a promising technology to enhance capacity and coverage. However, unplanned deployment may cause high interference levels and high number of unnecessary handovers, which in turn will result in an increase in the signalling overhead. To guarantee service continuity, minimize unnecessary handovers, and reduce signalling overhead in heterogeneous networks, it is essential to properly model the handover decision problem. In this paper, we model the handover decision according to Multiple Attribute Decision Making (MADM) method, specifically Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS). In this paper, we propose a hybrid TOPSIS method to control the handover in heterogeneous network. The proposed method adopts a hybrid weighting, which is a combination of entropy and standard deviation. A hybrid weighting control parameter is introduced to balance the impact of the standard deviation and entropy weighting on the network selection process and the overall performance. Our proposed method shows better performance, in terms of the number of frequent handovers and the mean user throughput, compared to the existing methods.

Keywords: handover, HetNets, interference, MADM, small cells, TOPSIS, weight

Procedia PDF Downloads 149
7418 Develop a Conceptual Data Model of Geotechnical Risk Assessment in Underground Coal Mining Using a Cloud-Based Machine Learning Platform

Authors: Reza Mohammadzadeh

Abstract:

The major challenges in geotechnical engineering in underground spaces arise from uncertainties and different probabilities. The collection, collation, and collaboration of existing data to incorporate them in analysis and design for given prospect evaluation would be a reliable, practical problem solving method under uncertainty. Machine learning (ML) is a subfield of artificial intelligence in statistical science which applies different techniques (e.g., Regression, neural networks, support vector machines, decision trees, random forests, genetic programming, etc.) on data to automatically learn and improve from them without being explicitly programmed and make decisions and predictions. In this paper, a conceptual database schema of geotechnical risks in underground coal mining based on a cloud system architecture has been designed. A new approach of risk assessment using a three-dimensional risk matrix supported by the level of knowledge (LoK) has been proposed in this model. Subsequently, the model workflow methodology stages have been described. In order to train data and LoK models deployment, an ML platform has been implemented. IBM Watson Studio, as a leading data science tool and data-driven cloud integration ML platform, is employed in this study. As a Use case, a data set of geotechnical hazards and risk assessment in underground coal mining were prepared to demonstrate the performance of the model, and accordingly, the results have been outlined.

Keywords: data model, geotechnical risks, machine learning, underground coal mining

Procedia PDF Downloads 274
7417 R-Killer: An Email-Based Ransomware Protection Tool

Authors: B. Lokuketagoda, M. Weerakoon, U. Madushan, A. N. Senaratne, K. Y. Abeywardena

Abstract:

Ransomware has become a common threat in past few years and the recent threat reports show an increase of growth in Ransomware infections. Researchers have identified different variants of Ransomware families since 2015. Lack of knowledge of the user about the threat is a major concern. Ransomware detection methodologies are still growing through the industry. Email is the easiest method to send Ransomware to its victims. Uninformed users tend to click on links and attachments without much consideration assuming the emails are genuine. As a solution to this in this paper R-Killer Ransomware detection tool is introduced. Tool can be integrated with existing email services. The core detection Engine (CDE) discussed in the paper focuses on separating suspicious samples from emails and handling them until a decision is made regarding the suspicious mail. It has the capability of preventing execution of identified ransomware processes. On the other hand, Sandboxing and URL analyzing system has the capability of communication with public threat intelligence services to gather known threat intelligence. The R-Killer has its own mechanism developed in its Proactive Monitoring System (PMS) which can monitor the processes created by downloaded email attachments and identify potential Ransomware activities. R-killer is capable of gathering threat intelligence without exposing the user’s data to public threat intelligence services, hence protecting the confidentiality of user data.

Keywords: ransomware, deep learning, recurrent neural networks, email, core detection engine

Procedia PDF Downloads 215
7416 A New Concept for Deriving the Expected Value of Fuzzy Random Variables

Authors: Liang-Hsuan Chen, Chia-Jung Chang

Abstract:

Fuzzy random variables have been introduced as an imprecise concept of numeric values for characterizing the imprecise knowledge. The descriptive parameters can be used to describe the primary features of a set of fuzzy random observations. In fuzzy environments, the expected values are usually represented as fuzzy-valued, interval-valued or numeric-valued descriptive parameters using various metrics. Instead of the concept of area metric that is usually adopted in the relevant studies, the numeric expected value is proposed by the concept of distance metric in this study based on two characters (fuzziness and randomness) of FRVs. Comparing with the existing measures, although the results show that the proposed numeric expected value is same with those using the different metric, if only triangular membership functions are used. However, the proposed approach has the advantages of intuitiveness and computational efficiency, when the membership functions are not triangular types. An example with three datasets is provided for verifying the proposed approach.

Keywords: fuzzy random variables, distance measure, expected value, descriptive parameters

Procedia PDF Downloads 343
7415 “The Unbearable Lightness of Being” Book as an Interdisciplinary Study Basis for Students’ Learning Process about Love and Politics at Old Communist Czechoslovakia

Authors: Clarissa Valença Travassos da Silva

Abstract:

In this article, it is intended to study the book “The unbearable Lightness of Being” by the Czech Republican writer Milan Kundera. The main objective is to be an interdisciplinary study basis for students in the world about love and politics at old communist Czechoslovakia. Love is presented by discussing the relationship between Tomas and Tereza and the discovery of true love. Furthermore, it is debated the Russian invasion in Czechoslovakia and the outcomes of it for the personages, all this related to the contradiction of lightness and heaviness in life. For the production of this didactic material, the researcher based her work on the original book, “The Unbearable Lightness of Being” by Kundera, Milan Kundera’s interviews, Friedrich Nietzche, Zygmunt Bauman and George Orwell, among Brazilian and international articles on the issue.

Keywords: lightness, heaviness, Russia, Czechoslovakia, love

Procedia PDF Downloads 400
7414 Graph Clustering Unveiled: ClusterSyn - A Machine Learning Framework for Predicting Anti-Cancer Drug Synergy Scores

Authors: Babak Bahri, Fatemeh Yassaee Meybodi, Changiz Eslahchi

Abstract:

In the pursuit of effective cancer therapies, the exploration of combinatorial drug regimens is crucial to leverage synergistic interactions between drugs, thereby improving treatment efficacy and overcoming drug resistance. However, identifying synergistic drug pairs poses challenges due to the vast combinatorial space and limitations of experimental approaches. This study introduces ClusterSyn, a machine learning (ML)-powered framework for classifying anti-cancer drug synergy scores. ClusterSyn employs a two-step approach involving drug clustering and synergy score prediction using a fully connected deep neural network. For each cell line in the training dataset, a drug graph is constructed, with nodes representing drugs and edge weights denoting synergy scores between drug pairs. Drugs are clustered using the Markov clustering (MCL) algorithm, and vectors representing the similarity of drug pairs to each cluster are input into the deep neural network for synergy score prediction (synergy or antagonism). Clustering results demonstrate effective grouping of drugs based on synergy scores, aligning similar synergy profiles. Subsequently, neural network predictions and synergy scores of the two drugs on others within their clusters are used to predict the synergy score of the considered drug pair. This approach facilitates comparative analysis with clustering and regression-based methods, revealing the superior performance of ClusterSyn over state-of-the-art methods like DeepSynergy and DeepDDS on diverse datasets such as Oniel and Almanac. The results highlight the remarkable potential of ClusterSyn as a versatile tool for predicting anti-cancer drug synergy scores.

Keywords: drug synergy, clustering, prediction, machine learning., deep learning

Procedia PDF Downloads 79
7413 Projective Lag Synchronization in Drive-Response Dynamical Networks via Hybrid Feedback Control

Authors: Mohd Salmi Md Noorani, Ghada Al-Mahbashi, Sakhinah Abu Bakar

Abstract:

This paper investigates projective lag synchronization (PLS) behavior in drive response dynamical networks (DRDNs) model with identical nodes. A hybrid feedback control method is designed to achieve the PLS with mismatch and without mismatch terms. The stability of the error dynamics is proven theoretically using the Lyapunov stability theory. Finally, analytical results show that the states of the dynamical network with non-delayed coupling can be asymptotically synchronized onto a desired scaling factor under the designed controller. Moreover, the numerical simulations results demonstrate the validity of the proposed method.

Keywords: drive-response dynamical network, projective lag synchronization, hybrid feedback control, stability theory

Procedia PDF Downloads 391
7412 Using Biofunctool® Index to Assess Soil Quality after Eight Years of Conservation Agriculture in New Caledonia

Authors: Remy Kulagowski, Tobias Sturm, Audrey Leopold, Aurelie Metay, Josephine Peigne, Alexis Thoumazeau, Alain Brauman, Bruno Fogliani, Florent Tivet

Abstract:

A major challenge for agriculture is to enhance productivity while limiting the impact on the environment. Conservation agriculture (CA) is one strategy whereby both sustainability and productivity can be achieved by preserving and improving the soil quality. Soils provide and regulate a large number of ecosystem services (ES) such as agricultural productivity and climate change adaptation and mitigation. The aim of this study is to assess the impacts of contrasted CA crop management on soil functions for maize (Zea mays L.) cultivation in an eight years field experiment (2010-2018). The study included two CA practices: direct seeding in dead mulch (DM) and living mulch (LM), and conventional plough-based tillage (CT) practices on a fluvisol in New Caledonia (French Archipelago in the South Pacific). In 2018, soil quality of the cropping systems were evaluated with the Biofunctool® set of indicators, that consists in twelve integrative, in-field, and low-tech indicators assessing the biological, physical and chemical properties of soils. Main soil functions were evaluated including (i) carbon transformation, (ii) structure maintenance, and (iii) nutrient cycling in the ten first soil centimeters. The results showed significant higher score for soil structure maintenance (e.g., aggregate stability, water infiltration) and carbon transformation function (e.g., soil respiration, labile carbon) under CA in DM and LM when compared with CT. Score of carbon transformation index was higher in DM compared with LM. However, no significant effect of cropping systems was observed on nutrient cycling (i.e., nitrogen and phosphorus). In conclusion, the aggregated synthetic scores of soil multi-functions evaluated with Biofunctool® demonstrate that CA cropping systems lead to a better soil functioning. Further analysis of the results with agronomic performance of the soil-crop systems would allow to better understand the links between soil functioning and production ES of CA.

Keywords: conservation agriculture, cropping systems, ecosystem services, soil functions

Procedia PDF Downloads 157
7411 Simulation of Flow through Dam Foundation by FEM and ANN Methods Case Study: Shahid Abbaspour Dam

Authors: Mehrdad Shahrbanozadeh, Gholam Abbas Barani, Saeed Shojaee

Abstract:

In this study, a finite element (Seep3D model) and an artificial neural network (ANN) model were developed to simulate flow through dam foundation. Seep3D model is capable of simulating three-dimensional flow through a heterogeneous and anisotropic, saturated and unsaturated porous media. Flow through the Shahid Abbaspour dam foundation has been used as a case study. The FEM with 24960 triangular elements and 28707 nodes applied to model flow through foundation of this dam. The FEM being made denser in the neighborhood of the curtain screen. The ANN model developed for Shahid Abbaspour dam is a feedforward four layer network employing the sigmoid function as an activator and the back-propagation algorithm for the network learning. The water level elevations of the upstream and downstream of the dam have been used as input variables and the piezometric heads as the target outputs in the ANN model. The two models are calibrated and verified using the Shahid Abbaspour’s dam piezometric data. Results of the models were compared with those measured by the piezometers which are in good agreement. The model results also revealed that the ANN model performed as good as and in some cases better than the FEM.

Keywords: seepage, dam foundation, finite element method, neural network, seep 3D model

Procedia PDF Downloads 474
7410 Proteome-Wide Convergent Evolution on Vocal Learning Birds Reveals Insight into cAMP-Based Learning Pathway

Authors: Chul Lee, Seoae Cho, Erich D. Jarvis, Heebal Kim

Abstract:

Vocal learning, the ability to imitate vocalizations based on auditory experience, is a homoplastic character state observed in different independent lineages of animals such as songbirds, parrots, hummingbirds and human. It has now become possible to perform genome-wide molecular analyses across vocal learners and vocal non-learners with the recent expansion of avian genome data. It was analyzed the whole genomes of human and 48 avian species including those belonging to the three avian vocal learning lineages, to determine if behavior and neural convergence are associated with molecular convergence in divergent species of vocal learners. Analyses of 8295 orthologous genes across bird species revealed 141 genes with amino acid substitutions specific to vocal learners. Out of these, 25 genes have vocal learner specific genetic homoplasies, and their functions were enriched for learning. Several sites in these genes are estimated under convergent evolution and positive selection. A potential role for a subset of these genes in vocal learning was supported by associations with gene expression profiles in vocal learning brain regions of songbirds and human disease that cause language dysfunctions. The key candidate gene with multiple independent lines of the evidences specific to vocal learners was DRD5. Our findings suggest cAMP-based learning pathway in avian vocal learners, indicating molecular homoplastic changes associated with a complex behavioral trait, vocal learning.

Keywords: amino acid substitutions, convergent evolution, positive selection, vocal learning

Procedia PDF Downloads 341
7409 Comparative Analysis of Predictive Models for Customer Churn Prediction in the Telecommunication Industry

Authors: Deepika Christopher, Garima Anand

Abstract:

To determine the best model for churn prediction in the telecom industry, this paper compares 11 machine learning algorithms, namely Logistic Regression, Support Vector Machine, Random Forest, Decision Tree, XGBoost, LightGBM, Cat Boost, AdaBoost, Extra Trees, Deep Neural Network, and Hybrid Model (MLPClassifier). It also aims to pinpoint the top three factors that lead to customer churn and conducts customer segmentation to identify vulnerable groups. According to the data, the Logistic Regression model performs the best, with an F1 score of 0.6215, 81.76% accuracy, 68.95% precision, and 56.57% recall. The top three attributes that cause churn are found to be tenure, Internet Service Fiber optic, and Internet Service DSL; conversely, the top three models in this article that perform the best are Logistic Regression, Deep Neural Network, and AdaBoost. The K means algorithm is applied to establish and analyze four different customer clusters. This study has effectively identified customers that are at risk of churn and may be utilized to develop and execute strategies that lower customer attrition.

Keywords: attrition, retention, predictive modeling, customer segmentation, telecommunications

Procedia PDF Downloads 57
7408 Optimal Simultaneous Sizing and Siting of DGs and Smart Meters Considering Voltage Profile Improvement in Active Distribution Networks

Authors: T. Sattarpour, D. Nazarpour

Abstract:

This paper investigates the effect of simultaneous placement of DGs and smart meters (SMs), on voltage profile improvement in active distribution networks (ADNs). A substantial center of attention has recently been on responsive loads initiated in power system problem studies such as distributed generations (DGs). Existence of responsive loads in active distribution networks (ADNs) would have undeniable effect on sizing and siting of DGs. For this reason, an optimal framework is proposed for sizing and siting of DGs and SMs in ADNs. SMs are taken into consideration for the sake of successful implementing of demand response programs (DRPs) such as direct load control (DLC) with end-side consumers. Looking for voltage profile improvement, the optimization procedure is solved by genetic algorithm (GA) and tested on IEEE 33-bus distribution test system. Different scenarios with variations in the number of DG units, individual or simultaneous placing of DGs and SMs, and adaptive power factor (APF) mode for DGs to support reactive power have been established. The obtained results confirm the significant effect of DRPs and APF mode in determining the optimal size and site of DGs to be connected in ADN resulting to the improvement of voltage profile as well.

Keywords: active distribution network (ADN), distributed generations (DGs), smart meters (SMs), demand response programs (DRPs), adaptive power factor (APF)

Procedia PDF Downloads 302
7407 Multiobjective Economic Dispatch Using Optimal Weighting Method

Authors: Mandeep Kaur, Fatehgarh Sahib

Abstract:

The purpose of economic load dispatch is to allocate the required load demand between the available generation units such that the cost of operation is minimized. It is an optimization problem to find the most economical schedule of the generating units while satisfying load demand and operational constraints. The multiobjective optimization problem in which the engineer’s goal is to maximize or minimize not a single objective function but several objective functions simultaneously. The purpose of multiobjective problems in the mathematical programming framework is to optimize the different objective functions. Many approaches and methods have been proposed in recent years to solve multiobjective optimization problems. Weighting method has been applied to convert multiobjective optimization problems into scalar optimization. MATLAB 7.10 has been used to write the code for the complete algorithm with the help of genetic algorithm (GA). The validity of the proposed method has been demonstrated on a three-unit power system.

Keywords: economic load dispatch, genetic algorithm, generating units, multiobjective optimization, weighting method

Procedia PDF Downloads 150
7406 Multi Criteria Authentication Method in Cognitive Radio Networks

Authors: Shokoufeh Monjezi Kouchak

Abstract:

Cognitive radio network (CRN) is future network .Without this network wireless devices can’t work appropriately in the next decades. Today, wireless devices use static spectrum access methods and these methods don’t use spectrums optimum so we need use dynamic spectrum access methods to solve shortage spectrum challenge and CR is a great device for DSA but first of all its challenges should be solved .security is one of these challenges .In this paper we provided a survey about CR security. You can see this survey in tables 1 to 7 .After that we proposed a multi criteria authentication method in CRN. Our criteria in this method are: sensing results, following sending data rules, position of secondary users and no talk zone. Finally we compared our method with other authentication methods.

Keywords: authentication, cognitive radio, security, radio networks

Procedia PDF Downloads 392
7405 Optimized Cluster Head Selection Algorithm Based on LEACH Protocol for Wireless Sensor Networks

Authors: Wided Abidi, Tahar Ezzedine

Abstract:

Low-Energy Adaptive Clustering Hierarchy (LEACH) has been considered as one of the effective hierarchical routing algorithms that optimize energy and prolong the lifetime of network. Since the selection of Cluster Head (CH) in LEACH is carried out randomly, in this paper, we propose an approach of electing CH based on LEACH protocol. In other words, we present a formula for calculating the threshold responsible for CH election. In fact, we adopt three principle criteria: the remaining energy of node, the number of neighbors within cluster range and the distance between node and CH. Simulation results show that our proposed approach beats LEACH protocol in regards of prolonging the lifetime of network and saving residual energy.

Keywords: wireless sensors networks, LEACH protocol, cluster head election, energy efficiency

Procedia PDF Downloads 330
7404 Reconstructability Analysis for Landslide Prediction

Authors: David Percy

Abstract:

Landslides are a geologic phenomenon that affects a large number of inhabited places and are constantly being monitored and studied for the prediction of future occurrences. Reconstructability analysis (RA) is a methodology for extracting informative models from large volumes of data that work exclusively with discrete data. While RA has been used in medical applications and social science extensively, we are introducing it to the spatial sciences through applications like landslide prediction. Since RA works exclusively with discrete data, such as soil classification or bedrock type, working with continuous data, such as porosity, requires that these data are binned for inclusion in the model. RA constructs models of the data which pick out the most informative elements, independent variables (IVs), from each layer that predict the dependent variable (DV), landslide occurrence. Each layer included in the model retains its classification data as a primary encoding of the data. Unlike other machine learning algorithms that force the data into one-hot encoding type of schemes, RA works directly with the data as it is encoded, with the exception of continuous data, which must be binned. The usual physical and derived layers are included in the model, and testing our results against other published methodologies, such as neural networks, yields accuracy that is similar but with the advantage of a completely transparent model. The results of an RA session with a data set are a report on every combination of variables and their probability of landslide events occurring. In this way, every combination of informative state combinations can be examined.

Keywords: reconstructability analysis, machine learning, landslides, raster analysis

Procedia PDF Downloads 66
7403 Flexible 3D Virtual Desktop Using Handles for Cloud Environments

Authors: J. K. Lee, S. L. Lee

Abstract:

Due to the improvement in performance of computer hardware and the development of operating systems, a multi-tasking for several programs has become one of the basic functions to computer users. It is natural for computer users to want more functional, convenient, and visual GUI functions (Graphic User Interface). In this paper, a 3D virtual desktop system was proposed to meet users’ requirements for cloud environments such as a virtual desktop function in the Windows environment. The proposed system uses the handles of the windows to hide or restore several windows. It connects the list of task spaces using the circular double linked list to manage the handles. Each handle list is registered in the corresponding task space being executed. The 3D virtual desktop is efficient and flexible in handling the numbers of task spaces and can help users to work under more comfortable environments. Acknowledgment: This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Korea government (MSIP) (NRF-2015R1D1A1A01057680).

Keywords: virtual desktop, GUI, cloud, virtualization

Procedia PDF Downloads 210
7402 A Comparative Study of Environmental, Social and Economic Cross-Border Cooperation in Post-Conflict Environments: The Israel-Jordan Border

Authors: Tamar Arieli

Abstract:

Cross-border cooperation has long been hailed as a means for stabilizing and normalizing relations between former enemies. Cooperation in problem-solving and realizing of local interests in post-conflict environments can indeed serve as a basis for developing dialogue and meaningful relations between neighbors across borders. Hence the potential for formerly sealed borders to serve as a basis for generating local and national perceptions of interdependence and as a buffer against the resume of conflict. Central questions which arise for policy-makers and third parties are how to facilitate cross-border cooperation and which areas of cooperation best serve to normalize post-conflict border regions. The Israel-Jordan border functions as a post-conflict border, in that it is a peaceful border since the 1994 Israel-Jordan peace treaty yet cross-border relations are defined but the highly securitized nature of the border region and the ongoing Arab-Israel regional conflict. This case study is based on long term qualitative research carried out in the border regions of both Israel and Jordan, which mapped and analyzed cross-border in a wide range of activities – social interactions sponsored by peace-facilitating NGOs, government sponsored agricultural cooperation, municipal initiated emergency planning in cross-border continuous urban settings, private cross-border business ventures and various environmental cooperative initiatives. These cooperative initiatives are evaluated through multiple interviews carried out with initiators and partners in cross-border cooperation as well as analysis of documentation, funding and media. These cooperative interactions are compared based on levels of cross-border local and official awareness and involvement as well as sustainability over time. This research identifies environmental cooperation as the most sustainable area of cross- border cooperation and as most conducive to generating perceptions of regional interdependence. This is a variation to the ‘New Middle East’ vision of business-based cooperation leading to conflict amelioration and regional stability. Environmental cooperation serving the public good rather than personal profit enjoys social legitimization even in the face of widespread anti-normalization sentiments common in the post-conflict environment. This insight is examined in light of philosophical and social aspects of the natural environment and its social perceptions. This research has theoretical implications for better understanding dynamics of cooperation and conflict, as well as practical ramifications for practitioners in border region policy and management.

Keywords: borders, cooperation, post-conflict, security

Procedia PDF Downloads 314
7401 An Efficient Collocation Method for Solving the Variable-Order Time-Fractional Partial Differential Equations Arising from the Physical Phenomenon

Authors: Haniye Dehestani, Yadollah Ordokhani

Abstract:

In this work, we present an efficient approach for solving variable-order time-fractional partial differential equations, which are based on Legendre and Laguerre polynomials. First, we introduced the pseudo-operational matrices of integer and variable fractional order of integration by use of some properties of Riemann-Liouville fractional integral. Then, applied together with collocation method and Legendre-Laguerre functions for solving variable-order time-fractional partial differential equations. Also, an estimation of the error is presented. At last, we investigate numerical examples which arise in physics to demonstrate the accuracy of the present method. In comparison results obtained by the present method with the exact solution and the other methods reveals that the method is very effective.

Keywords: collocation method, fractional partial differential equations, legendre-laguerre functions, pseudo-operational matrix of integration

Procedia PDF Downloads 166