Search results for: optimization for learning and data analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15606

Search results for: optimization for learning and data analysis

12606 Online Forums Hotspot Detection and Analysis Using Aging Theory

Authors: K. Nirmala Devi, V. Murali Bhaskaran

Abstract:

The exponential growth of social media arouses much attention on public opinion information. The online forums, blogs, micro blogs are proving to be extremely valuable resources and are having bulk volume of information. However, most of the social media data is unstructured and semi structured form. So that it is more difficult to decipher automatically. Therefore, it is very much essential to understand and analyze those data for making a right decision. The online forums hotspot detection is a promising research field in the web mining and it guides to motivate the user to take right decision in right time. The proposed system consist of a novel approach to detect a hotspot forum for any given time period. It uses aging theory to find the hot terms and E-K-means for detecting the hotspot forum. Experimental results demonstrate that the proposed approach outperforms k-means for detecting the hotspot forums with the improved accuracy.

Keywords: Hotspot forums, Micro blog, Blog, Sentiment Analysis, Opinion Mining, Social media, Twitter, Web mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2161
12605 Convergence of ICT and Education

Authors: Raju Kumar

Abstract:

Information and communication technology (ICT) has become, within a very short time, one of the basic building blocks of modern society. Many countries now understanding the importance of ICT and mastering the basic skills and concepts of it as part of the core of education. Organizations, experts and practitioners in the education sector increasingly recognizing the importance of ICT in supporting educational improvement and reform. This paper addresses the convergence of ICT and education. When two technologies are converging to each other, together they will generate some great opportunities and challenges. This paper focuses on these issues. In introduction section, it explains the ICT, education, and ICT-enhanced education. In next section it describes need of ICT in education, relationship between ICT skills and education, and stages of teaching learning process. The next two sections describe opportunities and challenges in integrating ICT in education. Finally the concluding section summaries the idea and its usefulness.

Keywords: Education, Information and CommunicationTechnology, Learning, Teaching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3565
12604 Review of the Model-Based Supply Chain Management Research in the Construction Industry

Authors: Aspasia Koutsokosta, Stefanos Katsavounis

Abstract:

This paper reviews the model-based qualitative and quantitative Operations Management research in the context of Construction Supply Chain Management (CSCM). Construction industry has been traditionally blamed for low productivity, cost and time overruns, waste, high fragmentation and adversarial relationships. The construction industry has been slower than other industries to employ the Supply Chain Management (SCM) concept and develop models that support the decision-making and planning. However the last decade there is a distinct shift from a project-based to a supply-based approach of construction management. CSCM comes up as a new promising management tool of construction operations and improves the performance of construction projects in terms of cost, time and quality. Modeling the Construction Supply Chain (CSC) offers the means to reap the benefits of SCM, make informed decisions and gain competitive advantage. Different modeling approaches and methodologies have been applied in the multi-disciplinary and heterogeneous research field of CSCM. The literature review reveals that a considerable percentage of the CSC modeling research accommodates conceptual or process models which present general management frameworks and do not relate to acknowledged soft Operations Research methods. We particularly focus on the model-based quantitative research and categorize the CSCM models depending on their scope, objectives, modeling approach, solution methods and software used. Although over the last few years there has been clearly an increase of research papers on quantitative CSC models, we identify that the relevant literature is very fragmented with limited applications of simulation, mathematical programming and simulation-based optimization. Most applications are project-specific or study only parts of the supply system. Thus, some complex interdependencies within construction are neglected and the implementation of the integrated supply chain management is hindered. We conclude this paper by giving future research directions and emphasizing the need to develop optimization models for integrated CSCM. We stress that CSC modeling needs a multi-dimensional, system-wide and long-term perspective. Finally, prior applications of SCM to other industries have to be taken into account in order to model CSCs, but not without translating the generic concepts to the context of construction industry.

Keywords: Construction supply chain management, modeling, operations research, optimization and simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2807
12603 2D Numerical Analysis of Sao Paulo Tunnel

Authors: A.H. Akhaveissy

Abstract:

Nonlinear finite element method and Serendipity eight nodes element are used for determining of ground surface settlement due to tunneling. Linear element with elastic behavior is used for modeling of lining. Modified Generalized plasticity model with nonassociated flow rule is applied for analysis of a tunnel in Sao Paulo – Brazil. The tunnel had analyzed by Lades- model with 16 parameters. In this work modified Generalized Plasticity is used with 10 parameters, also Mohr-Coulomb model is used to analysis the tunnel. The results show good agreement with observed results of field data by modified Generalized Plasticity model than other models. The obtained result by Mohr-Coulomb model shows less settlement than other model due to excavation.

Keywords: Non-associated flow rule, Generalized plasticity, tunnel excavation, Excavation method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2605
12602 Web-Content Analysis of the Major Spanish Tourist Destinations Evaluation by Russian Tourists

Authors: Natalia Polkanova, Sergey Kazakov

Abstract:

In the second decade of the XXI century the role of tourism destination attractiveness is becoming increasingly important for destination management. Competition in tourism market moves from ordinary service quality to provision of unforgettable emotional experience for tourists. The main purpose of the present study is to identify the perception of the tourism destinations based on the number of factors related to its tourist attractiveness. The content analysis method was used to analyze the on-line tourist feedback data immensely available in Social Media and in travel related sites. The collected data made it possible to procure the information which is necessary to understand the perceived attractiveness of the destinations and key destination appeal factors that are important for Russian leisure travelers. Results of the present study demonstrate key attractiveness factors or destination ‘properties’ that were unveiled as the most important for Russian leisure tourists. The study targeted five main Spanish tourism destinations that initially were determined by in-depth interview with a number of Russian nationals who had visited Spain at least once. The research results can be useful for Spanish Tourism Organization Representation office in Russia as well as for the other national tourism organizations in order to promote their respective destinations for Russian travelers focusing on main attractiveness factors identified in this study.

Keywords: Tourism destination, destination attractiveness, destination competitiveness, content analysis, unstructured image.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2570
12601 Improved Data Warehousing: Lessons Learnt from the Systems Approach

Authors: Roelien Goede

Abstract:

Data warehousing success is not high enough. User dissatisfaction and failure to adhere to time frames and budgets are too common. Most traditional information systems practices are rooted in hard systems thinking. Today, the great systems thinkers are forgotten by information systems developers. A data warehouse is still a system and it is worth investigating whether systems thinkers such as Churchman can enhance our practices today. This paper investigates data warehouse development practices from a systems thinking perspective. An empirical investigation is done in order to understand the everyday practices of data warehousing professionals from a systems perspective. The paper presents a model for the application of Churchman-s systems approach in data warehouse development.

Keywords: Data warehouse development, Information systemsdevelopment, Interpretive case study, Systems thinking

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1570
12600 Centralized Resource Management for Network Infrastructure Including Ip Telephony by Integrating a Mediator Between the Heterogeneous Data Sources

Authors: Mohammed Fethi Khalfi, Malika Kandouci

Abstract:

Over the past decade, mobile has experienced a revolution that will ultimately change the way we communicate.All these technologies have a common denominator exploitation of computer information systems, but their operation can be tedious because of problems with heterogeneous data sources.To overcome the problems of heterogeneous data sources, we propose to use a technique of adding an extra layer interfacing applications of management or supervision at the different data sources.This layer will be materialized by the implementation of a mediator between different host applications and information systems frequently used hierarchical and relational manner such that the heterogeneity is completely transparent to the VoIP platform.

Keywords: TOIP, Data Integration, Mediation, informationcomputer system, heterogeneous data sources

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1309
12599 Optimization of Process Parameters for Friction Stir Welding of Cast Alloy AA7075 by Taguchi Method

Authors: Dhairya Partap Sing, Vikram Singh, Sudhir Kumar

Abstract:

This investigation proposes Friction stir welding technique to solve the fusion welding problems. Objectives of this investigation are fabrication of AA7075-10%wt. Silicon carbide (SiC) aluminum metal matrix composite and optimization of optimal process parameters of friction stir welded AA7075-10%wt. SiC Composites. Composites were prepared by the mechanical stir casting process. Experiments were performed with four process parameters such as tool rotational speed, weld speed, axial force and tool geometry considering three levels of each. The quality characteristics considered is joint efficiency (JE). The welding experiments were conducted using L27 orthogonal array. An orthogonal array and design of experiments were used to give best possible welding parameters that give optimal JE. The fabricated welded joints using rotational speed of 1500 rpm, welding speed (1.3 mm/sec), axial force (7 k/n) of and tool geometry (square) give best possible results. Experimental result reveals that the tool rotation speed, welding speed and axial force are the significant process parameters affecting the welding performance. The predicted optimal value of percentage JE is 95.621. The confirmation tests also have been done for verifying the results.

Keywords: Metal matrix composite, axial force, joint efficiency, rotational speed, traverse speed, tool geometry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 845
12598 Secure Multiparty Computations for Privacy Preserving Classifiers

Authors: M. Sumana, K. S. Hareesha

Abstract:

Secure computations are essential while performing privacy preserving data mining. Distributed privacy preserving data mining involve two to more sites that cannot pool in their data to a third party due to the violation of law regarding the individual. Hence in order to model the private data without compromising privacy and information loss, secure multiparty computations are used. Secure computations of product, mean, variance, dot product, sigmoid function using the additive and multiplicative homomorphic property is discussed. The computations are performed on vertically partitioned data with a single site holding the class value.

Keywords: Homomorphic property, secure product, secure mean and variance, secure dot product, vertically partitioned data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 904
12597 SWOT Analysis of Cassava Sector in Cameroon

Authors: Elise Stephanie Mvodo Meyo, Dapeng Liang

Abstract:

Cassava is one of the top five crops in Cameroon. Its evolution has remained constant since the independence period and the production has more than tripled. It is a crop with multiple industrial capacities but the sector-s business opportunities are underexploited. Using Strengths, Weaknesses, Opportunities and Threats analysis method, this paper examines the cassava actual state. It appraises the sector-s strengths (S), considers suitable measures to strengthen weaknesses (W), evaluates strategies to fully benefit from the sector numerous business opportunities (O) and explore means to convert threats (T) into opportunities. Data were collected from the ministry of agriculture and rural development and different actors. The results show that cassava sector embodies many business opportunities and stands as a raw material provider for many industries but ultimately requires challenges to be tackled appropriately.

Keywords: Business opportunities, cassava sector, rural development, SWOT analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7228
12596 Geostatistical Analysis and Mapping of Groundlevel Ozone in a Medium Sized Urban Area

Authors: F. J. Moral García, P. Valiente González, F. López Rodríguez

Abstract:

Ground-level tropospheric ozone is one of the air pollutants of most concern. It is mainly produced by photochemical processes involving nitrogen oxides and volatile organic compounds in the lower parts of the atmosphere. Ozone levels become particularly high in regions close to high ozone precursor emissions and during summer, when stagnant meteorological conditions with high insolation and high temperatures are common. In this work, some results of a study about urban ozone distribution patterns in the city of Badajoz, which is the largest and most industrialized city in Extremadura region (southwest Spain) are shown. Fourteen sampling campaigns, at least one per month, were carried out to measure ambient air ozone concentrations, during periods that were selected according to favourable conditions to ozone production, using an automatic portable analyzer. Later, to evaluate the ozone distribution at the city, the measured ozone data were analyzed using geostatistical techniques. Thus, first, during the exploratory analysis of data, it was revealed that they were distributed normally, which is a desirable property for the subsequent stages of the geostatistical study. Secondly, during the structural analysis of data, theoretical spherical models provided the best fit for all monthly experimental variograms. The parameters of these variograms (sill, range and nugget) revealed that the maximum distance of spatial dependence is between 302-790 m and the variable, air ozone concentration, is not evenly distributed in reduced distances. Finally, predictive ozone maps were derived for all points of the experimental study area, by use of geostatistical algorithms (kriging). High prediction accuracy was obtained in all cases as cross-validation showed. Useful information for hazard assessment was also provided when probability maps, based on kriging interpolation and kriging standard deviation, were produced.

Keywords: Kriging, map, tropospheric ozone, variogram.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1840
12595 Segmenting Ultrasound B-Mode Images Using RiIG Distributions and Stochastic Optimization

Authors: N. Mpofu, M. Sears

Abstract:

In this paper, we propose a novel algorithm for delineating the endocardial wall from a human heart ultrasound scan. We assume that the gray levels in the ultrasound images are independent and identically distributed random variables with different Rician Inverse Gaussian (RiIG) distributions. Both synthetic and real clinical data will be used for testing the algorithm. Algorithm performance will be evaluated using the expert radiologist evaluation of a soft copy of an ultrasound scan during the scanning process and secondly, doctor’s conclusion after going through a printed copy of the same scan. Successful implementation of this algorithm should make it possible to differentiate normal from abnormal soft tissue and help disease identification, what stage the disease is in and how best to treat the patient. We hope that an automated system that uses this algorithm will be useful in public hospitals especially in Third World countries where problems such as shortage of skilled radiologists and shortage of ultrasound machines are common. These public hospitals are usually the first and last stop for most patients in these countries.

Keywords: Endorcardial Wall, Rician Inverse Distributions, Segmentation, Ultrasound Images.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1558
12594 Design of Buffer Management for Industry to Avoid Sensor Data- Conflicts

Authors: Dae-ho Won, Jong-wook Hong, Yeon-Mo Yang, Jinung An

Abstract:

To reduce accidents in the industry, WSNs(Wireless Sensor networks)- sensor data is used. WSNs- sensor data has the persistence and continuity. therefore, we design and exploit the buffer management system that has the persistence and continuity to avoid and delivery data conflicts. To develop modules, we use the multi buffers and design the buffer management modules that transfer sensor data through the context-aware methods.

Keywords: safe management system, buffer management, context-aware, input data stream

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535
12593 Design and Simulation of a New Self-Learning Expert System for Mobile Robot

Authors: Rabi W. Yousif, Mohd Asri Hj Mansor

Abstract:

In this paper, we present a novel technique called Self-Learning Expert System (SLES). Unlike Expert System, where there is a need for an expert to impart experiences and knowledge to create the knowledge base, this technique tries to acquire the experience and knowledge automatically. To display this technique at work, a simulation of a mobile robot navigating through an environment with obstacles is employed using visual basic. The mobile robot will move through this area without colliding with any obstacle and save the path that it took. If the mobile robot has to go through a similar environment again, then it will apply this experience to help it move through quicker without having to check for collision.

Keywords: Expert system, knowledge base, mobile robot, visual basic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1510
12592 Security in Resource Constraints Network Light Weight Encryption for Z-MAC

Authors: Mona Almansoori, Ahmed Mustafa, Ahmad Elshamy

Abstract:

Wireless sensor network was formed by a combination of nodes, systematically it transmitting the data to their base stations, this transmission data can be easily compromised if the limited processing power and the data consistency from these nodes are kept in mind; there is always a discussion to address the secure data transfer or transmission in actual time. This will present a mechanism to securely transmit the data over a chain of sensor nodes without compromising the throughput of the network by utilizing available battery resources available in the sensor node. Our methodology takes many different advantages of Z-MAC protocol for its efficiency, and it provides a unique key by sharing the mechanism using neighbor node MAC address. We present a light weighted data integrity layer which is embedded in the Z-MAC protocol to prove that our protocol performs well than Z-MAC when we introduce the different attack scenarios.

Keywords: Hybrid MAC protocol, data integrity, lightweight encryption, Neighbor based key sharing, Sensor node data processing, Z-MAC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 534
12591 Data Recording for Remote Monitoring of Autonomous Vehicles

Authors: Rong-Terng Juang

Abstract:

Autonomous vehicles offer the possibility of significant benefits to social welfare. However, fully automated cars might not be going to happen in the near further. To speed the adoption of the self-driving technologies, many governments worldwide are passing laws requiring data recorders for the testing of autonomous vehicles. Currently, the self-driving vehicle, (e.g., shuttle bus) has to be monitored from a remote control center. When an autonomous vehicle encounters an unexpected driving environment, such as road construction or an obstruction, it should request assistance from a remote operator. Nevertheless, large amounts of data, including images, radar and lidar data, etc., have to be transmitted from the vehicle to the remote center. Therefore, this paper proposes a data compression method of in-vehicle networks for remote monitoring of autonomous vehicles. Firstly, the time-series data are rearranged into a multi-dimensional signal space. Upon the arrival, for controller area networks (CAN), the new data are mapped onto a time-data two-dimensional space associated with the specific CAN identity. Secondly, the data are sampled based on differential sampling. Finally, the whole set of data are encoded using existing algorithms such as Huffman, arithmetic and codebook encoding methods. To evaluate system performance, the proposed method was deployed on an in-house built autonomous vehicle. The testing results show that the amount of data can be reduced as much as 1/7 compared to the raw data.

Keywords: Autonomous vehicle, data recording, remote monitoring, controller area network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1320
12590 Multivariate Statistical Analysis of Decathlon Performance Results in Olympic Athletes (1988-2008)

Authors: Jaebum Park, Vladimir M. Zatsiorsky

Abstract:

The performance results of the athletes competed in the 1988-2008 Olympic Games were analyzed (n = 166). The data were obtained from the IAAF official protocols. In the principal component analysis, the first three principal components explained 70% of the total variance. In the 1st principal component (with 43.1% of total variance explained) the largest factor loadings were for 100m (0.89), 400m (0.81), 110m hurdle run (0.76), and long jump (–0.72). This factor can be interpreted as the 'sprinting performance'. The loadings on the 2nd factor (15.3% of the total variance) presented a counter-intuitive throwing-jumping combination: the highest loadings were for throwing events (javelin throwing 0.76; shot put 0.74; and discus throwing 0.73) and also for jumping events (high jump 0.62; pole vaulting 0.58). On the 3rd factor (11.6% of total variance), the largest loading was for 1500 m running (0.88); all other loadings were below 0.4.

Keywords: Decathlon, principal component analysis, Olympic Games, multivariate statistical analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2788
12589 Chaos Theory and Application in Foreign Exchange Rates vs. IRR (Iranian Rial)

Authors: M. A. Torkamani, S. Mahmoodzadeh, S. Pourroostaei, C. Lucas

Abstract:

Daily production of information and importance of the sequence of produced data in forecasting future performance of market causes analysis of data behavior to become a problem of analyzing time series. But time series that are very complicated, usually are random and as a result their changes considered being unpredictable. While these series might be products of a deterministic dynamical and nonlinear process (chaotic) and as a result be predictable. Point of Chaotic theory view, complicated systems have only chaotically face and as a result they seem to be unregulated and random, but it is possible that they abide by a specified math formula. In this article, with regard to test of strange attractor and biggest Lyapunov exponent probability of chaos on several foreign exchange rates vs. IRR (Iranian Rial) has been investigated. Results show that data in this market have complex chaotic behavior with big degree of freedom.

Keywords: Chaos, Exchange Rate, Nonlinear Models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2459
12588 Anomaly Detection and Characterization to Classify Traffic Anomalies Case Study: TOT Public Company Limited Network

Authors: O. Siriporn, S. Benjawan

Abstract:

This paper represents four unsupervised clustering algorithms namely sIB, RandomFlatClustering, FarthestFirst, and FilteredClusterer that previously works have not been used for network traffic classification. The methodology, the result, the products of the cluster and evaluation of these algorithms with efficiency of each algorithm from accuracy are shown. Otherwise, the efficiency of these algorithms considering form the time that it use to generate the cluster quickly and correctly. Our work study and test the best algorithm by using classify traffic anomaly in network traffic with different attribute that have not been used before. We analyses the algorithm that have the best efficiency or the best learning and compare it to the previously used (K-Means). Our research will be use to develop anomaly detection system to more efficiency and more require in the future.

Keywords: Unsupervised, clustering, anomaly, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2089
12587 Optimization of Material Removal Rate in Electrical Discharge Machining Using Fuzzy Logic

Authors: Amit Kohli, Aashim Wadhwa, Tapan Virmani, Ujjwal Jain

Abstract:

The objective of present work is to stimulate the machining of material by electrical discharge machining (EDM) to give effect of input parameters like discharge current (Ip), pulse on time (Ton), pulse off time (Toff) which can bring about changes in the output parameter, i.e. material removal rate. Experimental data was gathered from die sinking EDM process using copper electrode and Medium Carbon Steel (AISI 1040) as work-piece. The rules of membership function (MF) and the degree of closeness to the optimum value of the MMR are within the upper and lower range of the process parameters. It was found that proposed fuzzy model is in close agreement with the experimental results. By Intelligent, model based design and control of EDM process parameters in this study will help to enable dramatically decreased product and process development cycle times.

Keywords: Electrical discharge Machining (EDM), Fuzzy Logic, Material removal rate (MRR), Membership functions (MF).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2727
12586 Development of a New Method for T-joint Specimens Testing under Shear Loading

Authors: R. Doubrava, R. Růžek

Abstract:

Nonstandard tests are necessary for analyses and verification of new developed structural and technological solutions with application of composite materials. One of the most critical primary structural parts of a typical aerospace structure is T-joint. This structural element is loaded mainly in shear, bending, peel and tension. The paper is focused on the shear loading simulations. The aim of the work is to obtain a representative uniform distribution of shear loads along T-joint during the mechanical testing. A new design of T-joint test procedure, numerical simulation and optimization of representative boundary conditions are presented. The different conditions and inaccuracies both in simulations and experiments are discussed. The influence of different parameters on stress and strain distributions is demonstrated on T-joint made of CFRP (carbon fibre reinforced plastic). A special test rig designed by VZLU (Aerospace Research and Test Establishment) for T-shear test procedure is presented.

Keywords: T-joint, shear, composite, mechanical testing, Finite Element analysis, methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2637
12585 Analysis of Possible Causes of Fukushima Disaster

Authors: Abid Hossain Khan, Syam Hasan, M. A. R. Sarkar

Abstract:

Fukushima disaster is one of the most publicly exposed accidents in a nuclear facility which has changed the outlook of people towards nuclear power. Some have used it as an example to establish nuclear energy as an unsafe source, while others have tried to find the real reasons behind this accident. Many papers have tried to shed light on the possible causes, some of which are purely based on assumptions while others rely on rigorous data analysis. To our best knowledge, none of the works can say with absolute certainty that there is a single prominent reason that has paved the way to this unexpected incident. This paper attempts to compile all the apparent reasons behind Fukushima disaster and tries to analyze and identify the most likely one.

Keywords: Fuel meltdown, Fukushima disaster, manmade calamity, nuclear facility, tsunami.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2147
12584 Using Genetic Algorithms to Outline Crop Rotations and a Cropping-System Model

Authors: Nicolae Bold, Daniel Nijloveanu

Abstract:

The idea of cropping-system is a method used by farmers. It is an environmentally-friendly method, protecting the natural resources (soil, water, air, nutritive substances) and increase the production at the same time, taking into account some crop particularities. The combination of this powerful method with the concepts of genetic algorithms results into a possibility of generating sequences of crops in order to form a rotation. The usage of this type of algorithms has been efficient in solving problems related to optimization and their polynomial complexity allows them to be used at solving more difficult and various problems. In our case, the optimization consists in finding the most profitable rotation of cultures. One of the expected results is to optimize the usage of the resources, in order to minimize the costs and maximize the profit. In order to achieve these goals, a genetic algorithm was designed. This algorithm ensures the finding of several optimized solutions of cropping-systems possibilities which have the highest profit and, thus, which minimize the costs. The algorithm uses genetic-based methods (mutation, crossover) and structures (genes, chromosomes). A cropping-system possibility will be considered a chromosome and a crop within the rotation is a gene within a chromosome. Results about the efficiency of this method will be presented in a special section. The implementation of this method would bring benefits into the activity of the farmers by giving them hints and helping them to use the resources efficiently.

Keywords: Genetic algorithm, chromosomes, genes, cropping, agriculture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587
12583 Prediction Heating Values of Lignocellulosics from Biomass Characteristics

Authors: Kaltima Phichai, Pornchanoke Pragrobpondee, Thaweesak Khumpart, Samorn Hirunpraditkoon

Abstract:

The paper provides biomasses characteristics by proximate analysis (volatile matter, fixed carbon and ash) and ultimate analysis (carbon, hydrogen, nitrogen and oxygen) for the prediction of the heating value equations. The heating value estimation of various biomasses can be used as an energy evaluation. Thirteen types of biomass were studied. Proximate analysis was investigated by mass loss method and infrared moisture analyzer. Ultimate analysis was analyzed by CHNO analyzer. The heating values varied from 15 to 22.4MJ kg-1. Correlations of the calculated heating value with proximate and ultimate analyses were undertaken using multiple regression analysis and summarized into three and two equations, respectively. Correlations based on proximate analysis illustrated that deviation of calculated heating values from experimental heating values was higher than the correlations based on ultimate analysis.

Keywords: Heating value equation, Proximate analysis, Ultimate analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3685
12582 A Reasoning Method of Cyber-Attack Attribution Based on Threat Intelligence

Authors: Li Qiang, Yang Ze-Ming, Liu Bao-Xu, Jiang Zheng-Wei

Abstract:

With the increasing complexity of cyberspace security, the cyber-attack attribution has become an important challenge of the security protection systems. The difficult points of cyber-attack attribution were forced on the problems of huge data handling and key data missing. According to this situation, this paper presented a reasoning method of cyber-attack attribution based on threat intelligence. The method utilizes the intrusion kill chain model and Bayesian network to build attack chain and evidence chain of cyber-attack on threat intelligence platform through data calculation, analysis and reasoning. Then, we used a number of cyber-attack events which we have observed and analyzed to test the reasoning method and demo system, the result of testing indicates that the reasoning method can provide certain help in cyber-attack attribution.

Keywords: Reasoning, Bayesian networks, cyber-attack attribution, kill chain, threat intelligence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2650
12581 Issues in Organizational Assessment: The Case of Frustration Tolerance Measurement in Mexico

Authors: David Ruiz, Carlos Nava, Roberto Carbajal

Abstract:

The psychological profile has become one of the most important sources of information when it comes to individual selection and the hiring process in any organization. Psychological instruments are used to collect data about variables that are considered critically important for performance in work. However, because of conceptual chaos in organizational psychology, most of the information provided by psychological testing is not directly useful for Mexican human resources professionals to take hiring decisions. The aims of this paper are 1) to underline the lack of conceptual precision in theoretical testing foundations in Mexico and 2) presenting a reliability and validity analysis of a frustration tolerance instrument created as an alternative to a heuristically conduct individual assessment in organizations. First, a description of assessment conditions in Mexico is made. Second, an instrument and a theoretical framework is presented as an alternative to the assessment practices in the country. A total of 65 Psychology Iztacala Superior Studies Faculty students were assessed. Cronbach´s alpha coefficient was calculated and an exploratory factor analysis was carried out to prove the scale unidimensionality. Reliability analysis revealed good internal consistency of the scale (Cronbach’s α = 0.825). Factor analysis produced 4 factors for the scale. However, factor loadings and explained variation give proof to the scale unidimensionality. It is concluded that the instrument has good psychometric properties that will allow human resources professionals to collect useful data. Different possibilities to conduct psychological assessment are suggested for future development.

Keywords: Psychological assessment, frustration tolerance, human resources, organizational psychology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1889
12580 Online Topic Model for Broadcasting Contents Using Semantic Correlation Information

Authors: Chang-Uk Kwak, Sun-Joong Kim, Seong-Bae Park, Sang-Jo Lee

Abstract:

This paper proposes a method of learning topics for broadcasting contents. There are two kinds of texts related to broadcasting contents. One is a broadcasting script, which is a series of texts including directions and dialogues. The other is blogposts, which possesses relatively abstracted contents, stories, and diverse information of broadcasting contents. Although two texts range over similar broadcasting contents, words in blogposts and broadcasting script are different. When unseen words appear, it needs a method to reflect to existing topic. In this paper, we introduce a semantic vocabulary expansion method to reflect unseen words. We expand topics of the broadcasting script by incorporating the words in blogposts. Each word in blogposts is added to the most semantically correlated topics. We use word2vec to get the semantic correlation between words in blogposts and topics of scripts. The vocabularies of topics are updated and then posterior inference is performed to rearrange the topics. In experiments, we verified that the proposed method can discover more salient topics for broadcasting contents.

Keywords: Broadcasting script analysis, topic expansion, semantic correlation analysis, word2vec.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1742
12579 Cloud Computing Databases: Latest Trends and Architectural Concepts

Authors: Tarandeep Singh, Parvinder S. Sandhu

Abstract:

The Economic factors are leading to the rise of infrastructures provides software and computing facilities as a service, known as cloud services or cloud computing. Cloud services can provide efficiencies for application providers, both by limiting up-front capital expenses, and by reducing the cost of ownership over time. Such services are made available in a data center, using shared commodity hardware for computation and storage. There is a varied set of cloud services available today, including application services (salesforce.com), storage services (Amazon S3), compute services (Google App Engine, Amazon EC2) and data services (Amazon SimpleDB, Microsoft SQL Server Data Services, Google-s Data store). These services represent a variety of reformations of data management architectures, and more are on the horizon.

Keywords: Data Management in Cloud, AWS, EC2, S3, SQS, TQG.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1964
12578 Searching the Efficient Frontier for the Coherent Covering Location Problem

Authors: Felipe Azocar Simonet, Luis Acosta Espejo

Abstract:

In this article, we will try to find an efficient boundary approximation for the bi-objective location problem with coherent coverage for two levels of hierarchy (CCLP). We present the mathematical formulation of the model used. Supported efficient solutions and unsupported efficient solutions are obtained by solving the bi-objective combinatorial problem through the weights method using a Lagrangean heuristic. Subsequently, the results are validated through the DEA analysis with the GEM index (Global efficiency measurement).

Keywords: Coherent covering location problem, efficient frontier, Lagrangian relaxation, data envelopment analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 790
12577 Comparing Field Displacement History with Numerical Results to Estimate Geotechnical Parameters: Case Study of Arash-Esfandiar-Niayesh under Passing Tunnel, 2.5 Traffic Lane Tunnel, Tehran, Iran

Authors: A. Golshani, M. Gharizade Varnusefaderani, S. Majidian

Abstract:

Underground structures are of those structures that have uncertainty in design procedures. That is due to the complexity of soil condition around. Under passing tunnels are also such affected structures. Despite geotechnical site investigations, lots of uncertainties exist in soil properties due to unknown events. As results, it possibly causes conflicting settlements in numerical analysis with recorded values in the project. This paper aims to report a case study on a specific under passing tunnel constructed by New Austrian Tunnelling Method in Iran. The intended tunnel has an overburden of about 11.3m, the height of 12.2m and, the width of 14.4m with 2.5 traffic lane. The numerical modeling was developed by a 2D finite element program (PLAXIS Version 8). Comparing displacement histories at the ground surface during the entire installation of initial lining, the estimated surface settlement was about four times the field recorded one, which indicates that some local unknown events affect that value. Also, the displacement ratios were in a big difference between the numerical and field data. Consequently, running several numerical back analyses using laboratory and field tests data, the geotechnical parameters were accurately revised to match with the obtained monitoring data. Finally, it was found that usually the values of soil parameters are conservatively low-estimated up to 40 percent by typical engineering judgment. Additionally, it could be attributed to inappropriate constitutive models applied for the specific soil condition.

Keywords: NATM, surface displacement history, soil tests, monitoring data, numerical back-analysis, geotechnical parameters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 769