Search results for: Integral Max Flow Problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5873

Search results for: Integral Max Flow Problem

1283 A Study on Finding Similar Document with Multiple Categories

Authors: R. Saraçoğlu, N. Allahverdi

Abstract:

Searching similar documents and document management subjects have important place in text mining. One of the most important parts of similar document research studies is the process of classifying or clustering the documents. In this study, a similar document search approach that includes discussion of out the case of belonging to multiple categories (multiple categories problem) has been carried. The proposed method that based on Fuzzy Similarity Classification (FSC) has been compared with Rocchio algorithm and naive Bayes method which are widely used in text mining. Empirical results show that the proposed method is quite successful and can be applied effectively. For the second stage, multiple categories vector method based on information of categories regarding to frequency of being seen together has been used. Empirical results show that achievement is increased almost two times, when proposed method is compared with classical approach.

Keywords: Document similarity, Fuzzy classification, Multiple categories, Text mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1707
1282 Investigation of the Main Trends of Tourist Expenses in Georgia

Authors: Nino Abesadze, Marine Mindorashvili, Nino Paresashvili

Abstract:

The main purpose of the article is to make complex statistical analysis of tourist expenses of foreign visitors. We used mixed technique of selection that implies rules of random and proportional selection. Computer software SPSS was used to compute statistical data for corresponding analysis. Corresponding methodology of tourism statistics was implemented according to international standards. Important information was collected and grouped from the major Georgian airports. Techniques of statistical observation were prepared. A representative population of foreign visitors and a rule of selection of respondents were determined. We have a trend of growth of tourist numbers and share of tourists from post-soviet countries constantly increases. Level of satisfaction with tourist facilities and quality of service has grown, but still we have a problem of disparity between quality of service and prices. The design of tourist expenses of foreign visitors is diverse; competitiveness of tourist products of Georgian tourist companies is higher.

Keywords: Tourist, expenses, methods, statistics, analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 948
1281 A Weighted Sum Technique for the Joint Optimization of Performance and Power Consumption in Data Centers

Authors: Samee Ullah Khan, C.Ardil

Abstract:

With data centers, end-users can realize the pervasiveness of services that will be one day the cornerstone of our lives. However, data centers are often classified as computing systems that consume the most amounts of power. To circumvent such a problem, we propose a self-adaptive weighted sum methodology that jointly optimizes the performance and power consumption of any given data center. Compared to traditional methodologies for multi-objective optimization problems, the proposed self-adaptive weighted sum technique does not rely on a systematical change of weights during the optimization procedure. The proposed technique is compared with the greedy and LR heuristics for large-scale problems, and the optimal solution for small-scale problems implemented in LINDO. the experimental results revealed that the proposed selfadaptive weighted sum technique outperforms both of the heuristics and projects a competitive performance compared to the optimal solution.

Keywords: Meta-heuristics, distributed systems, adaptive methods, resource allocation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1836
1280 Design and Sensitivity Analysis of Photovoltaic/Thermal Solar Collector

Authors: H. M. Farghally, N. M. Ahmed, H. T. El-Madany, D. M. Atia, F. H. Fahmy

Abstract:

Energy is required in almost every aspect of human activities and development of any nation in the world. Increasing fossil fuel price, energy security and climate change have important bearings on sustainable development of any nation. The renewable energy technology is considered one of the drastic approaches which taken over the world to reduce the energy problem. The preservation of vegetables by freezing is one of the most important methods of retaining quality in agricultural products over long-term storage periods. Freezing factories show high demand of energy for both heat and electricity; the hybrid Photovoltaic/Thermal (PV/T) systems could be used in order to meet this requirement. This paper presents PV/T system design for freezing factory. Also, the complete mathematical modeling and MATLAB SIMULINK of PV/T collector is introduced. The sensitivity analysis for the manufacturing parameters of PV/T collector is carried out to study their effect on both thermal and electrical efficiency.

Keywords: Renewable energy, Hybrid PV/T system, Sensitivity analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3894
1279 An MCDM Approach to Selection Scheduling Rule in Robotic Flexibe Assembly Cells

Authors: Khalid Abd, Kazem Abhary, Romeo Marian

Abstract:

Multiple criteria decision making (MCDM) is an approach to ranking the solutions and finding the best one when two or more solutions are provided. In this study, MCDM approach is proposed to select the most suitable scheduling rule of robotic flexible assembly cells (RFACs). Two MCDM approaches, Analytic Hierarchy Process (AHP) and Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) are proposed for solving the scheduling rule selection problem. The AHP method is employed to determine the weights of the evaluation criteria, while the TOPSIS method is employed to obtain final ranking order of scheduling rules. Four criteria are used to evaluate the scheduling rules. Also, four scheduling policies of RFAC are examined to choose the most appropriate one for this purpose. A numerical example illustrates applications of the suggested methodology. The results show that the methodology is practical and works in RFAC settings.

Keywords: AHP, TOPSIS, Scheduling rules selection

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1812
1278 Grocery Customer Behavior Analysis using RFID-based Shopping Paths Data

Authors: In-Chul Jung, Young S. Kwon

Abstract:

Knowing about the customer behavior in a grocery has been a long-standing issue in the retailing industry. The advent of RFID has made it easier to collect moving data for an individual shopper's behavior. Most of the previous studies used the traditional statistical clustering technique to find the major characteristics of customer behavior, especially shopping path. However, in using the clustering technique, due to various spatial constraints in the store, standard clustering methods are not feasible because moving data such as the shopping path should be adjusted in advance of the analysis, which is time-consuming and causes data distortion. To alleviate this problem, we propose a new approach to spatial pattern clustering based on the longest common subsequence. Experimental results using real data obtained from a grocery confirm the good performance of the proposed method in finding the hot spot, dead spot and major path patterns of customer movements.

Keywords: customer path, shopping behavior, exploratoryanalysis, LCS, RFID

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3149
1277 Adjustment of a PET Scanner for PEPT

Authors: Alireza Sadrmomtaz

Abstract:

Positron emission particle tracking (PEPT) is a technique in which a single radioactive tracer particle can be accurately tracked as it moves. A limitation of PET is that in order to reconstruct a tomographic image it is necessary to acquire a large volume of data (millions of events), so it is difficult to study rapidly changing systems. By considering this fact, PEPT is a very fast process compared with PET. In PEPT detecting both photons defines a line and the annihilation is assumed to have occurred somewhere along this line. The location of the tracer can be determined to within a few mm from coincident detection of a small number of pairs of back-to-back gamma rays and using triangulation. This can be achieved many times per second and the track of a moving particle can be reliably followed. This technique was invented at the University of Birmingham [1]. The attempt in PEPT is not to form an image of the tracer particle but simply to determine its location with time. If this tracer is followed for a long enough period within a closed, circulating system it explores all possible types of motion. The application of PEPT to industrial process systems carried out at the University of Birmingham is categorized in two subjects: the behaviour of granular materials and viscous fluids. Granular materials are processed in industry for example in the manufacture of pharmaceuticals, ceramics, food, polymers and PEPT has been used in a number of ways to study the behaviour of these systems [2]. PEPT allows the possibility of tracking a single particle within the bed [3]. Also PEPT has been used for studying systems such as: fluid flow, viscous fluids in mixers [4], using a neutrally-buoyant tracer particle [5].

Keywords: PET, BGO, Particle Tracking, ECAT 931, List mode, PEPT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1403
1276 Approximation of PE-MOCVD to ALD for TiN Concerning Resistivity and Chemical Composition

Authors: D. Geringswald, B. Hintze

Abstract:

The miniaturization of circuits is advancing. During chip manufacturing, structures are filled for example by metal organic chemical vapor deposition (MOCVD). Since this process reaches its limits in case of very high aspect ratios, the use of alternatives such as the atomic layer deposition (ALD) is possible, requiring the extension of existing coating systems. However, it is an unsolved question to what extent MOCVD can achieve results similar as an ALD process. In this context, this work addresses the characterization of a metal organic vapor deposition of titanium nitride. Based on the current state of the art, the film properties coating thickness, sheet resistance, resistivity, stress and chemical composition are considered. The used setting parameters are temperature, plasma gas ratio, plasma power, plasma treatment time, deposition time, deposition pressure, number of cycles and TDMAT flow. The derived process instructions for unstructured wafers and inside a structure with high aspect ratio include lowering the process temperature and increasing the number of cycles, the deposition and the plasma treatment time as well as the plasma gas ratio of hydrogen to nitrogen (H2:N2). In contrast to the current process configuration, the deposited titanium nitride (TiN) layer is more uniform inside the entire test structure. Consequently, this paper provides approaches to employ the MOCVD for structures with increasing aspect ratios.

Keywords: ALD, high aspect ratio, PE-MOCVD, TiN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1507
1275 Tool for Metadata Extraction and Content Packaging as Endorsed in OAIS Framework

Authors: Payal Abichandani, Rishi Prakash, Paras Nath Barwal, B. K. Murthy

Abstract:

Information generated from various computerization processes is a potential rich source of knowledge for its designated community. To pass this information from generation to generation without modifying the meaning is a challenging activity. To preserve and archive the data for future generations it’s very essential to prove the authenticity of the data. It can be achieved by extracting the metadata from the data which can prove the authenticity and create trust on the archived data. Subsequent challenge is the technology obsolescence. Metadata extraction and standardization can be effectively used to resolve and tackle this problem. Metadata can be categorized at two levels i.e. Technical and Domain level broadly. Technical metadata will provide the information that can be used to understand and interpret the data record, but only this level of metadata isn’t sufficient to create trustworthiness. We have developed a tool which will extract and standardize the technical as well as domain level metadata. This paper is about the different features of the tool and how we have developed this.  

Keywords: Digital Preservation, Metadata, OAIS, PDI, XML.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1824
1274 Studies on the Blended Concrete Prepared with Tannery Effluent

Authors: K. Nirmalkumar

Abstract:

There is a acute water problem especially in the dry season in and around Perundurai (Erode district, Tamil Nadu, India) where there are more number of tannery units. Hence an attempt was made to use the waste water from tannery industry for construction purpose. The mechanical properties such as compressive strength, tensile strength, flexural strength etc were studied by casting various concrete specimens in form of cube, cylinders and beams etc and were found to be satisfactory. Hence some special properties such as chloride attack, sulphate attack and chemical attack are considered and comparatively studied with the conventional potable water. In this experimental study the results of specimens prepared by using treated and untreated tannery effluent were compared with the concrete specimens prepared by using potable water. It was observed that the concrete had some reduction in strength while subjected to chloride attack, sulphate attack and chemical attack. So admixtures were selected and optimized in suitable proportion to counter act the adverse effects and the results were found to be satisfactory.

Keywords: Calcium nitrite, concrete, fly ash.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
1273 Role of Direct and Secondary Traumatic Experience on Later Functioning

Authors: Pamela L. Knox, Linda R. Guthrie

Abstract:

Trauma in early life is widely regarded as a cause for adult mental health problems. This study explores the role of secondary trauma on later functioning in a sample of 359 university students enrolled in undergraduate psychology classes in the United States. Participants were initially divided into four groups based on 1) having directly experienced trauma (assaultive violence), 2) having directly experienced trauma and secondary traumatization through the unanticipated death of a close friend or family member or witnessing of an injury or shocking even), 3) having no experience of direct trauma but having experienced indirect trauma (secondary trauma), or 4) reporting no exposure. Participants completed a battery of measures on concepts associated with psychological functioning which included measures of psychological well-being, problem solving, coping and resiliency. Findings discuss differences in psychological functioning and resilience based on participants who experienced secondary traumatization and assaultive violence versus secondary traumatization alone.

Keywords: Psychological Functioning, Resiliency, Trauma, Abuse

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2054
1272 Application of a New Hybrid Optimization Algorithm on Cluster Analysis

Authors: T. Niknam, M. Nayeripour, B.Bahmani Firouzi

Abstract:

Clustering techniques have received attention in many areas including engineering, medicine, biology and data mining. The purpose of clustering is to group together data points, which are close to one another. The K-means algorithm is one of the most widely used techniques for clustering. However, K-means has two shortcomings: dependency on the initial state and convergence to local optima and global solutions of large problems cannot found with reasonable amount of computation effort. In order to overcome local optima problem lots of studies done in clustering. This paper is presented an efficient hybrid evolutionary optimization algorithm based on combining Particle Swarm Optimization (PSO) and Ant Colony Optimization (ACO), called PSO-ACO, for optimally clustering N object into K clusters. The new PSO-ACO algorithm is tested on several data sets, and its performance is compared with those of ACO, PSO and K-means clustering. The simulation results show that the proposed evolutionary optimization algorithm is robust and suitable for handing data clustering.

Keywords: Ant Colony Optimization (ACO), Data clustering, Hybrid evolutionary optimization algorithm, K-means clustering, Particle Swarm Optimization (PSO).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2198
1271 Automating the Testing of Object Behaviour: A Statechart-Driven Approach

Authors: Dong He Nam, Eric C. Mousset, David C. Levy

Abstract:

The evolution of current modeling specifications gives rise to the problem of generating automated test cases from a variety of application tools. Past endeavours on behavioural testing of UML statecharts have not systematically leveraged the potential of existing graph theory for testing of objects. Therefore there exists a need for a simple, tool-independent, and effective method for automatic test generation. An architecture, codenamed ACUTE-J (Automated stateChart Unit Testing Engine for Java), for automating the unit test generation process is presented. A sequential approach for converting UML statechart diagrams to JUnit test classes is described, with the application of existing graph theory. Research byproducts such as a universal XML Schema and API for statechart-driven testing are also proposed. The result from a Java implementation of ACUTE-J is discussed in brief. The Chinese Postman algorithm is utilised as an illustration for a run-through of the ACUTE-J architecture.

Keywords: Automated testing, model based testing, statechart testing, UML, unit testing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1973
1270 Study on Connecting Method of Box Pontoons

Authors: Young-Jun You, Youn-Ju Jeong, Min-Su Park, Du-Ho Lee

Abstract:

Due to a lot of limited conditions, a large box type floating structure is inevitably constructed by connecting many pontoons. When a floating structure is made with concrete, concrete shear key with saw-teeth shape is often used to carry shear force. Match casting for the shear key and precise construction on a sea are very important for making separated two pontoons as one body but those are not easy work and may increase construction time and cost. To solve this problem, one-way shear key is studied in this paper for a connected part where there is some difference between upward and downward shear force. It has only one inclined plane and can resist shear force in one direction. Big shear force is resisted by concrete which forms an inclined plane and small shear force is resisted by steel bar. This system can reduce manufacturing cost of individual pontoon and construction time and cost for constructing a floating structure on a sea. In this paper, the feasibility study about one-way shear key system is performed by comparing with design example.

Keywords: Connection, floating container terminal, pontoon, pre-stressing, shear key.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3112
1269 A Hybrid Approach for Quantification of Novelty in Rule Discovery

Authors: Vasudha Bhatnagar, Ahmed Sultan Al-Hegami, Naveen Kumar

Abstract:

Rule Discovery is an important technique for mining knowledge from large databases. Use of objective measures for discovering interesting rules lead to another data mining problem, although of reduced complexity. Data mining researchers have studied subjective measures of interestingness to reduce the volume of discovered rules to ultimately improve the overall efficiency of KDD process. In this paper we study novelty of the discovered rules as a subjective measure of interestingness. We propose a hybrid approach that uses objective and subjective measures to quantify novelty of the discovered rules in terms of their deviations from the known rules. We analyze the types of deviation that can arise between two rules and categorize the discovered rules according to the user specified threshold. We implement the proposed framework and experiment with some public datasets. The experimental results are quite promising.

Keywords: Knowledge Discovery in Databases (KDD), Data Mining, Rule Discovery, Interestingness, Subjective Measures, Novelty Measure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1354
1268 Re-Optimization MVPP Using Common Subexpression for Materialized View Selection

Authors: Boontita Suchyukorn, Raweewan Auepanwiriyakul

Abstract:

A Data Warehouses is a repository of information integrated from source data. Information stored in data warehouse is the form of materialized in order to provide the better performance for answering the queries. Deciding which appropriated views to be materialized is one of important problem. In order to achieve this requirement, the constructing search space close to optimal is a necessary task. It will provide effective result for selecting view to be materialized. In this paper we have proposed an approach to reoptimize Multiple View Processing Plan (MVPP) by using global common subexpressions. The merged queries which have query processing cost not close to optimal would be rewritten. The experiment shows that our approach can help to improve the total query processing cost of MVPP and sum of query processing cost and materialized view maintenance cost is reduced as well after views are selected to be materialized.

Keywords: Data Warehouse, materialized views, query rewriting, common subexpressions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678
1267 The Status Info Processing and Keeping System for Production Equipment

Authors: So Jeong Nam, Seung Woo Lee, Jai-Kyung Lee

Abstract:

With the globalized production and logistics environment, the need for reducing the product development interval and lead time, having a faster response to orders, conforming to quality standards, fair tracking, and boosting information exchanging activities with customers and partners, and coping with changes in the management environment, manufacturers are in dire need of an information management system in their manufacturing environments. There are lots of information systems that have been designed to manage the condition or operation of equipment in the field but existing systems have a decentralized architecture, which is not unified. Also, these systems cannot effectively handle the status data extraction process upon encountering a problem related to protocols or changes in the equipment or the setting. In this regard, this paper will introduce a system for processing and saving the status info of production equipment, which uses standard representation formats, to enable flexible responses to and support for variables in the field equipment. This system can be used for a variety of manufacturing and equipment settings and is capable of interacting with higher-tier systems such as MES.

Keywords: DAS, Equipment Status, Regular Expression

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1548
1266 Lubricating Grease from Waste Cooking Oil and Waste Motor Sludge

Authors: Aseem Rajvanshi, Pankaj Kumar Pandey

Abstract:

Increase in population has increased the demand of energy to fulfill all its needs. This will result in burden on fossil fuels especially crude oil. Waste oil due to its disposal problem creates environmental degradation. In this context, this paper studies utilization of waste cooking oil and waste motor sludge for making lubricating grease. Experimental studies have been performed by variation in time and concentration of mixture of waste cooking oil and waste motor sludge. The samples were analyzed using penetration test (ASTM D-217), dropping point (ASTM D-566), work penetration (ASTM D-217) and copper strip test (ASTM D-408). Among 6 samples, sample 6 gives the best results with a good drop point and a fine penetration value. The dropping point and penetration test values were found to be 205 °C and 315, respectively. The penetration value falls under the category of NLGI (National Lubricating Grease Institute) consistency number 1.

Keywords: Crude oil, copper strip corrosion test, dropping point, penetration test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3115
1265 Establishing Pairwise Keys Using Key Predistribution Schemes for Sensor Networks

Authors: Y. Harold Robinson, M. Rajaram

Abstract:

Designing cost-efficient, secure network protocols for Wireless Sensor Networks (WSNs) is a challenging problem because sensors are resource-limited wireless devices. Security services such as authentication and improved pairwise key establishment are critical to high efficient networks with sensor nodes. For sensor nodes to correspond securely with each other efficiently, usage of cryptographic techniques is necessary. In this paper, two key predistribution schemes that enable a mobile sink to establish a secure data-communication link, on the fly, with any sensor nodes. The intermediate nodes along the path to the sink are able to verify the authenticity and integrity of the incoming packets using a predicted value of the key generated by the sender’s essential power. The proposed schemes are based on the pairwise key with the mobile sink, our analytical results clearly show that our schemes perform better in terms of network resilience to node capture than existing schemes if used in wireless sensor networks with mobile sinks.

Keywords: Wireless Sensor Networks, predistribution scheme, cryptographic techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1589
1264 Analysis of Impact of Land Use Regulations against Urban Spatial Structure - Centering around Shiheung City

Authors: Chang-il Kang, Yoon-Hong Park, Tae-Hyun Kim, Yu Wann

Abstract:

In this paper, we analyzed the pattern of urban spatial structure of Siheung City that had been divided into two parts and presented alternative plans in order to get rid of these phenomena. Concerning patterns of urban spatial structure, we examined it through means of analyzing status of land use, population density and distribution of residence, status of distribution of main facilities, medical facilities, status of distribution of cultural facilities, distribution of land prices and traffic volume trends. The results of study revealed that status of facilities distribution and distribution of land prices, etc. were bisected by the surrounding area of former municipal office and the district of Sihwa, which were both regarded as one apex of the city divide, forming a duo-centric city. In order to get rid of this problem concerned with urban spatial structure that has been bisected, it is required that measures in order to expand facilities in Siheung City should be taken.

Keywords: Urban Spatial Structure, Duo-centric City, Siheung City.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1242
1263 Linear Quadratic Gaussian/Loop Transfer Recover Control Flight Control on a Nonlinear Model

Authors: T. Sanches, K. Bousson

Abstract:

As part of the development of a 4D autopilot system for unmanned aerial vehicles (UAVs), i.e. a time-dependent robust trajectory generation and control algorithm, this work addresses the problem of optimal path control based on the flight sensors data output that may be unreliable due to noise on data acquisition and/or transmission under certain circumstances. Although several filtering methods, such as the Kalman-Bucy filter or the Linear Quadratic Gaussian/Loop Transfer Recover Control (LQG/LTR), are available, the utter complexity of the control system, together with the robustness and reliability required of such a system on a UAV for airworthiness certifiable autonomous flight, required the development of a proper robust filter for a nonlinear system, as a way of further mitigate errors propagation to the control system and improve its ,performance. As such, a nonlinear algorithm based upon the LQG/LTR, is validated through computational simulation testing, is proposed on this paper.

Keywords: Autonomous flight, LQG/LTR, nonlinear state estimator, robust flight control and stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 695
1262 Adaptive Naïve Bayesian Anti-Spam Engine

Authors: Wojciech P. Gajewski

Abstract:

The problem of spam has been seriously troubling the Internet community during the last few years and currently reached an alarming scale. Observations made at CERN (European Organization for Nuclear Research located in Geneva, Switzerland) show that spam mails can constitute up to 75% of daily SMTP traffic. A naïve Bayesian classifier based on a Bag Of Words representation of an email is widely used to stop this unwanted flood as it combines good performance with simplicity of the training and classification processes. However, facing the constantly changing patterns of spam, it is necessary to assure online adaptability of the classifier. This work proposes combining such a classifier with another NBC (naïve Bayesian classifier) based on pairs of adjacent words. Only the latter will be retrained with examples of spam reported by users. Tests are performed on considerable sets of mails both from public spam archives and CERN mailboxes. They suggest that this architecture can increase spam recall without affecting the classifier precision as it happens when only the NBC based on single words is retrained.

Keywords: Text classification, naïve Bayesian classification, spam, email.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4415
1261 Uniform Distribution of Ductility Demand in Irregular Bridges using Shape Memory Alloy

Authors: Seyed Mohyeddin Ghodratian, Mehdi Ghassemieh, Mohammad Khanmohammadi

Abstract:

Excessive ductility demand on shorter piers is a common problem for irregular bridges subjected to strong ground motion. Various techniques have been developed to reduce the likelihood of collapse of bridge due to failure of shorter piers. This paper presents the new approach to improve the seismic behavior of such bridges using Nitinol shape memory alloys (SMAs). Superelastic SMAs have the ability to remain elastic under very large deformation due to martensitic transformation. This unique property leads to enhanced performance of controlled bridge compared with the performance of the reference bridge. To evaluate the effectiveness of the devices, nonlinear time history analysis is performed on a RC single column bent highway bridge using a suite of representative ground motions. The results show that this method is very effective in limiting the ductility demand of shorter pier.

Keywords: bridge, ductility demand, irregularity, shape memory alloy

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1871
1260 A Case Study on the Value of Corporate Social Responsibility Systems

Authors: José M. Brotons, Manuel E. Sansalvador

Abstract:

The relationship between Corporate Social Responsibility (CSR) and financial performance (FP) is a subject of great interest that has not yet been resolved. In this work, we have developed a new and original tool to measure this relation. The tool quantifies the value contributed to companies that are committed to CSR. The theoretical model used is the fuzzy discounted cash flow method. Two assumptions have been considered, the first, the company has implemented the IQNet SR10 certification, and the second, the company has not implemented that certification. For the first one, the growth rate used for the time horizon is the rate maintained by the company after obtaining the IQNet SR10 certificate. For the second one, both, the growth rates company prior to the implementation of the certification, and the evolution of the sector will be taken into account. By using triangular fuzzy numbers, it is possible to deal adequately with each company’s forecasts as well as the information corresponding to the sector. Once the annual growth rate of the sales is obtained, the profit and loss accounts are generated from the annual estimate sales. For the remaining elements of this account, their regression with the nets sales has been considered. The difference between these two valuations, made in a fuzzy environment, allows obtaining the value of the IQNet SR10 certification. Although this study presents an innovative methodology to quantify the relation between CSR and FP, the authors are aware that only one company has been analyzed. This is precisely the main limitation of this study which in turn opens up an interesting line for future research: to broaden the sample of companies.

Keywords: Corporate social responsibility, case study, financial performance, company valuation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 789
1259 RBF Modelling and Optimization Control for Semi-Batch Reactors

Authors: Magdi M. Nabi, Ding-Li Yu

Abstract:

This paper presents a neural network based model predictive control (MPC) strategy to control a strongly exothermic reaction with complicated nonlinear kinetics given by Chylla-Haase polymerization reactor that requires a very precise temperature control to maintain product uniformity. In the benchmark scenario, the operation of the reactor must be guaranteed under various disturbing influences, e.g., changing ambient temperatures or impurity of the monomer. Such a process usually controlled by conventional cascade control, it provides a robust operation, but often lacks accuracy concerning the required strict temperature tolerances. The predictive control strategy based on the RBF neural model is applied to solve this problem to achieve set-point tracking of the reactor temperature against disturbances. The result shows that the RBF based model predictive control gives reliable result in the presence of some disturbances and keeps the reactor temperature within a tight tolerance range around the desired reaction temperature.

Keywords: Chylla-Haase reactor, RBF neural network modelling, model predictive control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2496
1258 Enhancing Competition in Public Procurement for Sustained Growth: Applying a Double Selection Model to Road Procurement Auctions

Authors: Antonio Estache, Atsushi Iimi

Abstract:

Limited competition has been a serious concern in infrastructure procurement. Importantly, however, there are normally a number of potential bidders initially showing interest in proposed projects. This paper focuses on tackling the question why these initially interested bidders fade out. An empirical problem is that no bids of fading-out firms are observable. They could decide not to enter the process at the beginning of the tendering or may be technically disqualified at any point in the selection process. The paper applies the double selection model to procurement data from road development projects in developing countries and shows that competition ends up restricted, because bidders are self-selective and auctioneers also tend to limit participation depending on the size of contracts.Limited competition would likely lead to high infrastructure procurement costs, threatening fiscal sustainability and economic growth.

Keywords: Auction theory, endogenous bidder entry, infrastructure development, public procurement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1414
1257 A Self Adaptive Genetic Based Algorithm for the Identification and Elimination of Bad Data

Authors: A. A. Hossam-Eldin, E. N. Abdallah, M. S. El-Nozahy

Abstract:

The identification and elimination of bad measurements is one of the basic functions of a robust state estimator as bad data have the effect of corrupting the results of state estimation according to the popular weighted least squares method. However this is a difficult problem to handle especially when dealing with multiple errors from the interactive conforming type. In this paper, a self adaptive genetic based algorithm is proposed. The algorithm utilizes the results of the classical linearized normal residuals approach to tune the genetic operators thus instead of making a randomized search throughout the whole search space it is more likely to be a directed search thus the optimum solution is obtained at very early stages(maximum of 5 generations). The algorithm utilizes the accumulating databases of already computed cases to reduce the computational burden to minimum. Tests are conducted with reference to the standard IEEE test systems. Test results are very promising.

Keywords: Bad Data, Genetic Algorithms, Linearized Normal residuals, Observability, Power System State Estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1346
1256 Investigation on Novel Based Naturally-Inspired Swarm Intelligence Algorithms for Optimization Problems in Mobile Ad Hoc Networks

Authors: C. Rajan, K. Geetha, C. Rasi Priya, S. Geetha

Abstract:

Nature is the immense gifted source for solving complex problems. It always helps to find the optimal solution to solve the problem. Mobile Ad Hoc NETwork (MANET) is a wide research area of networks which has set of independent nodes. The characteristics involved in MANET’s are Dynamic, does not depend on any fixed infrastructure or centralized networks, High mobility. The Bio-Inspired algorithms are mimics the nature for solving optimization problems opening a new era in MANET. The typical Swarm Intelligence (SI) algorithms are Ant Colony Optimization (ACO), Artificial Bee Colony (ABC), Particle Swarm Optimization (PSO), Modified Termite Algorithm, Bat Algorithm (BA), Wolf Search Algorithm (WSA) and so on. This work mainly concentrated on nature of MANET and behavior of nodes. Also it analyses various performance metrics such as throughput, QoS and End-to-End delay etc.

Keywords: Ant Colony Algorithm, Artificial Bee Colony algorithm, Bio-Inspired algorithm, Modified Termite Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2471
1255 Passive Neutralization of Acid Mine Drainage Using Locally Produced Limestone

Authors: Reneiloe Seodigeng, Malwandla Hanabe, Haleden Chiririwa, Hilary Rutto, Tumisang Seodigeng

Abstract:

Neutralisation of acid-mine drainage (AMD) using limestone is cost effective, and good results can be obtained. However, this process has its limitations; it cannot be used for highly acidic water which consists of Fe(III). When Fe(III) reacts with CaCO3, it results in armoring. Armoring slows the reaction, and additional alkalinity can no longer be generated. Limestone is easily accessible, so this problem can be easily dealt with. Experiments were carried out to evaluate the effect of PVC pipe length on ferric and ferrous ions. It was found that the shorter the pipe length the more these dissolved metals precipitate. The effect of the pipe length on the hydrogen ions was also studied, and it was found that these two have an inverse relationship. Experimental data were further compared with the model prediction data to see if they behave in a similar fashion. The model was able to predict the behaviour of 1.5m and 2 m pipes in ferric and ferrous ion precipitation.

Keywords: Acid mine drainage, neutralization, limestone, modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1331
1254 Discovery and Capture of Organizational Knowledge from Unstructured Information

Authors: J. Gu, W.B. Lee, C.F. Cheung, E. Tsui, W.M. Wang

Abstract:

Knowledge of an organization does not merely reside in structured form of information and data; it is also embedded in unstructured form. The discovery of such knowledge is particularly difficult as the characteristic is dynamic, scattered, massive and multiplying at high speed. Conventional methods of managing unstructured information are considered too resource demanding and time consuming to cope with the rapid information growth. In this paper, a Multi-faceted and Automatic Knowledge Elicitation System (MAKES) is introduced for the purpose of discovery and capture of organizational knowledge. A trial implementation has been conducted in a public organization to achieve the objective of decision capture and navigation from a number of meeting minutes which are autonomously organized, classified and presented in a multi-faceted taxonomy map in both document and content level. Key concepts such as critical decision made, key knowledge workers, knowledge flow and the relationship among them are elicited and displayed in predefined knowledge model and maps. Hence, the structured knowledge can be retained, shared and reused. Conducting Knowledge Management with MAKES reduces work in searching and retrieving the target decision, saves a great deal of time and manpower, and also enables an organization to keep pace with the knowledge life cycle. This is particularly important when the amount of unstructured information and data grows extremely quickly. This system approach of knowledge management can accelerate value extraction and creation cycles of organizations.

Keywords: Knowledge-Based System, Knowledge Elicitation, Knowledge Management, Taxonomy, Unstructured Information Management

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1841