Search results for: Fact checking
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 788

Search results for: Fact checking

728 Improvement of Realization Quality of Aerospace Products Using Augmented Reality Technology

Authors: Nuran Bahar, Mehmet A. Akcayol

Abstract:

In the aviation industry, many faults may occur frequently during the maintenance processes and assembly operations of complex structured aircrafts because of their high dependencies of components. These faults affect the quality of aircraft parts or developed modules adversely. Technical employee requires long time and high labor force while checking the correctness of each component. In addition, the person must be trained regularly because of the ever-growing and changing technology. Generally, the cost of this training is very high. Augmented Reality (AR) technology reduces the cost of training radically and improves the effectiveness of the training. In this study, the usage of AR technology in the aviation industry has been investigated and the effectiveness of AR with heads-up display glasses has been examined. An application has been developed for comparison of production process with AR and manual one.

Keywords: Aerospace, assembly quality, augmented reality, heads-up display.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1720
727 Price Quoting Method for Contract Manufacturer

Authors: S. Homrossukon, W. Parinyasart

Abstract:

This is an applied research to propose the method for price quotation for a contract electronics manufacturer. It has had a precise price quoting method but such method could not quickly provide a result as the customer required. This reduces the ability of company to compete in this kind of business. In this case, the cause of long time quotation process was analyzed. A lot of product features have been demanded by customer. By checking routine processes, it was found that high fraction of quoting time was used for production time estimating which has effected to the manufacturing or production cost. Then the historical data of products including types, number of components, assembling method, and their assembling time were used to analyze the key components affecting to production time. The price quoting model then was proposed. The implementation of proposed model was able to remarkably reduce quoting time with an acceptable required precision.

Keywords: Price quoting, Contract manufacturer, Stepwise technique, Best subset technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4418
726 Compressible Flow Modeling in Pipes and Porous Media during Blowdown Experiment

Authors: Thomas Paris, Vincent Bruyere, Patrick Namy

Abstract:

A numerical model is developed to simulate gas blowdowns through a thin tube and a filter (porous media), separating a high pressure gas filled reservoir to low pressure ones. Based on a previous work, a one-dimensional approach is developed by using the finite element method to solve the transient compressible flow and to predict the pressure and temperature evolution in space and time. Mass, momentum, and energy conservation equations are solved in a fully coupled way in the reservoirs, the pipes and the porous media. Numerical results, such as pressure and temperature evolutions, are firstly compared with experimental data to validate the model for different configurations. Couplings between porous media and pipe flow are then validated by checking mass balance. The influence of the porous media and the nature of the gas is then studied for different initial high pressure values.

Keywords: Fluid mechanics, compressible flow, heat transfer, porous media.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1124
725 A Visualized Framework for Representing Uncertain and Incomplete Temporal Knowledge

Authors: Yue Wang, Jixin Ma, Brian Knight

Abstract:

This paper presents a visualized computer aided case tool for non-expert, called Visual Time, for representing and reasoning about incomplete and uncertain temporal information. It is both expressive and versatile, allowing logical conjunctions and disjunctions of both absolute and relative temporal relations, such as “Before”, “Meets”, “Overlaps”, “Starts”, “During”, and “Finishes”, etc. In terms of a visualized framework, Visual Time provides a user-friendly environment for describing scenarios with rich temporal structure in natural language, which can be formatted as structured temporal phrases and modeled in terms of Temporal Relationship Diagrams (TRD). A TRD can be automatically and visually transformed into a corresponding Time Graph, supported by automatic consistency checker that derives a verdict to confirm if a given scenario is temporally consistent or inconsistent.

Keywords: Time Visualization, Uncertainty, Incompleteness, Consistency Checking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1502
724 Extreme Rainfall Frequency Analysis for Meteorological Sub-Division 4 of India Using L-Moments

Authors: Th. Arti Devi, Parthasarthi Choudhury

Abstract:

Extreme rainfall frequency analysis for Meteorological Sub-Division 4 of India was analyzed using L-moments approach. Serial Correlation and Mann Kendall tests were conducted for checking serially independent and stationarity of the observations. The discordancy measure for the sites was conducted to detect the discordant sites. The regional homogeneity was tested by comparing with 500 generated homogeneous regions using a 4 parameter Kappa distribution. The best fit distribution was selected based on ZDIST statistics and L-moments ratio diagram from the five extreme value distributions GPD, GLO, GEV, P3 and LP3. The LN3 distribution was selected and regional rainfall frequency relationship was established using index-rainfall procedure. A regional mean rainfall relationship was developed using multiple linear regression with latitude and longitude of the sites as variables.

Keywords: L-moments, ZDIST statistics, Serial correlation, Mann Kendall test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450
723 Through the Lens of Forced Displacement: Refugee Women's Rights as Human Rights

Authors: Pearl K. Atuhaire, Sylvia Kaye

Abstract:

While the need for equal access to civil, political as well as economic, social and cultural rights is clear under the international law, the adoption of the Convention on the Elimination of all forms of Discrimination against women in 1979 made this even clearer. Despite this positive progress, the abuse of refugee women's rights is one of the basic underlying root causes of their marginalisation and violence in their countries of asylum. This paper presents a critical review on the development of refugee women's rights at the international levels and national levels. It provides an array of scholarly literature on this issue and examines the measures taken by the international community to curb the problem of violence against women in their various provisions through the instruments set. It is cognizant of the fact that even if conflict affects both refugee women and men, the effects on women refugees are deep-reaching, due to the cultural strongholds they face. An important aspect of this paper is that it is conceptualised against the fact that refugee women face the problem of sexual and gender based first as refugees and second as women, yet, their rights are stumbled upon. Often times they have been rendered "worthless victims" who are only in need of humanitarian assistance than active participants committed to change their plight through their participation in political, economic and social participation in their societies. Scholars have taken notice of the fact that women's rights in refugee settings have been marginalized and call for a need to incorporate their perspectives in the planning and management of refugee settings in which they live. Underpinning this discussion is feminism theory which gives a clear understanding of the root cause of refugee women's problems. Finally, this paper suggests that these policies should be translated into action at local, national international and regional levels to ensure sustainable peace.

Keywords: Feminism theory, human rights, refugee women, sexual and gender based violence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1775
722 Motion Control of a Ball Throwing Robot with a Flexible Robotic Arm

Authors: Yizhi Gai, Yukinori Kobayashi, Yohei Hoshino, Takanori Emaru

Abstract:

Motion control of flexible arms is more difficult than that of rigid arms, however utilizing its dynamics enables improved performance such as a fast motion in short operation time. This paper investigates a ball throwing robot with one rigid link and one flexible link. This robot throws a ball at a set speed with a proper control torque. A mathematical model of this ball throwing robot is derived through Hamilton’s principle. Several patterns of torque input are designed and tested through the proposed simulation models. The parameters of each torque input pattern is optimized and determined by chaos embedded vector evaluated particle swarm optimization (CEVEPSO). Then, the residual vibration of the manipulator after throwing is suppressed with input shaping technique. Finally, a real experiment is set up for the model checking.

Keywords: Motion control, flexible robotic arm, CEVEPSO, ball throwing robot.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4046
721 Heat Exchanger Design

Authors: Su Thet Mon Than, Khin Aung Lin, Mi Sandar Mon

Abstract:

This paper is intended to assist anyone with some general technical experience, but perhaps limited specific knowledge of heat transfer equipment. A characteristic of heat exchanger design is the procedure of specifying a design, heat transfer area and pressure drops and checking whether the assumed design satisfies all requirements or not. The purpose of this paper is how to design the oil cooler (heat exchanger) especially for shell-and-tube heat exchanger which is the majority type of liquid-to-liquid heat exchanger. General design considerations and design procedure are also illustrated in this paper and a flow diagram is provided as an aid of design procedure. In design calculation, the MatLAB and AutoCAD software are used. Fundamental heat transfer concepts and complex relationships involved in such exchanger are also presented in this paper. The primary aim of this design is to obtain a high heat transfer rate without exceeding the allowable pressure drop. This computer program is highly useful to design the shell-and-tube type heat exchanger and to modify existing deign.

Keywords: Shell-and-Tube Heat Exchanger, MatLAB and AutoCAD

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7912
720 Simulated Annealing Algorithm for Data Aggregation Trees in Wireless Sensor Networks and Comparison with Genetic Algorithm

Authors: Ladan Darougaran, Hossein Shahinzadeh, Hajar Ghotb, Leila Ramezanpour

Abstract:

In ad hoc networks, the main issue about designing of protocols is quality of service, so that in wireless sensor networks the main constraint in designing protocols is limited energy of sensors. In fact, protocols which minimize the power consumption in sensors are more considered in wireless sensor networks. One approach of reducing energy consumption in wireless sensor networks is to reduce the number of packages that are transmitted in network. The technique of collecting data that combines related data and prevent transmission of additional packages in network can be effective in the reducing of transmitted packages- number. According to this fact that information processing consumes less power than information transmitting, Data Aggregation has great importance and because of this fact this technique is used in many protocols [5]. One of the Data Aggregation techniques is to use Data Aggregation tree. But finding one optimum Data Aggregation tree to collect data in networks with one sink is a NP-hard problem. In the Data Aggregation technique, related information packages are combined in intermediate nodes and form one package. So the number of packages which are transmitted in network reduces and therefore, less energy will be consumed that at last results in improvement of longevity of network. Heuristic methods are used in order to solve the NP-hard problem that one of these optimization methods is to solve Simulated Annealing problems. In this article, we will propose new method in order to build data collection tree in wireless sensor networks by using Simulated Annealing algorithm and we will evaluate its efficiency whit Genetic Algorithm.

Keywords: Data aggregation, wireless sensor networks, energy efficiency, simulated annealing algorithm, genetic algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
719 Verification of Protocol Design using UML - SMV

Authors: Prashanth C.M., K. Chandrashekar Shet

Abstract:

In recent past, the Unified Modeling Language (UML) has become the de facto industry standard for object-oriented modeling of the software systems. The syntax and semantics rich UML has encouraged industry to develop several supporting tools including those capable of generating deployable product (code) from the UML models. As a consequence, ensuring the correctness of the model/design has become challenging and extremely important task. In this paper, we present an approach for automatic verification of protocol model/design. As a case study, Session Initiation Protocol (SIP) design is verified for the property, “the CALLER will not converse with the CALLEE before the connection is established between them ". The SIP is modeled using UML statechart diagrams and the desired properties are expressed in temporal logic. Our prototype verifier “UML-SMV" is used to carry out the verification. We subjected an erroneous SIP model to the UML-SMV, the verifier could successfully detect the error (in 76.26ms) and generate the error trace.

Keywords: Unified Modeling Language, Statechart, Verification, Protocol Design, Model Checking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1841
718 Evaluation Method for Information Security Levels of CIIP (Critical Information Infrastructure Protection)

Authors: Soon-Tai Park, Jong-Whoi Shin, Bog-Ki Min, Ik-Sub Lee, Gang-Shin Lee, Jae-Il Lee

Abstract:

As the information age matures, major social infrastructures such as communication, finance, military and energy, have become ever more dependent on information communication systems. And since these infrastructures are connected to the Internet, electronic intrusions such as hacking and viruses have become a new security threat. Especially, disturbance or neutralization of a major social infrastructure can result in extensive material damage and social disorder. To address this issue, many nations around the world are researching and developing various techniques and information security policies as a government-wide effort to protect their infrastructures from newly emerging threats. This paper proposes an evaluation method for information security levels of CIIP (Critical Information Infrastructure Protection), which can enhance the security level of critical information infrastructure by checking the current security status and establish security measures accordingly to protect infrastructures effectively.

Keywords: Information Security Evaluation Methodology, Critical Information Infrastructure Protection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1770
717 Multi-view Description of Real-Time Systems- Architecture

Authors: A. Bessam, M. T. Kimour

Abstract:

Real-time embedded systems should benefit from component-based software engineering to handle complexity and deal with dependability. In these systems, applications should not only be logically correct but also behave within time windows. However, in the current component based software engineering approaches, a few of component models handles time properties in a manner that allows efficient analysis and checking at the architectural level. In this paper, we present a meta-model for component-based software description that integrates timing issues. To achieve a complete functional model of software components, our meta-model focuses on four functional aspects: interface, static behavior, dynamic behavior, and interaction protocol. With each aspect we have explicitly associated a time model. Such a time model can be used to check a component-s design against certain properties and to compute the timing properties of component assemblies.

Keywords: Real-time systems, Software architecture, software component, dependability, time properties, ADL, metamodeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1622
716 Investigation of Slope Stability in Gravel Soils in Unsaturated State

Authors: Seyyed Abolhasan Naeini, Ehsan Azini

Abstract:

In this paper, we consider the stability of a slope of 10 meters in silty gravel soils with modeling in the Geostudio Software.  we intend to use the parameters of the volumetric water content and suction dependent permeability and provides relationships and graphs using the parameters obtained from gradation tests and Atterberg’s limits. Also, different conditions of the soil will be investigated, including: checking the factor of safety and deformation rates and pore water pressure in drained, non-drained and unsaturated conditions, as well as the effect of reducing the water level on other parameters. For this purpose, it is assumed that the groundwater level is at a depth of 2 meters from the ground.  Then, with decreasing water level, the safety factor of slope stability was investigated and it was observed that with decreasing water level, the safety factor increased.

Keywords: Slope stability analysis, factor of safety, matric suction, unsaturated silty gravel soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 814
715 Mechanisms in Regulating Language Practices in Electronics Engineering: A Program Plan for Outcomes-Based Education

Authors: Analiza Acuña-Villacorte

Abstract:

The underlying principle behind the harmonization in international education does not solely aim for the comparability but also the compatibility of outputs produced. The international standard in the different professions particularly in engineering defines the required graduate attributes to attain suitable qualifications and recognitions. This study described the language practices of the Electronics Engineering students of Bulacan State University, Philippines who will be deployed for their internship program. The purpose of the study was achieved by determining the language proficiency of the students in terms of speaking, listening, reading, and writing, and checking the adherence of the University to the commitment of intensifying community building for the Association of Southeast Asian Nation Vision 2020. The analysis of variance of the variables defined the significance between the causal variables and dependent variables. Thus, this study identified the mechanism that would regulate language practices in the Electronics Engineering program.

Keywords: Communicative competence, descriptive design, language practices, mechanisms, outcomes-based education.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575
714 Suitability of Entry into the Euro Area: An Excursion in Selected Economies

Authors: Ludk Benada, Jindika Sedova

Abstract:

The current situation in the eurozone raises a number of topics for discussion and to help in finding an answer to the question of whether a common currency is a more suitable means of coping with the impact of the financial crisis or whether national currencies are better suited to this. The economic situation in the EU is now considerably volatile and, due to problems with the fulfilment of the Maastricht convergence criteria, it is now being considered whether, in their further development, new member states will decide to distance themselves from the euro or will, in an attempt to overcome the crisis, speed up the adoption of the euro. The Czech Republic is one country with little interest in adopting the euro, justified by the fact that a better alternative to dealing with this crisis is an independent monetary policy and its ability to respond flexibly to the economic situation not only in Europe, but around the world. One attribute of the crisis in the Czech Republic and its mitigation is the freely floating exchange rate of the national currency. It is not only the Czech Republic that is attempting to alleviate the impact of the crisis, but also new EU member countries facing fresh questions to which theory have yet to provide wholly satisfactory answers. These questions undoubtedly include the problem of inflation targeting and the choice of appropriate instruments for achieving financial stability. The difficulty lies in the fact that these objectives may be contradictory and may require more than one means of achieving them. In this respect we may assume that membership of the euro zone might not in itself mitigate the development of the recession or protect the nation from future crises. We are of the opinion that the decisive factor in the development of any economy will continue to be the domestic economic policy and the operability of market economic mechanisms. We attempt to document this fact using selected countries as examples, these being the Czech Republic, Poland, Hungary, and Slovakia.

Keywords: Currency exchange rate, Maastricht convergence criteria, monetary union, public finances.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1670
713 The Methodology of Flip Chip Using Astro Place and Route Tool

Authors: Rohaya Abdul Wahab, Raja Mohd Fuad Tengku Aziz, Nazaliza Othman, Sharifah Saleh, Nabihah Razali, Rozaimah Baharim, Md Hanif Md Nasir

Abstract:

This paper will discuss flip chip methodology, in which I/O pads, standard cells, macros and bump cells array are placed in the floorplan, then routed using Astro place and route tool. Final DRC and LVS checking is done using Calibre verification tool. The design vehicle to run this methodology is an OpenRISC design targeted to Silterra 0.18 micrometer technology with 6 metal layers for routing. Astro has extensive support for flip chip placement and routing. Astro tool commands for flip chip are straightforward approach like the conventional standard wire bond packaging. However since we do not have flip chip commands in our Astro tool, no LEF file for bump cell and no LEF file for flip chip I/O pad, we create our own methodology to prepare for future flip chip tapeout. 

Keywords: Astro, bump cell, Calibre, flip chip, LEF, methodology, SCHEME, TCL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2744
712 Impact of Machining Parameters on the Surface Roughness of Machined PU Block

Authors: Louis Denis Kevin Catherine, Raja Aziz Raja Ma’arof, Azrina Arshad, Sangeeth Suresh

Abstract:

Machining parameters are very important in determining the surface quality of any material. In the past decade, some new engineering materials were developed for the manufacturing industry which created a need to conduct an investigation on the impact of the said parameters on their surface roughness. Polyurethane (PU) block is widely used in the automotive industry to manufacture parts such as checking fixtures that are used to verify the dimensional accuracy of automotive parts. In this paper, the design of experiment (DOE) was used to investigate on the effect of the milling parameters on the PU block. Furthermore, an analysis of the machined surface chemical composition was done using scanning electron microscope (SEM). It was found that the surface roughness of the PU block is severely affected when PU undergoes a flood machining process instead of a dry condition. In addition the stepover and the silicon content were found to be the most significant parameters that influence the surface quality of the PU block.

Keywords: Polyurethane (PU), design of experiment (DOE), scanning electron microscope (SEM), surface roughness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3591
711 Minimizing Mutant Sets by Equivalence and Subsumption

Authors: Samia Alblwi, Amani Ayad

Abstract:

Mutation testing is the art of generating syntactic variations of a base program and checking whether a candidate test suite can identify all the mutants that are not semantically equivalent to the base; this technique can be used to assess the quality of test suite. One of the main obstacles to the widespread use of mutation testing is cost, as even small programs (a few dozen lines of code) can give rise to a large number of mutants (up to hundreds); this has created an incentive to seek to reduce the number of mutants while preserving their collective effectiveness. Two criteria have been used to reduce the size of mutant sets: equivalence, which aims to partition the set of mutants into equivalence classes modulo semantic equivalence, and selecting one representative per class; and, subsumption, which aims to define a partial ordering among mutants that ranks mutants by effectiveness and seeks to select maximal elements in this ordering. In this paper, we analyze these two policies using analytical and empirical criteria.

Keywords: Mutation testing, mutant sets, mutant equivalence, mutant subsumption, mutant set minimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 157
710 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm

Authors: Ameur Abdelkader, Abed Bouarfa Hafida

Abstract:

Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.

Keywords: Predictive analysis, big data, predictive analysis algorithms. CART algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1055
709 Enhanced Imperialist Competitive Algorithm for the Cell Formation Problem Using Sequence Data

Authors: S. H. Borghei, E. Teymourian, M. Mobin, G. M. Komaki, S. Sheikh

Abstract:

Imperialist Competitive Algorithm (ICA) is a recent meta-heuristic method that is inspired by the social evolutions for solving NP-Hard problems. The ICA is a population-based algorithm which has achieved a great performance in comparison to other metaheuristics. This study is about developing enhanced ICA approach to solve the Cell Formation Problem (CFP) using sequence data. In addition to the conventional ICA, an enhanced version of ICA, namely EICA, applies local search techniques to add more intensification aptitude and embed the features of exploration and intensification more successfully. Suitable performance measures are used to compare the proposed algorithms with some other powerful solution approaches in the literature. In the same way, for checking the proficiency of algorithms, forty test problems are presented. Five benchmark problems have sequence data, and other ones are based on 0-1 matrices modified to sequence based problems. Computational results elucidate the efficiency of the EICA in solving CFP problems.

Keywords: Cell formation problem, Group technology, Imperialist competitive algorithm, Sequence data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573
708 Fault Detection via Stability Analysis for the Hybrid Control Unit of HEVs

Authors: Kyogun Chang, Yoon Bok Lee

Abstract:

Fault detection determines faultexistence and detecting time. This paper discusses two layered fault detection methods to enhance the reliability and safety. Two layered fault detection methods consist of fault detection methods of component level controllers and system level controllers. Component level controllers detect faults by using limit checking, model-based detection, and data-driven detection and system level controllers execute detection by stability analysis which can detect unknown changes. System level controllers compare detection results via stability with fault signals from lower level controllers. This paper addresses fault detection methods via stability and suggests fault detection criteria in nonlinear systems. The fault detection method applies tothe hybrid control unit of a military hybrid electric vehicleso that the hybrid control unit can detect faults of the traction motor.

Keywords: Two Layered Fault Detection, Stability Analysis, Fault-Tolerant Control

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1702
707 Economic Loss due to Ganoderma Disease in Oil Palm

Authors: K. Assis, K. P. Chong, A. S. Idris, C. M. Ho

Abstract:

Oil palm or Elaeis guineensis is considered as the golden crop in Malaysia. But oil palm industry in this country is now facing with the most devastating disease called as Ganoderma Basal Stem Rot disease. The objective of this paper is to analyze the economic loss due to this disease. There were three commercial oil palm sites selected for collecting the required data for economic analysis. Yield parameter used to measure the loss was the total weight of fresh fruit bunch in six months. The predictors include disease severity, change in disease severity, number of infected neighbor palms, age of palm, planting generation, topography, and first order interaction variables. The estimation model of yield loss was identified by using backward elimination based regression method. Diagnostic checking was conducted on the residual of the best yield loss model. The value of mean absolute percentage error (MAPE) was used to measure the forecast performance of the model. The best yield loss model was then used to estimate the economic loss by using the current monthly price of fresh fruit bunch at mill gate.

Keywords: Ganoderma, oil palm, regression model, yield loss, economic loss.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3206
706 Formal Verification of Cache System Using a Novel Cache Memory Model

Authors: Guowei Hou, Lixin Yu, Wei Zhuang, Hui Qin, Xue Yang

Abstract:

Formal verification is proposed to ensure the correctness of the design and make functional verification more efficient. As cache plays a vital role in the design of System on Chip (SoC), and cache with Memory Management Unit (MMU) and cache memory unit makes the state space too large for simulation to verify, then a formal verification is presented for such system design. In the paper, a formal model checking verification flow is suggested and a new cache memory model which is called “exhaustive search model” is proposed. Instead of using large size ram to denote the whole cache memory, exhaustive search model employs just two cache blocks. For cache system contains data cache (Dcache) and instruction cache (Icache), Dcache memory model and Icache memory model are established separately using the same mechanism. At last, the novel model is employed to the verification of a cache which is module of a custom-built SoC system that has been applied in practical, and the result shows that the cache system is verified correctly using the exhaustive search model, and it makes the verification much more manageable and flexible.

Keywords: Cache system, formal verification, novel model, System on Chip (SoC).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2279
705 Assembly and Alignment of Ship Power Plants in Modern Shipbuilding

Authors: A. O. Mikhailov, K. N. Morozov

Abstract:

Fine alignment of main ship power plants mechanisms and shaft lines provides long-term and failure-free performance of propulsion system while fast and high-quality installation of mechanisms and shaft lines decreases common labor intensity. For checking shaft line allowed stress and setting its alignment it is required to perform calculations considering various stages of life cycle. In 2012 JSC SSTC developed special software complex “Shaftline” for calculation of alignment of having its own I/O interface and display of shaft line 3D model. Alignment of shaft line as per bearing loads is rather labor-intensive procedure. In order to decrease its duration, JSC SSTC developed automated alignment system from ship power plants mechanisms. System operation principle is based on automatic simulation of design load on bearings. Initial data for shaft line alignment can be exported to automated alignment system from PC “Shaft line”.

Keywords: ANSYS, propulsion shaft, shaftline alignment, ship power plants.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3099
704 Measured versus Default Interstate Traffic Data in New Mexico, USA

Authors: M. A. Hasan, M. R. Islam, R. A. Tarefder

Abstract:

This study investigates how the site specific traffic data differs from the Mechanistic Empirical Pavement Design Software default values. Two Weigh-in-Motion (WIM) stations were installed in Interstate-40 (I-40) and Interstate-25 (I-25) to developed site specific data. A computer program named WIM Data Analysis Software (WIMDAS) was developed using Microsoft C-Sharp (.Net) for quality checking and processing of raw WIM data. A complete year data from November 2013 to October 2014 was analyzed using the developed WIM Data Analysis Program. After that, the vehicle class distribution, directional distribution, lane distribution, monthly adjustment factor, hourly distribution, axle load spectra, average number of axle per vehicle, axle spacing, lateral wander distribution, and wheelbase distribution were calculated. Then a comparative study was done between measured data and AASHTOWare default values. It was found that the measured general traffic inputs for I-40 and I-25 significantly differ from the default values.

Keywords: AASHTOWare, Traffic, Weigh-in-Motion, Axle load Distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1684
703 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved

Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben

Abstract:

Model transformation, as a pivotal aspect of Modeldriven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: crossdomain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.

Keywords: Automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2182
702 A Differential Calculus Based Image Steganography with Crossover

Authors: Srilekha Mukherjee, Subha Ash, Goutam Sanyal

Abstract:

Information security plays a major role in uplifting the standard of secured communications via global media. In this paper, we have suggested a technique of encryption followed by insertion before transmission. Here, we have implemented two different concepts to carry out the above-specified tasks. We have used a two-point crossover technique of the genetic algorithm to facilitate the encryption process. For each of the uniquely identified rows of pixels, different mathematical methodologies are applied for several conditions checking, in order to figure out all the parent pixels on which we perform the crossover operation. This is done by selecting two crossover points within the pixels thereby producing the newly encrypted child pixels, and hence the encrypted cover image. In the next lap, the first and second order derivative operators are evaluated to increase the security and robustness. The last lap further ensures reapplication of the crossover procedure to form the final stego-image. The complexity of this system as a whole is huge, thereby dissuading the third party interferences. Also, the embedding capacity is very high. Therefore, a larger amount of secret image information can be hidden. The imperceptible vision of the obtained stego-image clearly proves the proficiency of this approach.

Keywords: Steganography, Crossover, Differential Calculus, Peak Signal to Noise Ratio, Cross-correlation Coefficient.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1373
701 Predicting Protein-Protein Interactions from Protein Sequences Using Phylogenetic Profiles

Authors: Omer Nebil Yaveroglu, Tolga Can

Abstract:

In this study, a high accuracy protein-protein interaction prediction method is developed. The importance of the proposed method is that it only uses sequence information of proteins while predicting interaction. The method extracts phylogenetic profiles of proteins by using their sequence information. Combining the phylogenetic profiles of two proteins by checking existence of homologs in different species and fitting this combined profile into a statistical model, it is possible to make predictions about the interaction status of two proteins. For this purpose, we apply a collection of pattern recognition techniques on the dataset of combined phylogenetic profiles of protein pairs. Support Vector Machines, Feature Extraction using ReliefF, Naive Bayes Classification, K-Nearest Neighborhood Classification, Decision Trees, and Random Forest Classification are the methods we applied for finding the classification method that best predicts the interaction status of protein pairs. Random Forest Classification outperformed all other methods with a prediction accuracy of 76.93%

Keywords: Protein Interaction Prediction, Phylogenetic Profile, SVM , ReliefF, Decision Trees, Random Forest Classification

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1599
700 Multi-models Approach for Describing and Verifying Constraints Based Interactive Systems

Authors: Mamoun Sqali, Mohamed Wassim Trojet

Abstract:

The requirements analysis, modeling, and simulation have consistently been one of the main challenges during the development of complex systems. The scenarios and the state machines are two successful models to describe the behavior of an interactive system. The scenarios represent examples of system execution in the form of sequences of messages exchanged between objects and are a partial view of the system. In contrast, state machines can represent the overall system behavior. The automation of processing scenarios in the state machines provide some answers to various problems such as system behavior validation and scenarios consistency checking. In this paper, we propose a method for translating scenarios in state machines represented by Discreet EVent Specification and procedure to detect implied scenarios. Each induced DEVS model represents the behavior of an object of the system. The global system behavior is described by coupling the atomic DEVS models and validated through simulation. We improve the validation process with integrating formal methods to eliminate logical inconsistencies in the global model. For that end, we use the Z notation.

Keywords: Scenarios, DEVS, synthesis, validation and verification, simulation, formal verification, z notation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
699 Computing Continuous Skyline Queries without Discriminating between Static and Dynamic Attributes

Authors: Ibrahim Gomaa, Hoda M. O. Mokhtar

Abstract:

Although most of the existing skyline queries algorithms focused basically on querying static points through static databases; with the expanding number of sensors, wireless communications and mobile applications, the demand for continuous skyline queries has increased. Unlike traditional skyline queries which only consider static attributes, continuous skyline queries include dynamic attributes, as well as the static ones. However, as skyline queries computation is based on checking the domination of skyline points over all dimensions, considering both the static and dynamic attributes without separation is required. In this paper, we present an efficient algorithm for computing continuous skyline queries without discriminating between static and dynamic attributes. Our algorithm in brief proceeds as follows: First, it excludes the points which will not be in the initial skyline result; this pruning phase reduces the required number of comparisons. Second, the association between the spatial positions of data points is examined; this phase gives an idea of where changes in the result might occur and consequently enables us to efficiently update the skyline result (continuous update) rather than computing the skyline from scratch. Finally, experimental evaluation is provided which demonstrates the accuracy, performance and efficiency of our algorithm over other existing approaches.

Keywords: Continuous query processing, dynamic database, moving object, skyline queries.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1231