Search results for: component based software engineering
33554 Coupling Concept of Two Parallel Research Codes for Two and Three Dimensional Fluid Structure Interaction Analysis
Authors: Luciano Garelli, Marco Schauer, Jorge D’Elia, Mario A. Storti, Sabine C. Langer
Abstract:
This paper discuss a coupling strategy of two different software packages to provide fluid structure interaction (FSI) analysis. The basic idea is to combine the advantages of the two codes to create a powerful FSI solver for two and three dimensional analysis. The fluid part is computed by a program called PETSc-FEM, a software developed at Centro de Investigación de Métodos Computacionales (CIMEC). The structural part of the coupled process is computed by the research code elementary Parallel Solver (elPaSo) of the Technische Universität Braunschweig, Institut für Konstruktionstechnik (IK).Keywords: computational fluid dynamics (CFD), fluid structure interaction (FSI), finite element method (FEM), software
Procedia PDF Downloads 55333553 Site Selection of CNG Station by Using FUZZY-AHP Model (Case Study: Gas Zone 4, Tehran City Iran)
Authors: Hamidrza Joodaki
Abstract:
The most complex issue in urban land use planning is site selection that needs to assess the verity of elements and factors. Multi Criteria Decision Making (MCDM) methods are the best approach to deal with complex problems. In this paper, combination of the analytical hierarchy process (AHP) model and FUZZY logic was used as MCDM methods to select the best site for gas station in the 4th gas zone of Tehran. The first and the most important step in FUZZY-AHP model is selection of criteria and sub-criteria. Population, accessibility, proximity and natural disasters were considered as the main criteria in this study. After choosing the criteria, they were weighted based on AHP by EXPERT CHOICE software, and FUZZY logic was used to enhance accuracy and to approach the reality. After these steps, criteria layers were produced and weighted based on FUZZY-AHP model in GIS. Finally, through ARC GIS software, the layers were integrated and the 4th gas zone in TEHRAN was selected as the best site to locate gas station.Keywords: multiple criteria decision making (MCDM), analytic hierarchy process (AHP), FUZZY logic, geographic information system (GIS)
Procedia PDF Downloads 36333552 Multi-Objective Optimization of Assembly Manufacturing Factory Setups
Authors: Andreas Lind, Aitor Iriondo Pascual, Dan Hogberg, Lars Hanson
Abstract:
Factory setup lifecycles are most often described and prepared in CAD environments; the preparation is based on experience and inputs from several cross-disciplinary processes. Early in the factory setup preparation, a so-called block layout is created. The intention is to describe a high-level view of the intended factory setup and to claim area reservations and allocations. Factory areas are then blocked, i.e., targeted to be used for specific intended resources and processes, later redefined with detailed factory setup layouts. Each detailed layout is based on the block layout and inputs from cross-disciplinary preparation processes, such as manufacturing sequence, productivity, workers’ workplace requirements, and resource setup preparation. However, this activity is often not carried out with all variables considered simultaneously, which might entail a risk of sub-optimizing the detailed layout based on manual decisions. Therefore, this work aims to realize a digital method for assembly manufacturing layout planning where productivity, area utilization, and ergonomics can be considered simultaneously in a cross-disciplinary manner. The purpose of the digital method is to support engineers in finding optimized designs of detailed layouts for assembly manufacturing factories, thereby facilitating better decisions regarding setups of future factories. Input datasets are company-specific descriptions of required dimensions for specific area reservations, such as defined dimensions of a worker’s workplace, material façades, aisles, and the sequence to realize the product assembly manufacturing process. To test and iteratively develop the digital method, a demonstrator has been developed with an adaptation of existing software that simulates and proposes optimized designs of detailed layouts. Since the method is to consider productivity, ergonomics, area utilization, and constraints from the automatically generated block layout, a multi-objective optimization approach is utilized. In the demonstrator, the input data are sent to the simulation software industrial path solutions (IPS). Based on the input and Lua scripts, the IPS software generates a block layout in compliance with the company’s defined dimensions of area reservations. Communication is then established between the IPS and the software EPP (Ergonomics in Productivity Platform), including intended resource descriptions, assembly manufacturing process, and manikin (digital human) resources. Using multi-objective optimization approaches, the EPP software then calculates layout proposals that are sent iteratively and simulated and rendered in IPS, following the rules and regulations defined in the block layout as well as productivity and ergonomics constraints and objectives. The software demonstrator is promising. The software can handle several parameters to optimize the detailed layout simultaneously and can put forward several proposals. It can optimize multiple parameters or weight the parameters to fine-tune the optimal result of the detailed layout. The intention of the demonstrator is to make the preparation between cross-disciplinary silos transparent and achieve a common preparation of the assembly manufacturing factory setup, thereby facilitating better decisions.Keywords: factory setup, multi-objective, optimization, simulation
Procedia PDF Downloads 15333551 Requirement Engineering for Intrusion Detection Systems in Wireless Sensor Networks
Authors: Afnan Al-Romi, Iman Al-Momani
Abstract:
The urge of applying the Software Engineering (SE) processes is both of vital importance and a key feature in critical, complex large-scale systems, for example, safety systems, security service systems, and network systems. Inevitably, associated with this are risks, such as system vulnerabilities and security threats. The probability of those risks increases in unsecured environments, such as wireless networks in general and in Wireless Sensor Networks (WSNs) in particular. WSN is a self-organizing network of sensor nodes connected by wireless links. WSNs consist of hundreds to thousands of low-power, low-cost, multi-function sensor nodes that are small in size and communicate over short-ranges. The distribution of sensor nodes in an open environment that could be unattended in addition to the resource constraints in terms of processing, storage and power, make such networks in stringent limitations such as lifetime (i.e. period of operation) and security. The importance of WSN applications that could be found in many militaries and civilian aspects has drawn the attention of many researchers to consider its security. To address this important issue and overcome one of the main challenges of WSNs, security solution systems have been developed by researchers. Those solutions are software-based network Intrusion Detection Systems (IDSs). However, it has been witnessed, that those developed IDSs are neither secure enough nor accurate to detect all malicious behaviours of attacks. Thus, the problem is the lack of coverage of all malicious behaviours in proposed IDSs, leading to unpleasant results, such as delays in the detection process, low detection accuracy, or even worse, leading to detection failure, as illustrated in the previous studies. Also, another problem is energy consumption in WSNs caused by IDS. So, in other words, not all requirements are implemented then traced. Moreover, neither all requirements are identified nor satisfied, as for some requirements have been compromised. The drawbacks in the current IDS are due to not following structured software development processes by researches and developers when developing IDS. Consequently, they resulted in inadequate requirement management, process, validation, and verification of requirements quality. Unfortunately, WSN and SE research communities have been mostly impermeable to each other. Integrating SE and WSNs is a real subject that will be expanded as technology evolves and spreads in industrial applications. Therefore, this paper will study the importance of Requirement Engineering when developing IDSs. Also, it will study a set of existed IDSs and illustrate the absence of Requirement Engineering and its effect. Then conclusions are drawn in regard of applying requirement engineering to systems to deliver the required functionalities, with respect to operational constraints, within an acceptable level of performance, accuracy and reliability.Keywords: software engineering, requirement engineering, Intrusion Detection System, IDS, Wireless Sensor Networks, WSN
Procedia PDF Downloads 32433550 Oriented Strandboard-GEOGYPTM Undelayment, a Novel Composite Flooring System
Authors: B. Noruziaan, A. Shvarzman, R. Leahy
Abstract:
An innovative flooring underlayment was produced and tested. The composite system is made of common OSB boards and a layer of eco-friendly non-cement gypsum based material (GeoGypTM). It was found that the shear bond between the two materials is sufficient to secure the composite interaction between the two. The very high compressive strength and relatively high tensile strength of the non-cement based component together with its high modulus of elasticity provides enough strength and stiffness for the composite product to cover wider spacing between the joists. The initial findings of this study indicate that with joist spacing as wide as 800 mm, the flooring system provides enough strength without compromising the serviceability requirements of the building codes.Keywords: Composite, floor deck, gypsum based, lumber joist, non-cement, oriented strandboard, shear bond
Procedia PDF Downloads 42033549 Applications for Accounting of Inherited Object-Oriented Class Members
Authors: Jehad Al Dallal
Abstract:
A class in an Object-Oriented (OO) system is the basic unit of design, and it encapsulates a set of attributes and methods. In OO systems, instead of redefining the attributes and methods that are included in other classes, a class can inherit these attributes and methods and only implement its unique attributes and methods, which results in reducing code redundancy and improving code testability and maintainability. Such mechanism is called Class Inheritance. However, some software engineering applications may require accounting for all the inherited class members (i.e., attributes and methods). This paper explains how to account for inherited class members and discusses the software engineering applications that require such consideration.Keywords: class flattening, external quality attribute, inheritance, internal quality attribute, object-oriented design
Procedia PDF Downloads 27533548 Automatic Algorithm for Processing and Analysis of Images from the Comet Assay
Authors: Yeimy L. Quintana, Juan G. Zuluaga, Sandra S. Arango
Abstract:
The comet assay is a method based on electrophoresis that is used to measure DNA damage in cells and has shown important results in the identification of substances with a potential risk to the human population as innumerable physical, chemical and biological agents. With this technique is possible to obtain images like a comet, in which the tail of these refers to damaged fragments of the DNA. One of the main problems is that the image has unequal luminosity caused by the fluorescence microscope and requires different processing to condition it as well as to know how many optimal comets there are per sample and finally to perform the measurements and determine the percentage of DNA damage. In this paper, we propose the design and implementation of software using Image Processing Toolbox-MATLAB that allows the automation of image processing. The software chooses the optimum comets and measuring the necessary parameters to detect the damage.Keywords: artificial vision, comet assay, DNA damage, image processing
Procedia PDF Downloads 31233547 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks
Authors: Mst Shapna Akter, Hossain Shahriar
Abstract:
One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.Keywords: cyber security, vulnerability detection, neural networks, feature extraction
Procedia PDF Downloads 9133546 Dynamic Software Product Lines for Customer Centric Context Aware Business Process Management
Authors: Bochra Khiari, Lamia Labed
Abstract:
In the new digital marketplace, organizations are striving for a proactive position by leveraging the great potential of disruptive technologies to seize the full opportunity of the digital revolution in order to reshape their customer value propositions. New technologies such as big data analytics, which provide prediction of future events based on real-time information, are being integrated into BPM which urges the need for additional core values like capabilities for dynamic adaptation, autonomic behavior, runtime reconfiguration and post-deployment activities to manage unforeseen scenarios at runtime in a situated and changeable context. Dynamic Software Product Lines (DSPL) is an emerging paradigm that supports these runtime variability mechanisms. However, few works exploiting DSPLs principles and techniques in the BPM domain have been proposed so far. In this paper, we propose a conceptual approach DynPL4CBPM, which integrates DSPLs concepts along with the entire related dynamic properties, to the whole BPM lifecycle in order to dynamically adapt business processes according to different context conditions in an individual environment.Keywords: adaptive processes, context aware business process management, customer centric business process management, dynamic software product lines
Procedia PDF Downloads 16133545 Automated Java Testing: JUnit versus AspectJ
Authors: Manish Jain, Dinesh Gopalani
Abstract:
Growing dependency of mankind on software technology increases the need for thorough testing of the software applications and automated testing techniques that support testing activities. We have outlined our testing strategy for performing various types of automated testing of Java applications using AspectJ which has become the de-facto standard for Aspect Oriented Programming (AOP). Likewise JUnit, a unit testing framework is the most popular Java testing tool. In this paper, we have evaluated our proposed AOP approach for automated testing and JUnit on various parameters. First we have provided the similarity between the two approaches and then we have done a detailed comparison of the two testing techniques on factors like lines of testing code, learning curve, testing of private members etc. We established that our AOP testing approach using AspectJ has got several advantages and is thus particularly more effective than JUnit.Keywords: aspect oriented programming, AspectJ, aspects, JU-nit, software testing
Procedia PDF Downloads 33133544 Load Comparison between Different Positions during Elite Male Basketball Games: A Sport Metabolomics Approach
Authors: Kayvan Khoramipour, Abbas Ali Gaeini, Elham Shirzad, Øyvind Sandbakk
Abstract:
Basketball has different positions with individual movement profiles, which may influence metabolic demands. Accordingly, the present study aimed to compare the movement and metabolic load between different positions during elite male basketball games. Five main players of 14 teams (n = 70), who participated in the 2017-18 Iranian national basketball leagues, were selected as participants. The players were defined as backcourt (Posts 1-3) and frontcourt (Posts 4-5). Video based time motion analysis (VBTMA) was performed based on players’ individual running and shuffling speed using Dartfish software. Movements were classified into high and low intensity running with and without having the ball, as well as high and low-intensity shuffling and static movements. Mean frequency, duration, and distance were calculated for each class, except for static movements where only frequency was calculated. Saliva samples were collected from each player before and after 40-minute basketball games and analyzed using metabolomics. Principal component analysis (PCA) and Partial least square discriminant analysis (PLSDA) (for metabolomics data) and independent T-tests (for VBTMA) were used as statistical tests. Movement frequency, duration, and distance were higher in backcourt players (all p ≤ 0.05), while static movement frequency did not differ. Saliva samples showed that the levels of Taurine, Succinic acid, Citric acid, Pyruvate, Glycerol, Acetoacetic acid, Acetone, and Hypoxanthine were all higher in backcourt players, whereas Lactate, Alanine, 3-Metyl Histidine, and Methionine were higher in frontcourt players Based on metabolomics, we demonstrate that backcourt and frontcourt players have different metabolic profiles during games, where backcourt players move clearly more during games and therefore rely more on aerobic energy, whereas frontcourt players rely more on anaerobic energy systems in line with less dynamic but more static movement patterns.Keywords: basketball, metabolomics, saliva, sport loadomics
Procedia PDF Downloads 11633543 Methods of Categorizing Architectural Technical Debt
Authors: Blessing Igbadumhe
Abstract:
The continuous long- and short-term delivery of value to customers continues to be the overarching objective of software organizations. Software engineering professionals and organizations face challenges in the maintenance and evolution of software as a result of architectural, technical debt. The issues of architectural, technical debt continue to receive a significant amount of attention because of its important impact on successful system implementation. The cost of doing nothing as far as architectural, technical debt is concerned can be significant both in financial terms and impacts on customers. Different architectural, technical debt issues exist, and this qualitative research design reviewed existing literature on the subject to identify and categorize them. This research intends to contribute to the existing bludgeoning body of knowledge on categorizations and descriptive model of technical debt related issues related to system architecture. The results identify the most common characteristics of architectural and technical debt categories. Raising awareness of the intricacies of architectural and technical debt helps technology stakeholders reduce negative consequences and increase the system success rate.Keywords: architecture, categorizing TD, system design, technical debt
Procedia PDF Downloads 9133542 Introducing Design Principles for Clinical Decision Support Systems
Authors: Luca Martignoni
Abstract:
The increasing usage of clinical decision support systems in healthcare and the demand for software that enables doctors to take informed decisions is changing everyday clinical practice. However, as technology advances not only are the benefits of technology growing, but so are the potential risks. A growing danger is the doctors’ over-reliance on the proposed decision of the clinical decision support system, leading towards deskilling and rash decisions by doctors. In that regard, identifying doctors' requirements for software and developing approaches to prevent technological over-reliance is of utmost importance. In this paper, we report the results of a design science research study, focusing on the requirements and design principles of ultrasound software. We conducted a total of 15 interviews with experts about poten-tial ultrasound software functions. Subsequently, we developed meta-requirements and design principles to design future clinical decision support systems efficiently and as free from the occur-rence of technological over-reliance as possible.Keywords: clinical decision support systems, technological over-reliance, design principles, design science research
Procedia PDF Downloads 10133541 Diversification of Rice-Based Cropping Systems under Irrigated Condition
Authors: A. H. Nanher, N. P. Singh
Abstract:
In India, Agriculture is largely in rice- based cropping system. It has indicated decline in factor productivity along with emergence of multi - nutrient deficiency, buildup of soil pathogen and weed flora because it operates and removes nutrients from the same rooting depth. In designing alternative cropping systems, the common approaches are crop intensification, crop diversification and cultivar options. The intensification leads to the diversification of the cropping system. Intensification is achieved by introducing an additional component crop in a pre-dominant sequential system by desirable adjustments in cultivars of one or all the component crops. Invariably, this results in higher land use efficiency and productivity per unit time Crop Diversification through such crop and inclusion of fodder crops help to improve the economic situation of small and marginal farmers because of higher income. Inclusion of crops in sequential and intercropping systems reduces some obnoxious weeds through formation of canopies due to competitive planting pattern and thus provides an opportunity to utilize cropping systems as a tool of weed management with non-chemical means. Use of organic source not only acts as supplement for fertilizer (nitrogen) but also improve the physico-chemical properties of soils. Production and use of nitrogen rich biomass offer better prospect for supplementing chemical fertilizers on regular basis. Such biological diversity brings yield and economic stability because of its potential for compensation among components of the system. In a particular agro-climatic and resource condition, the identification of most suitable crop sequence is based on its productivity, stability, land use efficiency as well as production efficiency and its performance is chiefly judged in terms of productivity and net return.Keywords: integrated farming systems, sustainable intensification, system of crop intensification, wheat
Procedia PDF Downloads 42433540 Review of K0-Factors and Related Nuclear Data of the Selected Radionuclides for Use in K0-NAA
Authors: Manh-Dung Ho, Van-Giap Pham, Van-Doanh Ho, Quang-Thien Tran, Tuan-Anh Tran
Abstract:
The k0-factors and related nuclear data, i.e. the Q0-factors and effective resonance energies (Ēr) of the selected radionuclides which are used in the k0-based neutron activation analysis (k0-NAA), were critically reviewed to be integrated in the “k0-DALAT” software. The k0- and Q0-factors of some short-lived radionuclides: 46mSc, 110Ag, 116m2In, 165mDy, and 183mW, were experimentally determined at the Dalat research reactor. The other radionuclides selected are: 20F, 36S, 49Ca, 60mCo, 60Co, 75Se, 77mSe, 86mRb, 115Cd, 115mIn, 131Ba, 134mCs, 134Cs, 153Gd, 153Sm, 159Gd, 170Tm, 177mYb, 192Ir, 197mHg, 239U and 239Np. The reviewed data as compared with the literature data were biased within 5.6-7.3% in which the experimental re-determined factors were within 6.1 and 7.3%. The NIST standard reference materials: Oyster Tissue (1566b), Montana II Soil (2711a) and Coal Fly Ash (1633b) were used to validate the new reviewed data showing that the new data gave an improved k0-NAA using the “k0-DALAT” software with a factor of 4.5-6.8% for the investigated radionuclides.Keywords: neutron activation analysis, k0-based method, k0 factor, Q0 factor, effective resonance energy
Procedia PDF Downloads 12633539 Rethinking: Training Needs of Secondary School Teachers in Pakistan
Authors: Sidra Rizwan
Abstract:
The article focuses on the training needs of secondary school teachers related to the knowledge component of instructional planning and strategies as stated in the National professional standards for teachers in Pakistan. The study aimed to determine the training needs of secondary school teachers on different aspects of knowledge & understanding component of instructional planning and strategies. The target population of the study was the secondary school teachers across Pakistan. For this purpose, a sample of 400 secondary school teachers was selected through multistage sampling from all the four provinces and Federal capital area. Survey method was adopted to assess the training needs by using a self reporting tool. The tool helped to gauge the training needs through indirect inventory questions as well as a ranking list in which the respondents themselves prioritized their training areas. The results showed variation between the direct and indirect reporting of the teachers on the basis of which it was concluded that the secondary school teachers needed awareness about the knowledge component of instructional planning and strategies in order to redefine their actual training needs. The researcher further identified the training needs of secondary school teachers within each province and Islamabad capital territory; including an analysis of variations between strata. As teachers are considered agents of change, their training according to the professional standards should provide a solid base for “rethinking education”.Keywords: training needs, secondary school teachers, instructional planning & strategies, knowledge & understanding
Procedia PDF Downloads 9133538 Tracking Maximum Power Point Utilizing Artificial Immunity System
Authors: Marwa Ahmed Abd El Hamied
Abstract:
In this paper In this paper, a new technique based on Artificial Immunity System (AIS) technique has been developed to track Maximum Power Point (MPP). AIS system is implemented in a photovoltaic system that is subjected to variable temperature and insulation condition. The proposed novel is simulated using Mat Lab program. The results of simulation have been compared to those who are generated from Observation Controller. The proposed model shows promising results as it provide better accuracy comparing to classical model.Keywords: component, artificial immunity technique, solar energy, perturbation and observation, power based methods
Procedia PDF Downloads 42733537 Analysis of Rural Roads in Developing Countries Using Principal Component Analysis and Simple Average Technique in the Development of a Road Safety Performance Index
Authors: Muhammad Tufail, Jawad Hussain, Hammad Hussain, Imran Hafeez, Naveed Ahmad
Abstract:
Road safety performance index is a composite index which combines various indicators of road safety into single number. Development of a road safety performance index using appropriate safety performance indicators is essential to enhance road safety. However, a road safety performance index in developing countries has not been given as much priority as needed. The primary objective of this research is to develop a general Road Safety Performance Index (RSPI) for developing countries based on the facility as well as behavior of road user. The secondary objectives include finding the critical inputs in the RSPI and finding the better method of making the index. In this study, the RSPI is developed by selecting four main safety performance indicators i.e., protective system (seat belt, helmet etc.), road (road width, signalized intersections, number of lanes, speed limit), number of pedestrians, and number of vehicles. Data on these four safety performance indicators were collected using observation survey on a 20 km road section of the National Highway N-125 road Taxila, Pakistan. For the development of this composite index, two methods are used: a) Principal Component Analysis (PCA) and b) Equal Weighting (EW) method. PCA is used for extraction, weighting, and linear aggregation of indicators to obtain a single value. An individual index score was calculated for each road section by multiplication of weights and standardized values of each safety performance indicator. However, Simple Average technique was used for weighting and linear aggregation of indicators to develop a RSPI. The road sections are ranked according to RSPI scores using both methods. The two weighting methods are compared, and the PCA method is found to be much more reliable than the Simple Average Technique.Keywords: indicators, aggregation, principle component analysis, weighting, index score
Procedia PDF Downloads 15833536 Formex Algebra Adaptation into Parametric Design Tools: Dome Structures
Authors: Réka Sárközi, Péter Iványi, Attila B. Széll
Abstract:
The aim of this paper is to present the adaptation of the dome construction tool for formex algebra to the parametric design software Grasshopper. Formex algebra is a mathematical system, primarily used for planning structural systems such like truss-grid domes and vaults, together with the programming language Formian. The goal of the research is to allow architects to plan truss-grid structures easily with parametric design tools based on the versatile formex algebra mathematical system. To produce regular structures, coordinate system transformations are used and the dome structures are defined in spherical coordinate system. Owing to the abilities of the parametric design software, it is possible to apply further modifications on the structures and gain special forms. The paper covers the basic dome types, and also additional dome-based structures using special coordinate-system solutions based on spherical coordinate systems. It also contains additional structural possibilities like making double layer grids in all geometry forms. The adaptation of formex algebra and the parametric workflow of Grasshopper together give the possibility of quick and easy design and optimization of special truss-grid domes.Keywords: parametric design, structural morphology, space structures, spherical coordinate system
Procedia PDF Downloads 25533535 Simplifying the Migration of Architectures in Embedded Applications Introducing a Pattern Language to Support the Workforce
Authors: Farha Lakhani, Michael J. Pont
Abstract:
There are two main architectures used to develop software for modern embedded systems: these can be labelled as “event-triggered” (ET) and “time-triggered” (TT). The research presented in this paper is concerned with the issues involved in migration between these two architectures. Although TT architectures are widely used in safety-critical applications they are less familiar to developers of mainstream embedded systems. The research presented in this paper began from the premise that–for a broad class of systems that have been implemented using an ET architecture–migration to a TT architecture would improve reliability. It may be tempting to assume that conversion between ET and TT designs will simply involve converting all event-handling software routines into periodic activities. However, the required changes to the software architecture are, in many cases rather more profound. The main contribution of the work presented in this paper is to identify ways in which the significant effort involved in migrating between existing ET architectures and “equivalent” (and effective) TT architectures could be reduced. The research described in this paper has taken an innovative step in this regard by introducing the use of ‘Design patterns’ for this purpose for the first time.Keywords: embedded applications, software architectures, reliability, pattern
Procedia PDF Downloads 32933534 An E-Retailing System Architecture Based on Cloud Computing
Authors: Chanchai Supaartagorn
Abstract:
E-retailing is the sale of goods online that takes place over the Internet. The Internet has shrunk the entire World. The world e-retailing is growing at an exponential rate in the Americas, Europe, and Asia. However, e-retailing costs require expensive investment, such as hardware, software, and security systems. Cloud computing technology is internet-based computing for the management and delivery of applications and services. Cloud-based e-retailing application models allow enterprises to lower their costs with their effective implementation of e-retailing activities. In this paper, we describe the concept of cloud computing and present the architecture of cloud computing, combining the features of e-retailing. In addition, we propose a strategy for implementing cloud computing with e-retailing. Finally, we explain the benefits from the architecture.Keywords: architecture, cloud computing, e-retailing, internet-based
Procedia PDF Downloads 39733533 Leader-Member Exchange and Affective Commitment: The Moderating Role of Exchange Ideology
Authors: Seung Yeon Son
Abstract:
In today’s rapidly changing and increasingly complex environment, organizations have relied on their members’ positive attitude toward their employers. In particular, employees’ organizational commitment (primarily, the effective component) has been recognized as an essential component of organizational functioning and success. Hence, identifying the determinants of effective commitment is one of the most important research issues. This study tested the influence of leader-member exchange (LMX) and exchange ideology on employee’s effective commitment. In addition, the interactive effect of LMX and exchange ideology was examined. Data from 198 members of the Korean military supports each of the hypotheses. Lastly, implications for research and directions for future research are discussed.Keywords: affective commitment, exchange ideology, leader-member exchange, commitment
Procedia PDF Downloads 44233532 Principal Component Analysis Applied to the Electric Power Systems – Practical Guide; Practical Guide for Algorithms
Authors: John Morales, Eduardo Orduña
Abstract:
Currently the Principal Component Analysis (PCA) theory has been used to develop algorithms regarding to Electric Power Systems (EPS). In this context, this paper presents a practical tutorial of this technique detailed their concept, on-line and off-line mathematical foundations, which are necessary and desirables in EPS algorithms. Thus, features of their eigenvectors which are very useful to real-time process are explained, showing how it is possible to select these parameters through a direct optimization. On the other hand, in this work in order to show the application of PCA to off-line and on-line signals, an example step to step using Matlab commands is presented. Finally, a list of different approaches using PCA is presented, and some works which could be analyzed using this tutorial are presented.Keywords: practical guide; on-line; off-line, algorithms, faults
Procedia PDF Downloads 56333531 The Use of the Matlab Software as the Best Way to Recognize Penumbra Region in Radiotherapy
Authors: Alireza Shayegan, Morteza Amirabadi
Abstract:
The y tool was developed to quantitatively compare dose distributions, either measured or calculated. Before computing ɣ, the dose and distance scales of the two distributions, referred to as evaluated and reference, are re-normalized by dose and distance criteria, respectively. The re-normalization allows the dose distribution comparison to be conducted simultaneously along dose and distance axes. Several two-dimensional images were acquired using a Scanning Liquid Ionization Chamber EPID and Extended Dose Range (EDR2) films for regular and irregular radiation fields. The raw images were then converted into two-dimensional dose maps. Transitional and rotational manipulations were performed for images using Matlab software. As evaluated dose distribution maps, they were then compared with the corresponding original dose maps as the reference dose maps.Keywords: energetic electron, gamma function, penumbra, Matlab software
Procedia PDF Downloads 30133530 The Trigger-DAQ System in the Mu2e Experiment
Authors: Antonio Gioiosa, Simone Doanti, Eric Flumerfelt, Luca Morescalchi, Elena Pedreschi, Gianantonio Pezzullo, Ryan A. Rivera, Franco Spinella
Abstract:
The Mu2e experiment at Fermilab aims to measure the charged-lepton flavour violating neutrino-less conversion of a negative muon into an electron in the field of an aluminum nucleus. With the expected experimental sensitivity, Mu2e will improve the previous limit of four orders of magnitude. The Mu2e data acquisition (DAQ) system provides hardware and software to collect digitized data from the tracker, calorimeter, cosmic ray veto, and beam monitoring systems. Mu2e’s trigger and data acquisition system (TDAQ) uses otsdaq as its solution. developed at Fermilab, otsdaq uses the artdaq DAQ framework and art analysis framework, under-the-hood, for event transfer, filtering, and processing. Otsdaq is an online DAQ software suite with a focus on flexibility and scalability while providing a multi-user, web-based interface accessible through the Chrome or Firefox web browser. The detector read out controller (ROC) from the tracker and calorimeter stream out zero-suppressed data continuously to the data transfer controller (DTC). Data is then read over the PCIe bus to a software filter algorithm that selects events which are finally combined with the data flux that comes from a cosmic ray veto system (CRV).Keywords: trigger, daq, mu2e, Fermilab
Procedia PDF Downloads 15533529 Multi-Agent Railway Control System: Requirements Definitions of Multi-Agent System Using the Behavioral Patterns Analysis (BPA) Approach
Authors: Assem I. El-Ansary
Abstract:
This paper illustrates the event-oriented Behavioral Pattern Analysis (BPA) modeling approach in developing an Multi-Agent Railway Control System (MARCS). The Event defined in BPA is a real-life conceptual entity that is unrelated to any implementation. The major contributions of this research are the Behavioral Pattern Analysis (BPA) modeling methodology, and the development of an interactive software tool (DECISION), which is based on a combination of the Analytic Hierarchy Process (AHP) and the ELECTRE Multi-Criteria Decision Making (MCDM) methods.Keywords: analysis, multi-agent, railway control, modeling methodology, software modeling, event-oriented, behavioral pattern, use cases
Procedia PDF Downloads 54733528 Analysis of the Impact of NVivo and EndNote on Academic Research Productivity
Authors: Sujit K. Basak
Abstract:
The aim of this paper is to analyze the impact of literature review software on researchers. The aim of this study was achieved by analyzing models in terms of perceived usefulness, perceived ease of use, and acceptance level. Collected data was analyzed using WarpPLS 4.0 software. This study used two theoretical frameworks namely Technology Acceptance Model and the Training Needs Assessment Model. The study was experimental and was conducted at a public university in South Africa. The results of the study showed that acceptance level has a high impact on research workload and productivity followed by perceived usefulness and perceived ease of use.Keywords: technology acceptance model, training needs assessment model, literature review software, research productivity
Procedia PDF Downloads 50433527 Discover a New Technique for Cancer Recognition by Analysis and Determination of Fractal Dimension Images in Matlab Software
Authors: Saeedeh Shahbazkhany
Abstract:
Cancer is a terrible disease that, if not diagnosed early, therapy can be difficult while it is easily medicable if it is diagnosed in early stages. So it is very important for cancer diagnosis that medical procedures are performed. In this paper we introduce a new method. In this method, we only need pictures of healthy cells and cancer cells. In fact, where we suspect cancer, we take a picture of cells or tissue in that area, and then take some pictures of the surrounding tissues. Then, fractal dimension of images are calculated and compared. Cancer can be easily detected by comparing the fractal dimension of images. In this method, we use Matlab software.Keywords: Matlab software, fractal dimension, cancer, surrounding tissues, cells or tissue, new method
Procedia PDF Downloads 35433526 Mobility Management via Software Defined Networks (SDN) in Vehicular Ad Hoc Networks (VANETs)
Authors: Bilal Haider, Farhan Aadil
Abstract:
A Vehicular Ad hoc Network (VANET) provides various services to end-users traveling on the road at high speeds. However, this high-speed mobility of mobile nodes can cause frequent service disruptions. Various mobility management protocols exist for managing node mobility, but due to their centralized nature, they tend to suffer in the VANET environment. In this research, we proposed a distributed mobility management protocol using software-defined networks (SDN) for VANETs. Instead of relying on a centralized mobility anchor, the mobility functionality is distributed at multiple infrastructural nodes. The protocol is based on the classical Proxy Mobile IP version 6 (PMIPv6). It is evident from simulation results that this work has improved the network performance with respect to nodes throughput, delay, and packet loss.Keywords: SDN, VANET, mobility management, optimization
Procedia PDF Downloads 17133525 Life Cycle Assessment of Residential Buildings: A Case Study in Canada
Authors: Venkatesh Kumar, Kasun Hewage, Rehan Sadiq
Abstract:
Residential buildings consume significant amounts of energy and produce a large amount of emissions and waste. However, there is a substantial potential for energy savings in this sector which needs to be evaluated over the life cycle of residential buildings. Life Cycle Assessment (LCA) methodology has been employed to study the primary energy uses and associated environmental impacts of different phases (i.e., product, construction, use, end of life, and beyond building life) for residential buildings. Four different alternatives of residential buildings in Vancouver (BC, Canada) with a 50-year lifespan have been evaluated, including High Rise Apartment (HRA), Low Rise Apartment (LRA), Single family Attached House (SAH), and Single family Detached House (SDH). Life cycle performance of the buildings is evaluated for embodied energy, embodied environmental impacts, operational energy, operational environmental impacts, total life-cycle energy, and total life cycle environmental impacts. Estimation of operational energy and LCA are performed using DesignBuilder software and Athena Impact estimator software respectively. The study results revealed that over the life span of the buildings, the relationship between the energy use and the environmental impacts are identical. LRA is found to be the best alternative in terms of embodied energy use and embodied environmental impacts; while, HRA showed the best life-cycle performance in terms of minimum energy use and environmental impacts. Sensitivity analysis has also been carried out to study the influence of building service lifespan over 50, 75, and 100 years on the relative significance of embodied energy and total life cycle energy. The life-cycle energy requirements for SDH is found to be a significant component among the four types of residential buildings. The overall disclose that the primary operations of these buildings accounts for 90% of the total life cycle energy which far outweighs minor differences in embodied effects between the buildings.Keywords: building simulation, environmental impacts, life cycle assessment, life cycle energy analysis, residential buildings
Procedia PDF Downloads 475