Search results for: Generalized matrix approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6169

Search results for: Generalized matrix approach

3859 MIMO System Order Reduction Using Real-Coded Genetic Algorithm

Authors: Swadhin Ku. Mishra, Sidhartha Panda, Simanchala Padhy, C. Ardil

Abstract:

In this paper, real-coded genetic algorithm (RCGA) optimization technique has been applied for large-scale linear dynamic multi-input-multi-output (MIMO) system. The method is based on error minimization technique where the integral square error between the transient responses of original and reduced order models has been minimized by RCGA. The reduction procedure is simple computer oriented and the approach is comparable in quality with the other well-known reduction techniques. Also, the proposed method guarantees stability of the reduced model if the original high-order MIMO system is stable. The proposed approach of MIMO system order reduction is illustrated with the help of an example and the results are compared with the recently published other well-known reduction techniques to show its superiority.

Keywords: Multi-input-multi-output (MIMO) system.Modelorder reduction. Integral squared error (ISE). Real-coded geneticalgorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2262
3858 Semantic Indexing Approach of a Corpora Based On Ontology

Authors: Mohammed Erritali

Abstract:

The growth in the volume of text data such as books and articles in libraries for centuries has imposed to establish effective mechanisms to locate them. Early techniques such as abstraction, indexing and the use of classification categories have marked the birth of a new field of research called "Information Retrieval". Information Retrieval (IR) can be defined as the task of defining models and systems whose purpose is to facilitate access to a set of documents in electronic form (corpus) to allow a user to find the relevant ones for him, that is to say, the contents which matches with the information needs of the user. This paper presents a new semantic indexing approach of a documentary corpus. The indexing process starts first by a term weighting phase to determine the importance of these terms in the documents. Then the use of a thesaurus like Wordnet allows moving to the conceptual level. Each candidate concept is evaluated by determining its level of representation of the document, that is to say, the importance of the concept in relation to other concepts of the document. Finally, the semantic index is constructed by attaching to each concept of the ontology, the documents of the corpus in which these concepts are found.

Keywords: Semantic, indexing, corpora, WordNet, ontology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1368
3857 Creativity in the Use of Sinhala and English in Advertisements in Sri Lanka: A Morphological Analysis

Authors: Chamindi Dilkushi Senaratne

Abstract:

Sri Lanka has lived with the English language for more than 200 years. Although officially considered a link language, the phenomenal usage of English by the Sinhala-English bilingual has given rise to a mixed code with identifiable structural characteristics. The extensive use of the mixed language by the average Sri Lankan bilingual has resulted in it being used as a medium of communication by creative writers of bilingual advertisements in Sri Lanka. This study analyses the way in which English is used in bilingual advertisements in both print and electronic media in Sri Lanka. The theoretical framework for the study is based on Kachru’s analysis of the use of English by the bilingual, Muysken’s typology on code mixing theories in colonial settings and Myers-Scotton’s theory on the Matrix Language Framework Model. The study will look at a selection of Sinhala-English advertisements published in newspapers from 2015 to 2016. Only advertisements using both Sinhala and English are used for the analysis. To substantiate data collected from the newspapers, the study will select bilingual advertisements from television advertisements. The objective of the study is to analyze the mixed patterns used for creative purposes by advertisers. The results of the study will reveal the creativity used by the Sinhala –English bilingual and the morphological processes used by the creators of Sinhala-English bilingual advertisements to attract the masses.

Keywords: Bilingual, code mixing, mixed code, morphology, processes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1947
3856 Feature Point Detection by Combining Advantages of Intensity-based Approach and Edge-based Approach

Authors: Sungho Kim, Chaehoon Park, Yukyung Choi, Soon Kwon, In So Kweon

Abstract:

In this paper, a novel corner detection method is presented to stably extract geometrically important corners. Intensity-based corner detectors such as the Harris corner can detect corners in noisy environments but has inaccurate corner position and misses the corners of obtuse angles. Edge-based corner detectors such as Curvature Scale Space can detect structural corners but show unstable corner detection due to incomplete edge detection in noisy environments. The proposed image-based direct curvature estimation can overcome limitations in both inaccurate structural corner detection of the Harris corner detector (intensity-based) and the unstable corner detection of Curvature Scale Space caused by incomplete edge detection. Various experimental results validate the robustness of the proposed method.

Keywords: Feature, intensity, contour, hybrid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1831
3855 Using Structural Equation Modeling in Causal Relationship Design for Balanced-Scorecards' Strategic Map

Authors: A. Saghaei, R. Ghasemi

Abstract:

Through 1980s, management accounting researchers described the increasing irrelevance of traditional control and performance measurement systems. The Balanced Scorecard (BSC) is a critical business tool for a lot of organizations. It is a performance measurement system which translates mission and strategy into objectives. Strategy map approach is a development variant of BSC in which some necessary causal relations must be established. To recognize these relations, experts usually use experience. It is also possible to utilize regression for the same purpose. Structural Equation Modeling (SEM), which is one of the most powerful methods of multivariate data analysis, obtains more appropriate results than traditional methods such as regression. In the present paper, we propose SEM for the first time to identify the relations between objectives in the strategy map, and a test to measure the importance of relations. In SEM, factor analysis and test of hypotheses are done in the same analysis. SEM is known to be better than other techniques at supporting analysis and reporting. Our approach provides a framework which permits the experts to design the strategy map by applying a comprehensive and scientific method together with their experience. Therefore this scheme is a more reliable method in comparison with the previously established methods.

Keywords: BSC, SEM, Strategy map.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2705
3854 Digital Content Strategy: Detailed Review of the Key Content Components

Authors: Oksana Razina, Shakeel Ahmad, Jessie Qun Ren, Olufemi Isiaq

Abstract:

The modern life of businesses is categorically reliant on their established position online, where digital (and particularly website) content plays a significant role as the first point of information. Digital content, therefore, becomes essential – from making the first impression through to the building and development of client relationships. Despite a number of valuable papers suggesting a strategic approach when dealing with digital data, other sources often do not view or accept the approach to digital content as a holistic or continuous process. Associations are frequently made with merely a one-off marketing campaign or similar. The challenge is in establishing an agreed definition for the notion of Digital Content Strategy (DCS), which currently does not exist, as it is viewed from an excessive number of angles. A strategic approach to content, nonetheless, is required, both practically and contextually. We, therefore, aimed at attempting to identify the key content components, comprising a DCS, to ensure all the aspects were covered and strategically applied – from the company’s understanding of the content value to the ability to display flexibility of content and advances in technology. This conceptual project evaluated existing literature on the topic of DCS and related aspects, using PRISMA Systematic Review Method, Document Analysis, Inclusion and Exclusion Criteria, Scoping Review, Snow-Balling Technique and Thematic Analysis. The data were collected from academic and statistical sources, government and relevant trade publications. Based on the suggestions from academics and trading sources, related to the issues discussed, we revealed the key actions for content creation and attempted to define the notion of DCS. The major finding of the study presented Key Content Components of DCS and can be considered for implementation in a business retail setting.

Keywords: Digital content strategy, digital marketing strategy, key content components, websites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 228
3853 Digital Content Strategy: Detailed Review of the Key Content Components

Authors: Oksana Razina, Shakeel Ahmad, Jessie Qun Ren, Olufemi Isiaq

Abstract:

The modern life of businesses is categorically reliant on their established position online, where digital (and particularly website) content plays a significant role as the first point of information. Digital content, therefore, becomes essential – from making the first impression through to the building and development of client relationships. Despite a number of valuable papers suggesting a strategic approach when dealing with digital data, other sources often do not view or accept the approach to digital content as a holistic or continuous process. Associations are frequently made with merely a one-off marketing campaign or similar. The challenge is in establishing an agreed definition for the notion of Digital Content Strategy (DCS), which currently does not exist, as it is viewed from an excessive number of angles. A strategic approach to content, nonetheless, is required, both practically and contextually. We, therefore, aimed at attempting to identify the key content components, comprising a DCS, to ensure all the aspects were covered and strategically applied – from the company’s understanding of the content value to the ability to display flexibility of content and advances in technology. This conceptual project evaluated existing literature on the topic of DCS and related aspects, using PRISMA Systematic Review Method, Document Analysis, Inclusion and Exclusion Criteria, Scoping Review, Snow-Balling Technique and Thematic Analysis. The data were collected from academic and statistical sources, government and relevant trade publications. Based on the suggestions from academics and trading sources, related to the issues discussed, we revealed the key actions for content creation and attempted to define the notion of DCS. The major finding of the study presented Key Content Components of DCS and can be considered for implementation in a business retail setting.

Keywords: Digital content strategy, digital marketing strategy, key content components, websites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 214
3852 Enhanced Multi-Intensity Analysis in Multi-Scenery Classification-Based Macro and Micro Elements

Authors: R. Bremananth

Abstract:

Several computationally challenging issues are encountered while classifying complex natural scenes. In this paper, we address the problems that are encountered in rotation invariance with multi-intensity analysis for multi-scene overlapping. In the present literature, various algorithms proposed techniques for multi-intensity analysis, but there are several restrictions in these algorithms while deploying them in multi-scene overlapping classifications. In order to resolve the problem of multi-scenery overlapping classifications, we present a framework that is based on macro and micro basis functions. This algorithm conquers the minimum classification false alarm while pigeonholing multi-scene overlapping. Furthermore, a quadrangle multi-intensity decay is invoked. Several parameters are utilized to analyze invariance for multi-scenery classifications such as rotation, classification, correlation, contrast, homogeneity, and energy. Benchmark datasets were collected for complex natural scenes and experimented for the framework. The results depict that the framework achieves a significant improvement on gray-level matrix of co-occurrence features for overlapping in diverse degree of orientations while pigeonholing multi-scene overlapping.

Keywords: Automatic classification, contrast, homogeneity, invariant analysis, multi-scene analysis, overlapping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1121
3851 Blood Cell Dynamics in a Simple Shear Flow using an Implicit Fluid-Structure Interaction Method Based on the ALE Approach

Authors: Choeng-Ryul Choi, Chang-Nyung Kim, Tae-Hyub Hong

Abstract:

A numerical method is developed for simulating the motion of particles with arbitrary shapes in an effectively infinite or bounded viscous flow. The particle translational and angular motions are numerically investigated using a fluid-structure interaction (FSI) method based on the Arbitrary-Lagrangian-Eulerian (ALE) approach and the dynamic mesh method (smoothing and remeshing) in FLUENT ( ANSYS Inc., USA). Also, the effects of arbitrary shapes on the dynamics are studied using the FSI method which could be applied to the motions and deformations of a single blood cell and multiple blood cells, and the primary thrombogenesis caused by platelet aggregation. It is expected that, combined with a sophisticated large-scale computational technique, the simulation method will be useful for understanding the overall properties of blood flow from blood cellular level (microscopic) to the resulting rheological properties of blood as a mass (macroscopic).

Keywords: Blood Flow, Fluid-Structure Interaction (FSI), Micro-Channels, Arbitrary Shapes, Red Blood Cells (RBCs)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2310
3850 Organization of the Purchasing Function for Innovation

Authors: Jasna Prester, Ivana Rašić Bakarić, Božidar Matijević

Abstract:

Innovations not only contribute to competitiveness of the company but have also positive effects on revenues. On average, product innovations account to 14 percent of companies’ sales. Innovation management has substantially changed during the last decade, because of growing reliance on external partners. As a consequence, a new task for purchasing arises, as firms need to understand which suppliers actually do have high potential contributing to the innovativeness of the firm and which do not. Proper organization of the purchasing function is important since for the majority of manufacturing companies deal with substantial material costs which pass through the purchasing function. In the past the purchasing function was largely seen as a transaction-oriented, clerical function but today purchasing is the intermediate with supply chain partners contributing to innovations, be it product or process innovations. Therefore, purchasing function has to be organized differently to enable firm innovation potential. However, innovations are inherently risky. There are behavioral risk (that some partner will take advantage of the other party), technological risk in terms of complexity of products and processes of manufacturing and incoming materials and finally market risks, which in fact judge the value of the innovation. These risks are investigated in this work. Specifically, technological risks which deal with complexity of the products, and processes will be investigated more thoroughly. Buying components or such high edge technologies necessities careful investigation of technical features and therefore is usually conducted by a team of experts. Therefore it is hypothesized that higher the technological risk, higher will be the centralization of the purchasing function as an interface with other supply chain members. Main contribution of this research lies is in the fact that analysis was performed on a large data set of 1493 companies, from 25 countries collected in the GMRG 4 survey. Most analyses of purchasing function are done by case study analysis of innovative firms. Therefore this study contributes with empirical evaluations that can be generalized.

Keywords: Purchasing function organization, innovation, technological risk, GMRG 4 survey.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3721
3849 Analyses and Optimization of Physical and Mechanical Properties of Direct Recycled Aluminium Alloy (AA6061) Wastes by ANOVA Approach

Authors: Mohammed H. Rady, Mohd Sukri Mustapa, S Shamsudin, M. A. Lajis, A. Wagiman

Abstract:

The present study is aimed at investigating microhardness and density of aluminium alloy chips when subjected to various settings of preheating temperature and preheating time. Three values of preheating temperature were taken as 450 °C, 500 °C, and 550 °C. On the other hand, three values of preheating time were chosen (1, 2, 3) hours. The influences of the process parameters (preheating temperature and time) were analyzed using Design of Experiments (DOE) approach whereby full factorial design with center point analysis was adopted. The total runs were 11 and they comprise of two factors of full factorial design with 3 center points. The responses were microhardness and density. The results showed that the density and microhardness increased with decreasing the preheating temperature. The results also found that the preheating temperature is more important to be controlled rather than the preheating time in microhardness analysis while both the preheating temperature and preheating time are important in density analysis. It can be concluded that setting temperature at 450 °C for 1 hour resulted in the optimum responses.

Keywords: AA6061, density, DOE, hot extrusion, microhardness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 704
3848 A System Dynamic Based DSS for Ecological Urban Management in Alexandria, Egypt

Authors: Mona M. Salem, Khaled S. Al-Hagla, Hany M. Ayad

Abstract:

The concept of urban metabolism has increasingly been employed in a diverse range of disciplines as a mean to analyze and theorize the city. Urban ecology has a particular focus on the implications of applying the metabolism concept to the urban realm. This approach has been developed by a few researchers, though it has rarely if ever been used in policy development for city planning. The aim of this research is to use ecologically informed urban planning interventions to increase the sustainability of urban metabolism; with special focus on land stock as a most important city resource by developing a system dynamic based DSS. This model identifies two critical management strategy variables for the Strategic Urban Plan Alexandria SUP 2032. As a result, this comprehensive and precise quantitative approach is needed to monitor, measure, evaluate and observe dynamic urban changes working as a decision support system (DSS) for policy making.

Keywords: Alexandria SUP 2032, DSS, ecology, land resource, LULCC, management, metabolism, model, scenarios, System dynamics, urban development.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1195
3847 Semisolid Structure and Parameters for A360 Aluminum Alloy Prepared by Mechanical Stirring

Authors: MM.Kaykha, A. Kamarei, M. Safari, V. Arbabi

Abstract:

Semisolid metal processing uses solid–liquid slurries containing fine and globular solid particles uniformly distributed in a liquid matrix, which can be handled as a solid and flow like a liquid. In the recent years, many methods have been introduced for the production of semisolid slurries since it is scientifically sound and industrially viable with such preferred microstructures called thixotropic microstructures as feedstock materials. One such process that needs very low equipment investment and running costs is the cooling slope. In this research by using a mechanical stirrer slurry maker constructed by the authors, the effects of mechanical stirring parameters such as: stirring time, stirring temperature and stirring Speed on micro-structure and mechanical properties of A360 aluminum alloy in semi-solid forming, are investigated. It is determined that mold temperature and holding time of part in temperature of 580ºC have a great effect on micro-structure and mechanical properties(stirring temperature of 585ºC, stirring time of 20 minutes and stirring speed of 425 RPM). By optimizing the forming parameters, dendrite microstructure changes to globular and mechanical properties improves. This is because of breaking and globularzing dendrites of primary α-AL.

Keywords: Semi-Solid Forming, Mechanical properties, Shear Rate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2187
3846 Disaggregating and Forecasting the Total Energy Consumption of a Building: A Case Study of a High Cooling Demand Facility

Authors: Juliana Barcelos Cordeiro, Khashayar Mahani, Farbod Farzan, Mohsen A. Jafari

Abstract:

Energy disaggregation has been focused by many energy companies since energy efficiency can be achieved when the breakdown of energy consumption is known. Companies have been investing in technologies to come up with software and/or hardware solutions that can provide this type of information to the consumer. On the other hand, not all people can afford to have these technologies. Therefore, in this paper, we present a methodology for breaking down the aggregate consumption and identifying the highdemanding end-uses profiles. These energy profiles will be used to build the forecast model for optimal control purpose. A facility with high cooling load is used as an illustrative case study to demonstrate the results of proposed methodology. We apply a high level energy disaggregation through a pattern recognition approach in order to extract the consumption profile of its rooftop packaged units (RTUs) and present a forecast model for the energy consumption.  

Keywords: Energy consumption forecasting, energy efficiency, load disaggregation, pattern recognition approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1461
3845 A Formal Property Verification for Aspect-Oriented Programs in Software Development

Authors: Moustapha Bande, Hakima Ould-Slimane, Hanifa Boucheneb

Abstract:

Software development for complex systems requires efficient and automatic tools that can be used to verify the satisfiability of some critical properties such as security ones. With the emergence of Aspect-Oriented Programming (AOP), considerable work has been done in order to better modularize the separation of concerns in the software design and implementation. The goal is to prevent the cross-cutting concerns to be scattered across the multiple modules of the program and tangled with other modules. One of the key challenges in the aspect-oriented programs is to be sure that all the pieces put together at the weaving time ensure the satisfiability of the overall system requirements. Our paper focuses on this problem and proposes a formal property verification approach for a given property from the woven program. The approach is based on the control flow graph (CFG) of the woven program, and the use of a satisfiability modulo theories (SMT) solver to check whether each property (represented par one aspect) is satisfied or not once the weaving is done.

Keywords: Aspect-oriented programming, control flow graph, satisfiability modulo theories, property verification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 751
3844 Safety Culture Implementation Based on Occupational Health and Safety Assessment

Authors: Nyambayar Davaadorj, Ichiro Koshijima

Abstract:

Safety or the state of being safe can be described as a condition of being not dangerous or not harmful. It is necessary for an individual to avoid dangerous situations every day. Also, an organization is subject to legal requirements for the health and safety of persons inside and around the immediate workplace, or who are exposed to the workplace activities. Although it might be difficult to keep a situation where complete safety is ensured, efforts must nonetheless be made to consider ways of removing any potential danger within an organization. In order to ensure a safe working environment, the capability of responding (i.e., resilience) to signals (i.e., information concerning events that could pose future problems that must be taken into account) that occur in and around corporations is necessary. The ability to evaluate this essential point is thus one way in which safety and security can be managed. This study focuses on OHSAS18001, an internationally applied standard for the construction and operation of occupational health and safety management systems, by using IDEF0 for Function Modeling (IDEF0) and the Resilience Matrix originally made by Bracco. Further, this study discusses a method for evaluating a manner in which Occupational Health and Safety Assessment Series (OHSAS) systematically functions within corporations. Based on the findings, this study clarifies the potential structural objection for corporations when implementing and operating the OHSAS standard.

Keywords: OHSAS18001, IDEF0, safety culture, resilience engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1430
3843 Probabilistic Bhattacharya Based Active Contour Model in Structure Tensor Space

Authors: Hiren Mewada, Suprava Patnaik

Abstract:

Object identification and segmentation application requires extraction of object in foreground from the background. In this paper the Bhattacharya distance based probabilistic approach is utilized with an active contour model (ACM) to segment an object from the background. In the proposed approach, the Bhattacharya histogram is calculated on non-linear structure tensor space. Based on the histogram, new formulation of active contour model is proposed to segment images. The results are tested on both color and gray images from the Berkeley image database. The experimental results show that the proposed model is applicable to both color and gray images as well as both texture images and natural images. Again in comparing to the Bhattacharya based ACM in ICA space, the proposed model is able to segment multiple object too.

Keywords: Active Contour, Bhattacharya Histogram, Structure tensor, Image segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2058
3842 Preliminary Studies of MWCNT/PVDF Polymer Composites

Authors: Esther Lorrayne M. Pereira, Adriana Souza M. Batista, Fabíola A. S. Ribeiro, Adelina P. Santos, Clascídia A. Furtado, Luiz O. Faria

Abstract:

The combination of multi–walled carbon nanotubes (MWCNTs) with polymers offers an attractive route to reinforce the macromolecular compounds as well as the introduction of new properties based on morphological modifications or electronic interactions between the two constituents. As they are only a few nanometers in dimension, it offers ultra-large interfacial area per volume between the nano-element and polymer matrix. Nevertheless, the use of MWCNTs as a rough material in different applications has been largely limited by their poor processability, insolubility, and infusibility. Studies concerning the nanofiller reinforced polymer composites are justified in an attempt to overcome these limitations. This work presents one preliminary study of MWCNTs dispersion into the PVDF homopolymer. For preparation, the composite components were diluted in n,n-dimethylacetamide (DMAc) with mechanical agitation assistance. After complete dilution, followed by slow evaporation of the solvent at 60°C, the samples were dried. Films of about 80 μm were obtained. FTIR and UV-Vis spectroscopic techniques were used to characterize the nanocomposites. The appearance of absorption bands in the FTIR spectra of nanofilled samples, when compared to the spectrum of pristine PVDF samples, are discussed and compared with the UV-Vis measurements.

Keywords: Composites materials, FTIR, MWNTs, PVDF, UVVis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1916
3841 Monte Carlo Analysis and Fuzzy Sets for Uncertainty Propagation in SIS Performance Assessment

Authors: Fares Innal, Yves Dutuit, Mourad Chebila

Abstract:

The object of this work is the probabilistic performance evaluation of safety instrumented systems (SIS), i.e. the average probability of dangerous failure on demand (PFDavg) and the average frequency of failure (PFH), taking into account the uncertainties related to the different parameters that come into play: failure rate (λ), common cause failure proportion (β), diagnostic coverage (DC)... This leads to an accurate and safe assessment of the safety integrity level (SIL) inherent to the safety function performed by such systems. This aim is in keeping with the requirement of the IEC 61508 standard with respect to handling uncertainty. To do this, we propose an approach that combines (1) Monte Carlo simulation and (2) fuzzy sets. Indeed, the first method is appropriate where representative statistical data are available (using pdf of the relating parameters), while the latter applies in the case characterized by vague and subjective information (using membership function). The proposed approach is fully supported with a suitable computer code.

Keywords: Fuzzy sets, Monte Carlo simulation, Safety instrumented system, Safety integrity level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2779
3840 Using Design Sprint for Software Engineering Undergraduate Student Projects: A Method Paper

Authors: Sobhani U. Pilapitiya, Tharanga Peiris

Abstract:

Software engineering curriculums generally consist of industry-based practices such as project-based learning (PBL) which mainly focuses on efficient and innovative product development. These approaches can be tailored and used in project-based modules in software engineering curriculums. However, there are very limited attempts in the area especially related to Sri Lankan context. This paper describes a tailored pedagogical approach and its results of using design sprint which can be used for project-based modules in software engineering (SE) curriculums. A controlled group of second year software engineering students was selected for the study. The study results indicate that all of the students agreed that the design sprint approach is effective in group-based projects and 83% of students stated that it minimized the re-work compared to traditional project approaches. The tailored process was effective, easy to implement and produced desired results at the end of the session while providing students an enjoyable experience.

Keywords: design sprint, project-based learning, software engineering, curriculum

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 733
3839 A Hybrid Distributed Vision System for Robot Localization

Authors: Hsiang-Wen Hsieh, Chin-Chia Wu, Hung-Hsiu Yu, Shu-Fan Liu

Abstract:

Localization is one of the critical issues in the field of robot navigation. With an accurate estimate of the robot pose, robots will be capable of navigating in the environment autonomously and efficiently. In this paper, a hybrid Distributed Vision System (DVS) for robot localization is presented. The presented approach integrates odometry data from robot and images captured from overhead cameras installed in the environment to help reduce possibilities of fail localization due to effects of illumination, encoder accumulated errors, and low quality range data. An odometry-based motion model is applied to predict robot poses, and robot images captured by overhead cameras are then used to update pose estimates with HSV histogram-based measurement model. Experiment results show the presented approach could localize robots in a global world coordinate system with localization errors within 100mm.

Keywords: Distributed Vision System, Localization, Measurement model, Motion model

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1340
3838 Quantifying the Stability of Software Systems via Simulation in Dependency Networks

Authors: Weifeng Pan

Abstract:

The stability of a software system is one of the most important quality attributes affecting the maintenance effort. Many techniques have been proposed to support the analysis of software stability at the architecture, file, and class level of software systems, but little effort has been made for that at the feature (i.e., method and attribute) level. And the assumptions the existing techniques based on always do not meet the practice to a certain degree. Considering that, in this paper, we present a novel metric, Stability of Software (SoS), to measure the stability of object-oriented software systems by software change propagation analysis using a simulation way in software dependency networks at feature level. The approach is evaluated by case studies on eight open source Java programs using different software structures (one employs design patterns versus one does not) for the same object-oriented program. The results of the case studies validate the effectiveness of the proposed metric. The approach has been fully automated by a tool written in Java.

Keywords: Software stability, change propagation, design pattern, software maintenance, object-oriented (OO) software.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678
3837 Quantum-Like Approach for Deriving a Theory Describing the Concept of Interpretation

Authors: Yehuda Roth

Abstract:

In quantum theory, a system’s time evolution is predictable unless an observer performs measurement, as the measurement process can randomize the system. This randomness appears when the measuring device does not accurately describe the measured item, i.e., when the states characterizing the measuring device appear as a superposition of those being measured. When such a mismatch occurs, the measured data randomly collapse into a single eigenstate of the measuring device. This  scenario resembles the interpretation process in which the observer does not experience an objective reality but interprets it based on preliminary descriptions initially ingrained into his/her mind. This distinction is the motivation for the present study in which the collapse scenario is regarded as part of the interpretation process of the observer. By adopting the formalism of the quantum theory, we present a complete mathematical approach that describes the interpretation process. We demonstrate this process by applying the proposed interpretation formalism to the ambiguous image "My wife and mother-in-law" to identify whether a woman in the picture is young or old.

Keywords: Interpretation, ambiguous images, data reception, state matching, classification, determination.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 189
3836 Sperm Whale Signal Analysis: Comparison using the Auto Regressive model and the Daubechies 15 Wavelets Transform

Authors: Olivier Adam, Maciej Lopatka, Christophe Laplanche, Jean-François Motsch

Abstract:

This article presents the results using a parametric approach and a Wavelet Transform in analysing signals emitting from the sperm whale. The extraction of intrinsic characteristics of these unique signals emitted by marine mammals is still at present a difficult exercise for various reasons: firstly, it concerns non-stationary signals, and secondly, these signals are obstructed by interfering background noise. In this article, we compare the advantages and disadvantages of both methods: Auto Regressive models and Wavelet Transform. These approaches serve as an alternative to the commonly used estimators which are based on the Fourier Transform for which the hypotheses necessary for its application are in certain cases, not sufficiently proven. These modern approaches provide effective results particularly for the periodic tracking of the signal's characteristics and notably when the signal-to-noise ratio negatively effects signal tracking. Our objectives are twofold. Our first goal is to identify the animal through its acoustic signature. This includes recognition of the marine mammal species and ultimately of the individual animal (within the species). The second is much more ambitious and directly involves the intervention of cetologists to study the sounds emitted by marine mammals in an effort to characterize their behaviour. We are working on an approach based on the recordings of marine mammal signals and the findings from this data result from the Wavelet Transform. This article will explore the reasons for using this approach. In addition, thanks to the use of new processors, these algorithms once heavy in calculation time can be integrated in a real-time system.

Keywords: Autoregressive model, Daubechies Wavelet, Fourier Transform, marine mammals, signal processing, spectrogram, sperm whale, Wavelet Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2005
3835 CRLH and SRR Based Microwave Filter Design Useful for Communication Applications

Authors: Subal Kar, Amitesh Kumar, A. Majumder, S. K. Ghosh, S. Saha, S. S. Sikdar, T. K. Saha

Abstract:

CRLH (composite right/left-handed) based and SRR (split-ring resonator) based filters have been designed at microwave frequency which can provide better performance compared to conventional edge-coupled band-pass filter designed around the same frequency, 2.45 GHz. Both CRLH and SRR are unit cells used in metamaterial design. The primary aim of designing filters with such structures is to realize size reduction and also to realize novel filter performance. The CRLH based filter has been designed in microstrip transmission line, while the SRR based filter is designed with SRR loading in waveguide. The CRLH based filter designed at 2.45 GHz provides an insertion loss of 1.6 dB with harmonic suppression up to 10 GHz with 67 % size reduction when compared with a conventional edge-coupled band-pass filter designed around the same frequency. One dimensional (1-D) SRR matrix loaded in a waveguide shows the possibility of realizing a stop-band with sharp skirts in the pass-band while a stop-band in the pass-band of normal rectangular waveguide with tailoring of the dimensions of SRR unit cells. Such filters are expected to be very useful for communication systems at microwave frequency.

Keywords: BPF, CRLH, Harmonic, Metamaterial, SRR, Waveguide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1770
3834 Computer Proven Correctness of the Rabin Public-Key Scheme

Authors: Johannes Buchmann, Markus Kaiser

Abstract:

We decribe a formal specification and verification of the Rabin public-key scheme in the formal proof system Is-abelle/HOL. The idea is to use the two views of cryptographic verification: the computational approach relying on the vocabulary of probability theory and complexity theory and the formal approach based on ideas and techniques from logic and programming languages. The analysis presented uses a given database to prove formal properties of our implemented functions with computer support. Thema in task in designing a practical formalization of correctness as well as security properties is to cope with the complexity of cryptographic proving. We reduce this complexity by exploring a light-weight formalization that enables both appropriate formal definitions as well as eficient formal proofs. This yields the first computer-proved implementation of the Rabin public-key scheme in Isabelle/HOL. Consequently, we get reliable proofs with a minimal error rate augmenting the used database. This provides a formal basis for more computer proof constructions in this area.

Keywords: public-key encryption, Rabin public-key scheme, formalproof system, higher-order logic, formal verification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1591
3833 Modified Naïve Bayes Based Prediction Modeling for Crop Yield Prediction

Authors: Kefaya Qaddoum

Abstract:

Most of greenhouse growers desire a determined amount of yields in order to accurately meet market requirements. The purpose of this paper is to model a simple but often satisfactory supervised classification method. The original naive Bayes have a serious weakness, which is producing redundant predictors. In this paper, utilized regularization technique was used to obtain a computationally efficient classifier based on naive Bayes. The suggested construction, utilized L1-penalty, is capable of clearing redundant predictors, where a modification of the LARS algorithm is devised to solve this problem, making this method applicable to a wide range of data. In the experimental section, a study conducted to examine the effect of redundant and irrelevant predictors, and test the method on WSG data set for tomato yields, where there are many more predictors than data, and the urge need to predict weekly yield is the goal of this approach. Finally, the modified approach is compared with several naive Bayes variants and other classification algorithms (SVM and kNN), and is shown to be fairly good.

Keywords: Tomato yields prediction, naive Bayes, redundancy

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5109
3832 EU Families and Adolescents Quit Tobacco Focus Group Analysis in Hungary

Authors: Szilvia Gergely Seuss, Mihaela Nistor, Lilla Csáky, Péter Molnár

Abstract:

In the frame of the European Union project entitled EU-Families and Adolescents Quit Tobacco (www.eufaqt.eu) focus group analysis has been carried out in Hungary to acquire qualitative information on attitudes towards smoking in groups of adolescents, parents and educators, respectively. It rendered to identify methods for smoking prevention/ intervention with family approach. The results explored the role of the family in smoking behaviour. Teachers do not feel responsibility in prevention or cessation of smoking. Adolescents are not aware of the addictive effect of the cigarette. Water pipe is popular among adolescent, therefore spreading of more information needed on the harmful effects of water pipe. We outlined the requirement for professionals to provide interventions. Partnership of EU-FAQT project has worked out antismoking interventions for adolescents and their families conducted by psychologists to ensure skill development to prevent and quit tobacco.

Keywords: Smoking of adolescents, family approach, focus group analysis, water pipe.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1806
3831 Investigations of Flame Retardant Properties of Beneficiated Huntite and Hydromagnesite Mineral Reinforced Polymer Composites

Authors: H. Yilmaz Atay

Abstract:

Huntite and hydromagnesite minerals have been used as additive materials to achieve incombustible material due to their inflammability property. Those fire retardants materials can help to extinguish in the early stages of fire. Thus dispersion of the flame can be prevented even if the fire started. Huntite and hydromagnesite minerals are known to impart fire-proofing of the polymer composites. However, the additives used in the applications led to deterioration in the mechanical properties due to the usage of high amount of the powders in the composites. In this study, by enriching huntite and hydromagnesite, it was aimed to use purer minerals to reinforce the polymer composites. Thus, predictably, using purer mineral will lead to use lower amount of mineral powders. By this manner, the minerals free from impurities by various processes were added to the polymer matrix with different loading level and grades. Different types of samples were manufactured, and subsequently characterized by XRD, SEM-EDS, XRF and flame-retardant tests. Tensile strength and elongation at break values were determined according to loading levels and grades. Besides, a comparison on the properties of the polymer composites produced by using of minerals with and without impurities was performed. As a result of the work, it was concluded that it is required to use beneficiated minerals to provide better fire-proofing behaviors in the polymer composites.

Keywords: Huntite, hdromagnesite, flame retardant, mechanical property, polymeric composites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 845
3830 Pricing European Options under Jump Diffusion Models with Fast L-stable Padé Scheme

Authors: Salah Alrabeei, Mohammad Yousuf

Abstract:

The goal of option pricing theory is to help the investors to manage their money, enhance returns and control their financial future by theoretically valuing their options. Modeling option pricing by Black-School models with jumps guarantees to consider the market movement. However, only numerical methods can solve this model. Furthermore, not all the numerical methods are efficient to solve these models because they have nonsmoothing payoffs or discontinuous derivatives at the exercise price. In this paper, the exponential time differencing (ETD) method is applied for solving partial integrodifferential equations arising in pricing European options under Merton’s and Kou’s jump-diffusion models. Fast Fourier Transform (FFT) algorithm is used as a matrix-vector multiplication solver, which reduces the complexity from O(M2) into O(M logM). A partial fraction form of Pad`e schemes is used to overcome the complexity of inverting polynomial of matrices. These two tools guarantee to get efficient and accurate numerical solutions. We construct a parallel and easy to implement a version of the numerical scheme. Numerical experiments are given to show how fast and accurate is our scheme.

Keywords: Integral differential equations, L-stable methods, pricing European options, Jump–diffusion model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 499