Search results for: software product line
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10353

Search results for: software product line

10083 Analysis of the Level of Production Failures by Implementing New Assembly Line

Authors: Joanna Kochanska, Dagmara Gornicka, Anna Burduk

Abstract:

The article examines the process of implementing a new assembly line in a manufacturing enterprise of the household appliances industry area. At the initial stages of the project, a decision was made that one of its foundations should be the concept of lean management. Because of that, eliminating as many errors as possible in the first phases of its functioning was emphasized. During the start-up of the line, there were identified and documented all production losses (from serious machine failures, through any unplanned downtime, to micro-stops and quality defects). During 6 weeks (line start-up period), all errors resulting from problems in various areas were analyzed. These areas were, among the others, production, logistics, quality, and organization. The aim of the work was to analyze the occurrence of production failures during the initial phase of starting up the line and to propose a method for determining their critical level during its full functionality. There was examined the repeatability of the production losses in various areas and at different levels at such an early stage of implementation, by using the methods of statistical process control. Based on the Pareto analysis, there were identified the weakest points in order to focus improvement actions on them. The next step was to examine the effectiveness of the actions undertaken to reduce the level of recorded losses. Based on the obtained results, there was proposed a method for determining the critical failures level in the studied areas. The developed coefficient can be used as an alarm in case of imbalance of the production, which is caused by the increased failures level in production and production support processes in the period of the standardized functioning of the line.

Keywords: production failures, level of production losses, new production line implementation, assembly line, statistical process control

Procedia PDF Downloads 114
10082 The Different Improvement of Numerical Magnitude and Spatial Representation of Numbers to Symbolic Approximate Arithmetic: A Training Study of Preschooler

Authors: Yu Liang, Wei Wei

Abstract:

Spatial representation of numbers and numerical magnitude are important for preschoolers’ mathematical ability. Mental number line, a typical index to measure numbers spatial representation, and numerical comparison are both related to arithmetic obviously. However, they seem to rely on different mechanisms and probably influence arithmetic through different mechanisms. In line with this idea, preschool children were trained with two tasks to investigate which one is more important for approximate arithmetic. The training of numerical processing and number line estimation were proved to be effective. They both improved the ability of approximate arithmetic. When the difficulty of approximate arithmetic was taken into account, the performance in number line training group was not significantly different among three levels. However, two harder levels achieved significance in numerical comparison training group. Thus, comparing spatial representation ability, symbolic approximation arithmetic relies more on numerical magnitude. Educational implications of the study were discussed.

Keywords: approximate arithmetic, mental number line, numerical magnitude, preschooler

Procedia PDF Downloads 233
10081 Detection of Abnormal Process Behavior in Copper Solvent Extraction by Principal Component Analysis

Authors: Kirill Filianin, Satu-Pia Reinikainen, Tuomo Sainio

Abstract:

Frequent measurements of product steam quality create a data overload that becomes more and more difficult to handle. In the current study, plant history data with multiple variables was successfully treated by principal component analysis to detect abnormal process behavior, particularly, in copper solvent extraction. The multivariate model is based on the concentration levels of main process metals recorded by the industrial on-stream x-ray fluorescence analyzer. After mean-centering and normalization of concentration data set, two-dimensional multivariate model under principal component analysis algorithm was constructed. Normal operating conditions were defined through control limits that were assigned to squared score values on x-axis and to residual values on y-axis. 80 percent of the data set were taken as the training set and the multivariate model was tested with the remaining 20 percent of data. Model testing showed successful application of control limits to detect abnormal behavior of copper solvent extraction process as early warnings. Compared to the conventional techniques of analyzing one variable at a time, the proposed model allows to detect on-line a process failure using information from all process variables simultaneously. Complex industrial equipment combined with advanced mathematical tools may be used for on-line monitoring both of process streams’ composition and final product quality. Defining normal operating conditions of the process supports reliable decision making in a process control room. Thus, industrial x-ray fluorescence analyzers equipped with integrated data processing toolbox allows more flexibility in copper plant operation. The additional multivariate process control and monitoring procedures are recommended to apply separately for the major components and for the impurities. Principal component analysis may be utilized not only in control of major elements’ content in process streams, but also for continuous monitoring of plant feed. The proposed approach has a potential in on-line instrumentation providing fast, robust and cheap application with automation abilities.

Keywords: abnormal process behavior, failure detection, principal component analysis, solvent extraction

Procedia PDF Downloads 293
10080 A Simulative Approach for JIT Parts-Feeding Policies

Authors: Zhou BingHai, Fradet Victor

Abstract:

Lean philosophy follows the simple principle of “creating more value with fewer resources”. In accordance with this policy, material handling can be managed by the mean of Kanban which by triggering every feeding tour only when needed regulates the flow of material in one of the most efficient way. This paper focuses on Kanban Supermarket’s parameters and their optimization on a purely cost-based point of view. Number and size of forklifts, as well as size of the containers they carry, will be variables of the cost function which includes handling costs, inventory costs but also shortage costs. With an innovative computational approach encoded into industrial engineering software Tecnomatix and reproducing real-life conditions, a fictive assembly line is established and produces a random list of orders. Multi-scenarios are then run to study the impact of each change of parameter and the variation of costs it implies. Lastly, best-case scenarios financially speaking are selected.

Keywords: Kanban, supermarket, parts-feeding policies, multi-scenario simulation, assembly line

Procedia PDF Downloads 182
10079 Code Refactoring Using Slice-Based Cohesion Metrics and AOP

Authors: Jagannath Singh, Durga Prasad Mohapatra

Abstract:

Software refactoring is very essential for maintaining the software quality. It is an usual practice that we first design the software and then go for coding. But after coding is completed, if the requirement changes slightly or our expected output is not achieved, then we change the codes. For each small code change, we cannot change the design. In course of time, due to these small changes made to the code, the software design decays. Software refactoring is used to restructure the code in order to improve the design and quality of the software. In this paper, we propose an approach for performing code refactoring. We use slice-based cohesion metrics to identify the target methods which requires refactoring. After identifying the target methods, we use program slicing to divide the target method into two parts. Finally, we have used the concepts of Aspects to adjust the code structure so that the external behaviour of the original module does not change.

Keywords: software refactoring, program slicing, AOP, cohesion metrics, code restructure, AspectJ

Procedia PDF Downloads 488
10078 A Study of Growth Performance, Carcass Characteristic, Meat Quality and Association of Polymorphism in the ApoVLDL-II Gene with Fat Accumulation in the Female Broiler, Thai Native and Betong Chickens (KU Line)

Authors: C. Kridtayopas, W. Danvilai, P. Sopannarath, A. Kayan, W. Loongyai

Abstract:

Both Betong chicken (KU Line) and Thai Native chickens were the high quality of the meat and low carcass fat compared to broiler chickens. The objective of this study was to determine the growth performance, carcass characteristic, meat quality and association of polymorphism in the ApoVLDL-II gene with fat accumulation in the female broiler, Thai Native and Betong (KU line) chickens at 4-14 weeks. The chickens were used and reared under the same environment and management (100 chicks per breed). The results showed that body weight (BW) of broiler chickens was significantly higher than Thai Native and Betong (KU line) chickens (P < 0.01) through all the experiment. At 4-8 weeks of age, feed conversion ratio (FCR) of broiler chickens was significantly better than Thai Native and Betong (KU line) chickens (P < 0.01), then increased at week 8-14. The percentage of breast, abdominal fat and subcutaneous fat of broiler chickens was significantly greater than Thai Native and Betong (KU line) chickens (P < 0.01). However, Thai Native chickens showed the highest percentage of liver (P < 0.01) when compared to other breeds. In addition, the percentage of wing of Thai Native and Betong (KU line) chickens were significantly (P < 0.01) higher than broiler chickens. Meat quality was also determined and found that, pH of breast meat left from slaughter 45 minutes (pH45) and 24 hours (pH24) of broiler was significantly higher than Thai Native and Betong (KU line) (P < 0.01) whereas the percentage of drip loss, thawing loss, cooking loss and shear force was not significantly different between breeds. The polymerase chain reaction-restriction fragment length polymorphism (PCR-RFLP) technique was used to genotype the polymorphism in the ApoVLDL-II gene in the broiler, Thai Native and Betong (KU line) chickens. The results found that, the polymorphism in the ApoVLDL-II gene at VLDL6 loci was not associated with fat accumulation in those studied population.

Keywords: ApoVLDL-II gene, Betong (KU line) chickens, broiler chickens, carcass characteristic, growth performance, meat quality, Thai native chickens

Procedia PDF Downloads 186
10077 Study of the Effect of the Continuous Electric Field on the Rd Cancer Cell Line by Response Surface Methodology

Authors: Radia Chemlal, Salim Mehenni, Dahbia Leila Anes-boulahbal, Mohamed Kherat, Nabil Mameri

Abstract:

The application of the electric field is considered to be a very promising method in cancer therapy. Indeed, cancer cells are very sensitive to the electric field, although the cellular response is not entirely clear. The tests carried out consisted in subjecting the RD cell line under the effect of the continuous electric field while varying certain parameters (voltage, exposure time, and cell concentration). The response surface methodology (RSM) was used to assess the effect of the chosen parameters, as well as the existence of interactions between them. The results obtained showed that the voltage, the cell concentration as well as the interaction between voltage and exposure time have an influence on the mortality rate of the RD cell line.

Keywords: continuous electric field, RD cancer cell line, RSM, voltage

Procedia PDF Downloads 92
10076 User-Perceived Quality Factors for Certification Model of Web-Based System

Authors: Jamaiah H. Yahaya, Aziz Deraman, Abdul Razak Hamdan, Yusmadi Yah Jusoh

Abstract:

One of the most essential issues in software products is to maintain it relevancy to the dynamics of the user’s requirements and expectation. Many studies have been carried out in quality aspect of software products to overcome these problems. Previous software quality assessment models and metrics have been introduced with strengths and limitations. In order to enhance the assurance and buoyancy of the software products, certification models have been introduced and developed. From our previous experiences in certification exercises and case studies collaborating with several agencies in Malaysia, the requirements for user based software certification approach is identified and demanded. The emergence of social network applications, the new development approach such as agile method and other varieties of software in the market have led to the domination of users over the software. As software become more accessible to the public through internet applications, users are becoming more critical in the quality of the services provided by the software. There are several categories of users in web-based systems with different interests and perspectives. The classifications and metrics are identified through brain storming approach with includes researchers, users and experts in this area. The new paradigm in software quality assessment is the main focus in our research. This paper discusses the classifications of users in web-based software system assessment and their associated factors and metrics for quality measurement. The quality model is derived based on IEEE structure and FCM model. The developments are beneficial and valuable to overcome the constraints and improve the application of software certification model in future.

Keywords: software certification model, user centric approach, software quality factors, metrics and measurements, web-based system

Procedia PDF Downloads 387
10075 Product Form Bionic Design Based on Eye Tracking Data: A Case Study of Desk Lamp

Authors: Huan Lin, Liwen Pang

Abstract:

In order to reduce the ambiguity and uncertainty of product form bionic design, a product form bionic design method based on eye tracking is proposed. The eye-tracking experiment is designed to calculate the average time ranking of the specific parts of the bionic shape that the subjects are looking at. Key bionic shape is explored through the experiment and then applied to a desk lamp bionic design. During the design case, FAHP (Fuzzy Analytic Hierachy Process) and SD (Semantic Differential) method are firstly used to identify consumer emotional perception model toward desk lamp before product design. Through investigating different desk lamp design elements and consumer views, the form design factors on the desk lamp product are reflected and all design schemes are sequenced after caculation. Desk lamp form bionic design method is combined the key bionic shape extracted from eye-tracking experiment and priority of desk lamp design schemes. This study provides an objective and rational method to product form bionic design.

Keywords: Bionic design; Form; Eye tracking; FAHP; Desk lamp

Procedia PDF Downloads 197
10074 Linking Business Process Models and System Models Based on Business Process Modelling

Authors: Faisal A. Aburub

Abstract:

Organizations today need to invest in software in order to run their businesses, and to the organizations’ objectives, the software should be in line with the business process. This research presents an approach for linking process models and system models. Particularly, the new approach aims to synthesize sequence diagram based on role activity diagram (RAD) model. The approach includes four steps namely: Create business process model using RAD, identify computerized activities, identify entities in sequence diagram and identify messages in sequence diagram. The new approach has been validated using the process of student registration in University of Petra as a case study. Further research is required to validate the new approach using different domains.

Keywords: business process modelling, system models, role activity diagrams, sequence diagrams

Procedia PDF Downloads 366
10073 Degree of Approximation of Functions by Product Means

Authors: Hare Krishna Nigam

Abstract:

In this paper, for the first time, (E,q)(C,2) product summability method is introduced and two quite new results on degree of approximation of the function f belonging to Lip (alpha,r)class and W(L(r), xi(t)) class by (E,q)(C,2) product means of Fourier series, has been obtained.

Keywords: Degree of approximation, (E, q)(C, 2) means, Fourier series, Lebesgue integral, Lip (alpha, r)class, W(L(r), xi(t))class of functions

Procedia PDF Downloads 497
10072 New Standardized Framework for Developing Mobile Applications (Based On Real Case Studies and CMMI)

Authors: Ammar Khader Almasri

Abstract:

The software processes play a vital role for delivering a high quality software system that meets the user’s needs. There are many software development models which are used by most system developers, which can be categorized into two categories (traditional and new methodologies). Mobile applications like other desktop applications need appropriate and well-working software development process. Nevertheless, mobile applications have different features which limit their performance and efficiency like application size, mobile hardware features. Moreover, this research aims to help developers in using a standardized model for developing mobile applications.

Keywords: software development process, agile methods , moblile application development, traditional methods

Procedia PDF Downloads 371
10071 Secure Multiparty Computations for Privacy Preserving Classifiers

Authors: M. Sumana, K. S. Hareesha

Abstract:

Secure computations are essential while performing privacy preserving data mining. Distributed privacy preserving data mining involve two to more sites that cannot pool in their data to a third party due to the violation of law regarding the individual. Hence in order to model the private data without compromising privacy and information loss, secure multiparty computations are used. Secure computations of product, mean, variance, dot product, sigmoid function using the additive and multiplicative homomorphic property is discussed. The computations are performed on vertically partitioned data with a single site holding the class value.

Keywords: homomorphic property, secure product, secure mean and variance, secure dot product, vertically partitioned data

Procedia PDF Downloads 400
10070 Hybrid Approach for Software Defect Prediction Using Machine Learning with Optimization Technique

Authors: C. Manjula, Lilly Florence

Abstract:

Software technology is developing rapidly which leads to the growth of various industries. Now-a-days, software-based applications have been adopted widely for business purposes. For any software industry, development of reliable software is becoming a challenging task because a faulty software module may be harmful for the growth of industry and business. Hence there is a need to develop techniques which can be used for early prediction of software defects. Due to complexities in manual prediction, automated software defect prediction techniques have been introduced. These techniques are based on the pattern learning from the previous software versions and finding the defects in the current version. These techniques have attracted researchers due to their significant impact on industrial growth by identifying the bugs in software. Based on this, several researches have been carried out but achieving desirable defect prediction performance is still a challenging task. To address this issue, here we present a machine learning based hybrid technique for software defect prediction. First of all, Genetic Algorithm (GA) is presented where an improved fitness function is used for better optimization of features in data sets. Later, these features are processed through Decision Tree (DT) classification model. Finally, an experimental study is presented where results from the proposed GA-DT based hybrid approach is compared with those from the DT classification technique. The results show that the proposed hybrid approach achieves better classification accuracy.

Keywords: decision tree, genetic algorithm, machine learning, software defect prediction

Procedia PDF Downloads 316
10069 Assessing the Current State of Software Engineering and Information Technology in Ghana

Authors: David Yartel

Abstract:

Drawing on the current state of software engineering and information technology in Ghana, the study documents its significant contribution to the development of Ghanaian industries. The study focuses on the application of modern trends in technology and the barriers faced in the area of software engineering and information technology. A thorough analysis of a dozen of interviews with stakeholders in software engineering and information technology via interviews reveals how modern trends in software engineering pose challenges to the industry in Ghana. Results show that to meet the expectation of modern software engineering and information technology trends, stakeholders must have skilled professionals, adequate infrastructure, and enhanced support for technology startups. Again, individuals should be encouraged to pursue a career in software engineering and information technology, as it has the propensity to increase the efficiency and effectiveness of work-related activities. This study recommends that stakeholders in software engineering and technology industries should invest enough in training more professionals by collaborating with international institutions well-versed in the area by organizing frequent training and seminars. The government should also provide funding opportunities for small businesses in the technology sector to drive creativity and development in order to bring about growth and development.

Keywords: software engineering, information technology, Ghana, development

Procedia PDF Downloads 71
10068 Additive Manufacturing’s Impact on Product Design and Development: An Industrial Case Study

Authors: Ahmed Abdelsalam, Daniel Roozbahani, Marjan Alizadeh, Heikki Handroos

Abstract:

The aim of this study was to redesign a pressing air nozzle with lower weight and improved efficiency utilizing Selective Laser Melting (SLM) technology based on Design for Additive Manufacturing (DfAM) methods. The original pressing air nozzle was modified in SolidWorks 3D CAD, and two design concepts were introduced considering the DfAM approach. In the proposed designs, the air channels were amended. 3D models for the original pressing air nozzle and introduced designs were created to obtain the flow characteristic data using Ansys software. Results of CFD modeling for the original and two proposed designs were extracted, compared, and analyzed to demonstrate the impact of design on the development of a more efficient pressing air nozzle by AM process. Improved airflow was achieved by optimizing the pressing air nozzle's internal channel for both design concepts by providing 30% and 50.6% fewer pressure drops than the original design. Moreover, utilizing the presented designs, a significant reduction in product weight was attained. In addition, by applying the proposed designs, 48.3% and 70.3% reduction in product weight was attained compared to the original design. Therefore, pressing air nozzle with enhanced productivity and lowered weight was generated utilizing the DfAM-driven designs developed in this study. The main contribution of this study is to investigate the additional possibilities that can be achieved in designing modern parts using the advantage of SLM technology in producing that part. The approach presented in this study can be applied to almost any similar industrial application.

Keywords: additive manufacturing, design for additive manufacturing, design methods, product design, pressing air nozzle

Procedia PDF Downloads 111
10067 The Evaluation Model for the Quality of Software Based on Open Source Code

Authors: Li Donghong, Peng Fuyang, Yang Guanghua, Su Xiaoyan

Abstract:

Using open source code is a popular method of software development. How to evaluate the quality of software becomes more important. This paper introduces an evaluation model. The model evaluates the quality from four dimensions: technology, production, management, and development. Each dimension includes many indicators. The weight of indicator can be modified according to the purpose of evaluation. The paper also introduces a method of using the model. The evaluating result can provide good advice for evaluating or purchasing the software.

Keywords: evaluation model, software quality, open source code, evaluation indicator

Procedia PDF Downloads 369
10066 Impact of Long-Term Orientation on Product Quality in Supply Chain: An Empirical Analysis

Authors: Qingyu Zhang, Mei Cao

Abstract:

As the environments become increasingly uncertain, firms have attempted to achieve greater supply chain collaboration. Supply chain collaboration can generate significant benefits to its members, e.g., reducing risks and decreasing transaction costs. However, a strong relationship is often related to firm’s culture (e.g., short-term vs. long-term interests). The objective of the study is to explore the effect of long-term oriented culture on product quality in a supply chain. Data was collected through a Web survey of U.S. manufacturing firms. Structural equation modeling (LISREL) was used to analyze the data. The results support the mediating roles of goal congruence and communication in the relationship between long-term orientation and product quality in the supply chain. Goal congruence partially mediates the relationship between long-term orientation and communication; communication completely mediates the relationship between goal congruence and product quality. Without high levels of communication, goal congruence cannot improve product quality in a positive way.

Keywords: communication, long-term orientation, product quality, supply chain

Procedia PDF Downloads 318
10065 Test Suite Optimization Using an Effective Meta-Heuristic BAT Algorithm

Authors: Anuradha Chug, Sunali Gandhi

Abstract:

Regression Testing is a very expensive and time-consuming process carried out to ensure the validity of modified software. Due to the availability of insufficient resources to re-execute all the test cases in time constrained environment, efforts are going on to generate test data automatically without human efforts. Many search based techniques have been proposed to generate efficient, effective as well as optimized test data, so that the overall cost of the software testing can be minimized. The generated test data should be able to uncover all potential lapses that exist in the software or product. Inspired from the natural behavior of bat for searching her food sources, current study employed a meta-heuristic, search-based bat algorithm for optimizing the test data on the basis certain parameters without compromising their effectiveness. Mathematical functions are also applied that can effectively filter out the redundant test data. As many as 50 Java programs are used to check the effectiveness of proposed test data generation and it has been found that 86% saving in testing efforts can be achieved using bat algorithm while covering 100% of the software code for testing. Bat algorithm was found to be more efficient in terms of simplicity and flexibility when the results were compared with another nature inspired algorithms such as Firefly Algorithm (FA), Hill Climbing Algorithm (HC) and Ant Colony Optimization (ACO). The output of this study would be useful to testers as they can achieve 100% path coverage for testing with minimum number of test cases.

Keywords: regression testing, test case selection, test case prioritization, genetic algorithm, bat algorithm

Procedia PDF Downloads 355
10064 Pigging Operation in Two-Phase Flow Pipeline- Empirical and Simulation

Authors: Behnaz Jamshidi, Seyed Hassan Hashemabadi

Abstract:

The main objective of this study is to investigate on pigging operation of two phase flow pipeline and compare the empirical and simulation results for 108 km long , 0.7934 mm (32 inches) diameter sea line of "Phase 1 South Pars Gas Complex", located in south of Iran. The pigging time, pig velocity, the amount of slug and slug catcher pressure were calculated and monitored closely as the key parameters. Simulation was done by "OLGA" dynamic simulation software and obtained results were compared and validated with empirical data in real operation. The relative errors between empirical data and simulation of the process were 3 % and 9 % for pigging time and accumulated slug volume respectively. Simulated pig velocity and changes of slug catcher pressure were consistent with real values, too. It was also found the slug catcher and condensate stabilization units have been adequately sized for gas-liquid separation and handle the slug batch during transient conditions such as pigging and start up.

Keywords: sea line, pigging, slug catcher, two-phase flow, dynamic simulation

Procedia PDF Downloads 480
10063 Extending the AOP Joinpoint Model for Memory and Type Safety

Authors: Amjad Nusayr

Abstract:

Software security is a general term used to any type of software architecture or model in which security aspects are incorporated in this architecture. These aspects are not part of the main logic of the underlying program. Software security can be achieved using a combination of approaches, including but not limited to secure software designs, third part component validation, and secure coding practices. Memory safety is one feature in software security where we ensure that any object in memory has a valid pointer or a reference with a valid type. Aspect-Oriented Programming (AOP) is a paradigm that is concerned with capturing the cross-cutting concerns in code development. AOP is generally used for common cross-cutting concerns like logging and DB transaction managing. In this paper, we introduce the concepts that enable AOP to be used for the purpose of memory and type safety. We also present ideas for extending AOP in software security practices.

Keywords: aspect oriented programming, programming languages, software security, memory and type safety

Procedia PDF Downloads 112
10062 Developing a Modular Architecture of Apparel Product

Authors: Yu Zhao, Mengqin Sun, Yahui Zhang

Abstract:

Apparel products (or apparel) with the sense of aesthetics, usability (ergonomics) and function are fundamental and varied in people’s daily life. The numerous apparel thus produced by apparel industry, have been triggered many issues, such as the waste of sources and the environmental pollutions. In this study, a hybrid architecture called modular architecture of apparel (MAA) has been proposed to deal with the variety of apparel, and thus to overcome the aforementioned issues. Generally, the establishment of MAA takes advantage of the modular design of a general product that a product is assembled with many modules through their modular interface connector. The development of MAA is to first analyze the structure of apparel in terms of the necessity to form an apparel and the aesthetics, ergonomics, and function of apparel; then to divide apparel into many segments (or module in product design) based on the structure of apparel; to develop modular interfaces and modular interface connectors in terms of the features of apparel’s modules. It is noted that in the general product design, modules of a product are only about the function and ergonomics, but in MAA, the module of aesthetics is developed. Further, an apparel design with employing the MAA is carried out to validate its usefulness and efficiency. There are three contributions out of this study, the first is to overcome the aforementioned issues (i.e. waste of source and environmental pollutions); the second is the improvement of the modular design for product by considering aesthetics; the third is to add the value in realizing the personalized mass production of apparel in the near future.

Keywords: apparel, architecture, modular design, segment

Procedia PDF Downloads 264
10061 Recent Advances in Data Warehouse

Authors: Fahad Hanash Alzahrani

Abstract:

This paper describes some recent advances in a quickly developing area of data storing and processing based on Data Warehouses and Data Mining techniques, which are associated with software, hardware, data mining algorithms and visualisation techniques having common features for any specific problems and tasks of their implementation.

Keywords: data warehouse, data mining, knowledge discovery in databases, on-line analytical processing

Procedia PDF Downloads 382
10060 A Petri Net Model to Obtain the Throughput of Unreliable Production Lines in the Buffer Allocation Problem

Authors: Joselito Medina-Marin, Alexandr Karelin, Ana Tarasenko, Juan Carlos Seck-Tuoh-Mora, Norberto Hernandez-Romero, Eva Selene Hernandez-Gress

Abstract:

A production line designer faces with several challenges in manufacturing system design. One of them is the assignment of buffer slots in between every machine of the production line in order to maximize the throughput of the whole line, which is known as the Buffer Allocation Problem (BAP). The BAP is a combinatorial problem that depends on the number of machines and the total number of slots to be distributed on the production line. In this paper, we are proposing a Petri Net (PN) Model to obtain the throughput in unreliable production lines, based on PN mathematical tools and the decomposition method. The results obtained by this methodology are similar to those presented in previous works, and the number of machines is not a hard restriction.

Keywords: buffer allocation problem, Petri Nets, throughput, production lines

Procedia PDF Downloads 288
10059 Development of AUTOSAR Software Components of MDPS System

Authors: Jae-Woo Kim, Kyung-Joong Lee, Hyun-Sik Ahn

Abstract:

This paper describes the development of a Motor-Driven Power Steering (MDPS) system using Automotive Open System Architecture (AUTOSAR) methodology. The MDPS system is a new power steering technology for vehicles and it can enhance driver’s convenience and fuel efficiency. AUTOSAR defines common standards for the implementation of embedded automotive software. Some aspects of safety and timing requirements are analyzed. Through the AUTOSAR methodology, the embedded software becomes more flexible, reusable and maintainable than ever. Hence, we first design software components (SW-C) for MDPS control based on AUTOSAR and implement SW-Cs for MDPS control using authoring tool following AUTOSAR standards.

Keywords: AUTOSAR, MDPS, simulink, software component

Procedia PDF Downloads 337
10058 Conceptualization of Value Co-Creation for Shrimp Products in Bangladesh

Authors: Subarna Ferdous, Mitsuru Ikeda

Abstract:

For the shrimp companies to remain relevant to its local and international consumers, they must offer new shrimp product and services. It must work actively not just to create value for the consumer, but to involve the consumer in co-creating value for shrimp product innovation in the market. In this theoretical work, we conceptualize the business concept of value co-creation in the context of shrimp products, and propose a framework of value co-creation for shrimp product innovation in shrimp industries. With guidance on value co-creation in in shrimp industry, and shrimp value chain actors mapped to the co-creation cycle, companies can use the framework to offer new shrimp product to consumer communities. Although customer co-creation is known approach in the world, it is not commonly used by the companies in Bangladesh. This paper makes an original contribution by conceptualizing co-creation and set the examples of best co-creation practices in food sector. The results of the study provide management with guidelines for successful co-creation projects with an innovation- and market-oriented approach. The framework also provides a basis for further research in this area.

Keywords: bangladesh, shrimp industry, value co-creation, shrimp product

Procedia PDF Downloads 499
10057 Cognitive Weighted Polymorphism Factor: A New Cognitive Complexity Metric

Authors: T. Francis Thamburaj, A. Aloysius

Abstract:

Polymorphism is one of the main pillars of the object-oriented paradigm. It induces hidden forms of class dependencies which may impact software quality, resulting in higher cost factor for comprehending, debugging, testing, and maintaining the software. In this paper, a new cognitive complexity metric called Cognitive Weighted Polymorphism Factor (CWPF) is proposed. Apart from the software structural complexity, it includes the cognitive complexity on the basis of type. The cognitive weights are calibrated based on 27 empirical studies with 120 persons. A case study and experimentation of the new software metric shows positive results. Further, a comparative study is made and the correlation test has proved that CWPF complexity metric is a better, more comprehensive, and more realistic indicator of the software complexity than Abreu’s Polymorphism Factor (PF) complexity metric.

Keywords: cognitive complexity metric, object-oriented metrics, polymorphism factor, software metrics

Procedia PDF Downloads 433
10056 Improved Dynamic Bayesian Networks Applied to Arabic On Line Characters Recognition

Authors: Redouane Tlemsani, Abdelkader Benyettou

Abstract:

Work is in on line Arabic character recognition and the principal motivation is to study the Arab manuscript with on line technology. This system is a Markovian system, which one can see as like a Dynamic Bayesian Network (DBN). One of the major interests of these systems resides in the complete models training (topology and parameters) starting from training data. Our approach is based on the dynamic Bayesian Networks formalism. The DBNs theory is a Bayesians networks generalization to the dynamic processes. Among our objective, amounts finding better parameters, which represent the links (dependences) between dynamic network variables. In applications in pattern recognition, one will carry out the fixing of the structure, which obliges us to admit some strong assumptions (for example independence between some variables). Our application will relate to the Arabic isolated characters on line recognition using our laboratory database: NOUN. A neural tester proposed for DBN external optimization. The DBN scores and DBN mixed are respectively 70.24% and 62.50%, which lets predict their further development; other approaches taking account time were considered and implemented until obtaining a significant recognition rate 94.79%.

Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, computer vision

Procedia PDF Downloads 414
10055 The Development of Encrypted Near Field Communication Data Exchange Format Transmission in an NFC Passive Tag for Checking the Genuine Product

Authors: Tanawat Hongthai, Dusit Thanapatay

Abstract:

This paper presents the development of encrypted near field communication (NFC) data exchange format transmission in an NFC passive tag for the feasibility of implementing a genuine product authentication. We propose a research encryption and checking the genuine product into four major categories; concept, infrastructure, development and applications. This result shows the passive NFC-forum Type 2 tag can be configured to be compatible with the NFC data exchange format (NDEF), which can be automatically partially data updated when there is NFC field.

Keywords: near field communication, NFC data exchange format, checking the genuine product, encrypted NFC

Procedia PDF Downloads 264
10054 Gaming Mouse Redesign Based on Evaluation of Pragmatic and Hedonic Aspects of User Experience

Authors: Thedy Yogasara, Fredy Agus

Abstract:

In designing a product, it is currently crucial to focus not only on the product’s usability based on performance measures, but also on user experience (UX) that includes pragmatic and hedonic aspects of product use. These aspects play a significant role in fulfillment of user needs, both functionally and psychologically. Pragmatic quality refers to as product’s perceived ability to support the fulfillment of behavioral goals. It is closely linked to functionality and usability of the product. In contrast, hedonic quality is product’s perceived ability to support the fulfillment of psychological needs. Hedonic quality relates to the pleasure of ownership and use of the product, including stimulation for personal development and communication of user’s identity to others through the product. This study evaluates the pragmatic and hedonic aspects of gaming mice G600 and Razer Krait using AttrakDiff tool to create an improved design that is able to generate positive UX. AttrakDiff is a method that measures pragmatic and hedonic scores of a product with a scale between -3 to +3 through four attributes (i.e. Pragmatic Quality, Hedonic Quality-Identification, Hedonic Quality-Stimulation, and Attractiveness), represented by 28 pairs of opposite words. Based on data gathered from 15 participants, it is identified that gaming mouse G600 needs to be redesigned because of its low grades (pragmatic score: -0.838, hedonic score: 1, attractiveness score: 0.771). The redesign process focuses on the attributes with poor scores and takes into account improvement suggestions collected from interview with the participants. The redesigned mouse G600 is evaluated using the previous method. The result shows higher scores in pragmatic quality (1.929), hedonic quality (1.703), and attractiveness (1.667), indicating that the redesigned mouse is more capable of creating pleasurable experience of product use.

Keywords: AttrakDiff, hedonic aspect, pragmatic aspect, product design, user experience

Procedia PDF Downloads 140