Search results for: optimization for learning and data analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15650

Search results for: optimization for learning and data analysis

10460 Bayes Net Classifiers for Prediction of Renal Graft Status and Survival Period

Authors: Jiakai Li, Gursel Serpen, Steven Selman, Matt Franchetti, Mike Riesen, Cynthia Schneider

Abstract:

This paper presents the development of a Bayesian belief network classifier for prediction of graft status and survival period in renal transplantation using the patient profile information prior to the transplantation. The objective was to explore feasibility of developing a decision making tool for identifying the most suitable recipient among the candidate pool members. The dataset was compiled from the University of Toledo Medical Center Hospital patients as reported to the United Network Organ Sharing, and had 1228 patient records for the period covering 1987 through 2009. The Bayes net classifiers were developed using the Weka machine learning software workbench. Two separate classifiers were induced from the data set, one to predict the status of the graft as either failed or living, and a second classifier to predict the graft survival period. The classifier for graft status prediction performed very well with a prediction accuracy of 97.8% and true positive values of 0.967 and 0.988 for the living and failed classes, respectively. The second classifier to predict the graft survival period yielded a prediction accuracy of 68.2% and a true positive rate of 0.85 for the class representing those instances with kidneys failing during the first year following transplantation. Simulation results indicated that it is feasible to develop a successful Bayesian belief network classifier for prediction of graft status, but not the graft survival period, using the information in UNOS database.

Keywords: Bayesian network classifier, renal transplantation, graft survival period, United Network for Organ Sharing

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2098
10459 Evaluation of the Contribution of Starting Pitchers in a Professional Baseball Team by Grey Relational Analysis

Authors: Chih-Cheng Chen, Yung-Tan Lee, Shih-Yang Lee, Shih-Kuei Huang, Tien-Tze Chen, Qiu-Jun Chen

Abstract:

The evaluation of the contribution of professional baseball starting pitchers is a complex decision-making problem that includes several quantitative attributes. It is considered a type of multi-attribute or multi-criteria decision making (MADM/MCDM) problem. This study proposes a model using the Grey Relational Analysis (GRA) to evaluate the starting pitcher contribution for teams of the Chinese Professional Baseball League. The GRA calculates the individual grey relational degree of each alternative to the positive ideal alternative. An empirical analysis was conducted to show the use of the model for the starting pitcher contribution problem. The results demonstrate the effectiveness and feasibility of the proposed model.

Keywords: Starting pitchers, Grey Relational Analysis, Chinese Professional Baseball

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1617
10458 The Integration of Patient Health Record Generated from Wearable and Internet of Things Devices into Health Information Exchanges

Authors: Dalvin D. Hill, Hector M. Castro Garcia

Abstract:

A growing number of individuals utilize wearable devices on a daily basis. The usage and functionality of these wearable devices vary from user to user. One popular usage of said devices is to track health-related activities that are typically stored on a device’s memory or uploaded to an account in the cloud; based on the current trend, the data accumulated from the wearable device are stored in a standalone location. In many of these cases, this health related datum is not a factor when considering the holistic view of a user’s health lifestyle or record. This health-related data generated from wearable and Internet of Things (IoT) devices can serve as empirical information to a medical provider, as the standalone data can add value to the holistic health record of a patient. This paper proposes a solution to incorporate the data gathered from these wearable and IoT devices, with that a patient’s Personal Health Record (PHR) stored within the confines of a Health Information Exchange (HIE).

Keywords: Electronic health record, health information exchanges, Internet of Things, personal health records, wearable devices, wearables.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1742
10457 Neural Network-Based Control Strategies Applied to a Fed-Batch Crystallization Process

Authors: P. Georgieva, S. Feyo de Azevedo

Abstract:

This paper is focused on issues of process modeling and two model based control strategies of a fed-batch sugar crystallization process applying the concept of artificial neural networks (ANNs). The control objective is to force the operation into following optimal supersaturation trajectory. It is achieved by manipulating the feed flow rate of sugar liquor/syrup, considered as the control input. The control task is rather challenging due to the strong nonlinearity of the process dynamics and variations in the crystallization kinetics. Two control alternatives are considered – model predictive control (MPC) and feedback linearizing control (FLC). Adequate ANN process models are first built as part of the controller structures. MPC algorithm outperforms the FLC approach with respect to satisfactory reference tracking and smooth control action. However, the MPC is computationally much more involved since it requires an online numerical optimization, while for the FLC an analytical control solution was determined.

Keywords: artificial neural networks, nonlinear model control, process identification, crystallization process

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1824
10456 GEP Considering Purchase Prices, Profits of IPPs and Reliability Criteria Using Hybrid GA and PSO

Authors: H. Shayeghi, H. Hosseini, A. Shabani, M. Mahdavi

Abstract:

In this paper, optimal generation expansion planning (GEP) is investigated considering purchase prices, profits of independent power producers (IPPs) and reliability criteria using a new method based on hybrid coded Genetic Algorithm (GA) and Particle Swarm Optimization (PSO). In this approach, optimal purchase price of each IPP is obtained by HCGA and reliability criteria are calculated by PSO technique. It should be noted that reliability criteria and the rate of carbon dioxide (CO2) emission have been considered as constraints of the GEP problem. Finally, the proposed method has been tested on the case study system. The results evaluation show that the proposed method can simply obtain optimal purchase prices of IPPs and is a fast method for calculation of reliability criteria in expansion planning. Also, considering the optimal purchase prices and profits of IPPs in generation expansion planning are caused that the expansion costs are decreased and the problem is solved more exactly.

Keywords: GEP Problem, IPPs, Reliability Criteria, GA, PSO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1420
10455 Cooperative Data Caching in WSN

Authors: Narottam Chand

Abstract:

Wireless sensor networks (WSNs) have gained tremendous attention in recent years due to their numerous applications. Due to the limited energy resource, energy efficient operation of sensor nodes is a key issue in wireless sensor networks. Cooperative caching which ensures sharing of data among various nodes reduces the number of communications over the wireless channels and thus enhances the overall lifetime of a wireless sensor network. In this paper, we propose a cooperative caching scheme called ZCS (Zone Cooperation at Sensors) for wireless sensor networks. In ZCS scheme, one-hop neighbors of a sensor node form a cooperative cache zone and share the cached data with each other. Simulation experiments show that the ZCS caching scheme achieves significant improvements in byte hit ratio and average query latency in comparison with other caching strategies.

Keywords: Admission control, cache replacement, cooperative caching, WSN, zone cooperation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2746
10454 The Importance of Raising Awareness of Collocational Knowledge in ESL/EFL Classrooms

Authors: Mohammad ALAmro

Abstract:

The most crucial aspect that is closely related to vocabulary and the one that needs to be emphasized and investigated more than it has been up until now, is the ability to combine words that co-occur frequently in the language. Pedagogically, collocation is one of the error-provoking aspects in foreign language learning. This is indicative of the dire need to provide L2 learners with tools to help them improve their collocational knowledge. This paper pinpoints the role that collocations play in the English language. Furthermore, it presents pedagogical implications for ESL/EFL learners.

Keywords: Collocation, pedagogy, vocabulary knowledge.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2199
10453 Computer-Assisted Management of Building Climate and Microgrid with Model Predictive Control

Authors: Vinko Lešić, Mario Vašak, Anita Martinčević, Marko Gulin, Antonio Starčić, Hrvoje Novak

Abstract:

With 40% of total world energy consumption, building systems are developing into technically complex large energy consumers suitable for application of sophisticated power management approaches to largely increase the energy efficiency and even make them active energy market participants. Centralized control system of building heating and cooling managed by economically-optimal model predictive control shows promising results with estimated 30% of energy efficiency increase. The research is focused on implementation of such a method on a case study performed on two floors of our faculty building with corresponding sensors wireless data acquisition, remote heating/cooling units and central climate controller. Building walls are mathematically modeled with corresponding material types, surface shapes and sizes. Models are then exploited to predict thermal characteristics and changes in different building zones. Exterior influences such as environmental conditions and weather forecast, people behavior and comfort demands are all taken into account for deriving price-optimal climate control. Finally, a DC microgrid with photovoltaics, wind turbine, supercapacitor, batteries and fuel cell stacks is added to make the building a unit capable of active participation in a price-varying energy market. Computational burden of applying model predictive control on such a complex system is relaxed through a hierarchical decomposition of the microgrid and climate control, where the former is designed as higher hierarchical level with pre-calculated price-optimal power flows control, and latter is designed as lower level control responsible to ensure thermal comfort and exploit the optimal supply conditions enabled by microgrid energy flows management. Such an approach is expected to enable the inclusion of more complex building subsystems into consideration in order to further increase the energy efficiency.

Keywords: Energy-efficient buildings, Hierarchical model predictive control, Microgrid power flow optimization, Price-optimal building climate control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1511
10452 Analysis of a Coupled Hydro-Sedimentological Numerical Model for the Tombolo of GIENS

Authors: Yves Lacroix, Van Van Than, Didier Leandri, Pierre Liardet

Abstract:

The western Tombolo of the Giens peninsula in southern France, known as Almanarre beach, is subject to coastal erosion. We are trying to use computer simulation in order to propose solutions to stop this erosion. Our aim was first to determine the main factors for this erosion and successfully apply a coupled hydrosedimentological numerical model based on observations and measurements that have been performed on the site for decades. We have gathered all available information and data about waves, winds, currents, tides, bathymetry, coastal line, and sediments concerning the site. These have been divided into two sets: one devoted to calibrating a numerical model using Mike 21 software, the other to serve as a reference in order to numerically compare the present situation to what it could be if we implemented different types of underwater constructions. This paper presents the first part of the study: selecting and melting different sources into a coherent data basis, identifying the main erosion factors, and calibrating the coupled software model against the selected reference period. Our results bring calibration of the numerical model with good fitting coefficients. They also show that the winter South-Western storm events conjugated to depressive weather conditions constitute a major factor of erosion, mainly due to wave impact in the northern part of the Almanarre beach. Together, current and wind impact is shown negligible.

Keywords: Almanarre beach, coastal erosion, hydro-sedimentological, numerical model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2034
10451 An Improvement of Flow Forming Process for Pressure Vessels by Four Rollers Machine

Authors: P. Sawitri, S. Cdr. Sittha, T. Kritsana

Abstract:

Flow forming is widely used in many industries, especially in defence technology industries. Pressure vessels requirements are high precision, light weight, seamless and optimum strength. For large pressure vessels, flow forming by 3 rollers machine were used. In case of long range rocket motor case flow forming and welding of pressure vessels have been used for manufacturing. Due to complication of welding process, researchers had developed 4 meters length pressure vessels without weldment by 4 rollers flow forming machine. Design and preparation of preform work pieces are performed. The optimization of flow forming parameter such as feed rate, spindle speed and depth of cut will be discussed. The experimental result shown relation of flow forming parameters to quality of flow formed tube and prototype pressure vessels have been made.

Keywords: Flow forming, Pressure vessel, four rollers, feed rate, spindle speed, cold work.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2773
10450 A CFD Analysis of Flow through a High-Pressure Natural Gas Pipeline with an Undeformed and Deformed Orifice Plate

Authors: R. Kiš, M. Malcho, M. Janovcová

Abstract:

This work aims to present a numerical analysis of the natural gas which flows through a high-pressure pipeline and an orifice plate, through the use of CFD methods. The paper contains CFD calculations for the flow of natural gas in a pipe with different geometry used for the orifice plates. One of them has a standard geometry and a shape without any deformation and the other is deformed by the action of the pressure differential. It shows the behavior of natural gas in a pipeline using the velocity profiles and pressure fields of the gas in both models with their differences. The entire research is based on the elimination of any inaccuracy which should appear in the flow of the natural gas measured in the high-pressure pipelines of the gas industry and which is currently not given in the relevant standard.

Keywords: Orifice plate, high-pressure pipeline, natural gas, CFD analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3883
10449 Space-Time Variation in Rainfall and Runoff: Upper Betwa Catchment

Authors: Ritu Ahlawat

Abstract:

Among all geo-hydrological relationships, rainfallrunoff relationship is of utmost importance in any hydrological investigation and water resource planning. Spatial variation, lag time involved in obtaining areal estimates for the basin as a whole can affect the parameterization in design stage as well as in planning stage. In conventional hydrological processing of data, spatial aspect is either ignored or interpolated at sub-basin level. Temporal variation when analysed for different stages can provide clues for its spatial effectiveness. The interplay of space-time variation at pixel level can provide better understanding of basin parameters. Sustenance of design structures for different return periods and their spatial auto-correlations should be studied at different geographical scales for better management and planning of water resources. In order to understand the relative effect of spatio-temporal variation in hydrological data network, a detailed geo-hydrological analysis of Betwa river catchment falling in Lower Yamuna Basin is presented in this paper. Moreover, the exact estimates about the availability of water in the Betwa river catchment, especially in the wake of recent Betwa-Ken linkage project, need thorough scientific investigation for better planning. Therefore, an attempt in this direction is made here to analyse the existing hydrological and meteorological data with the help of SPSS, GIS and MS-EXCEL software. A comparison of spatial and temporal correlations at subcatchment level in case of upper Betwa reaches has been made to demonstrate the representativeness of rain gauges. First, flows at different locations are used to derive correlation and regression coefficients. Then, long-term normal water yield estimates based on pixel-wise regression coefficients of rainfall-runoff relationship have been mapped. The areal values obtained from these maps can definitely improve upon estimates based on point-based extrapolations or areal interpolations.

Keywords: Catchment's runoff estimates, influence area regional regression coefficients, runoff yield series,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2090
10448 Adopted Method of Information System Strategy for Knowledge Management System: A Literature Review

Authors: Elin Cahyaningsih, Dana Indra Sensuse, Wahyu Catur Wibowo, Sofiyanti Indriasari

Abstract:

Bureaucracy reform program drives Indonesian government to change their management to enhance their organizational performance. Information technology became one of strategic plan that organization tried to improve. Knowledge management system is one of information system that supporting knowledge management implementation in government which categorized as people perspective, because this system has high dependency in human interaction and participation. Strategic plan for developing knowledge management system can be determine using some of information system strategic methods. This research conducted to define type of strategic method of information system, stage of activity each method, strength and weakness. Literature review methods used to identify and classify strategic methods of information system, differentiate method type, categorize common activities, strength and weakness. Result of this research are determine and compare six strategic information system methods, Balanced Scorecard and Risk Analysis believe as common strategic method that usually used and have the highest excellence strength.

Keywords: Knowledge management system, balanced scorecard, five force, risk analysis, gap analysis, value chain analysis, SWOT analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2634
10447 Hybrid Artificial Immune System for Job Shop Scheduling Problem

Authors: Bin Cai, Shilong Wang, Haibo Hu

Abstract:

The job shop scheduling problem (JSSP) is a notoriously difficult problem in combinatorial optimization. This paper presents a hybrid artificial immune system for the JSSP with the objective of minimizing makespan. The proposed approach combines the artificial immune system, which has a powerful global exploration capability, with the local search method, which can exploit the optimal antibody. The antibody coding scheme is based on the operation based representation. The decoding procedure limits the search space to the set of full active schedules. In each generation, a local search heuristic based on the neighborhood structure proposed by Nowicki and Smutnicki is applied to improve the solutions. The approach is tested on 43 benchmark problems taken from the literature and compared with other approaches. The computation results validate the effectiveness of the proposed algorithm.

Keywords: Artificial immune system, Job shop scheduling problem, Local search, Metaheuristic algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1914
10446 The Automated Selective Acquisition System

Authors: Atisthan Wuttimanop, Suchada Rianmora

Abstract:

To support design process for launching the product on time, reverse engineering (RE) process has been introduced for quickly generating 3D CAD model from its physical object. The accuracy of the 3D CAD model depends upon the data acquisition technique selected, contact or non-contact methods. In order to reduce times used for acquiring surface and eliminating noises, the automated selective acquisition system has been developed and presented in this research as the alternative channel for non-contact acquisition technique where the data is selectively and locally scanned contour by contour without performing data reduction process. The results present as the organized contour points which are directly used to generate 3D virtual model. The comparison between the proposed technique and another non-contact scanning technique has been presented and discussed.

Keywords: Automated selective acquisition system, Non-contact acquisition, Reverse engineering, 3D scanners.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1595
10445 Node Insertion in Coalescence Hidden-Variable Fractal Interpolation Surface

Authors: Srijanani Anurag Prasad

Abstract:

The Coalescence Hidden-variable Fractal Interpolation Surface (CHFIS) was built by combining interpolation data from the Iterated Function System (IFS). The interpolation data in a CHFIS comprise a row and/or column of uncertain values when a single point is entered. Alternatively, a row and/or column of additional points are placed in the given interpolation data to demonstrate the node added CHFIS. There are three techniques for inserting new points that correspond to the row and/or column of nodes inserted, and each method is further classified into four types based on the values of the inserted nodes. As a result, numerous forms of node insertion can be found in a CHFIS.

Keywords: Fractal, interpolation, iterated function system, coalescence, node insertion, knot insertion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 313
10444 Transcriptional Evidence for the Involvement of MyD88 in Flagellin Recognition: Genomic Identification of Rock Bream MyD88 and Comparative Analysis

Authors: N. Umasuthan, S. D. N. K. Bathige, W. S. Thulasitha, I. Whang, J. Lee

Abstract:

The MyD88 is an evolutionarily conserved host-expressed adaptor protein that is essential for proper TLR/ IL1R immune-response signaling. A previously identified complete cDNA (1626 bp) of OfMyD88 comprised an ORF of 867 bp encoding a protein of 288 amino acids (32.9 kDa). The gDNA (3761 bp) of OfMyD88 revealed a quinquepartite genome organization composed of 5 exons (with the sizes of 310, 132, 178, 92 and 155 bp) separated by 4 introns. All the introns displayed splice signals consistent with the consensus GT/AG rule. A bipartite domain structure with two domains namely death domain (24-103) coded by 1st exon, and TIR domain (151-288) coded by last 3 exons were identified through in silico analysis. Moreover, homology modeling of these two domains revealed a similar quaternary folding nature between human and rock bream homologs. A comprehensive comparison of vertebrate MyD88 genes showed that they possess a 5-exonic structure.In this structure, the last three exons were strongly conserved, and this suggests that a rigid structure has been maintained during vertebrate evolution.A cluster of TATA box-like sequences were found 0.25 kb upstream of cDNA starting position. In addition, putative 5'-flanking region of OfMyD88 was predicted to have TFBS implicated with TLR signaling, including copies of NFkB1, APRF/ STAT3, Sp1, IRF1 and 2 and Stat1/2. Using qPCR technique, a ubiquitous mRNA expression was detected in liver and blood. Furthermore, a significantly up-regulated transcriptional expression of OfMyD88 was detected in head kidney (12-24 h; >2-fold), spleen (6 h; 1.5-fold), liver (3 h; 1.9-fold) and intestine (24 h; ~2-fold) post-Fla challenge. These data suggest a crucial role for MyD88 in antibacterial immunity of teleosts.

Keywords: MyD88, Innate immunity, Flagellin, Genomic analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897
10443 Using PFA in Feature Analysis and Selection for H.264 Adaptation

Authors: Nora A. Naguib, Ahmed E. Hussein, Hesham A. Keshk, Mohamed I. El-Adawy

Abstract:

Classification of video sequences based on their contents is a vital process for adaptation techniques. It helps decide which adaptation technique best fits the resource reduction requested by the client. In this paper we used the principal feature analysis algorithm to select a reduced subset of video features. The main idea is to select only one feature from each class based on the similarities between the features within that class. Our results showed that using this feature reduction technique the source video features can be completely omitted from future classification of video sequences.

Keywords: Adaptation, feature selection, H.264, Principal Feature Analysis (PFA)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1599
10442 Study on Guangzhou's Employment Subcentres and Polycentricity

Authors: L. Jiang

Abstract:

Since the late 1980s, the new phenomena of 'employment subcentres' or 'polycentricity' has appeared in the metropolises of North American and Western Europe and it has been an interesting topic for academics and researchers. This paper specifically uses one case study-Guangzhou to explore the development and the mechanism of employment subcentres and polycentricity in Chinese metropolises by spatial analysis method on the basis of the first economic census data. In conclusion, the paper regards that the employment subcentres and polycentricity has existed in Chinese metropolises. And that, the mechanism of them is mainly from the secondary industry instead of the tertiary industry in North American and Western Europe

Keywords: Employment Subcentre, Polycentricity, Guangzhou.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1825
10441 Optimizing Usability Testing with Collaborative Method in an E-Commerce Ecosystem

Authors: Markandeya Kunchi

Abstract:

Usability testing (UT) is one of the vital steps in the User-centred design (UCD) process when designing a product. In an e-commerce ecosystem, UT becomes primary as new products, features, and services are launched very frequently. And, there are losses attached to the company if an unusable and inefficient product is put out to market and is rejected by customers. This paper tries to answer why UT is important in the product life-cycle of an E-commerce ecosystem. Secondary user research was conducted to find out work patterns, development methods, type of stakeholders, and technology constraints, etc. of a typical E-commerce company. Qualitative user interviews were conducted with product managers and designers to find out the structure, project planning, product management method and role of the design team in a mid-level company. The paper tries to address the usual apprehensions of the company to inculcate UT within the team. As well, it stresses upon factors like monetary resources, lack of usability expert, narrow timelines, and lack of understanding of higher management as some primary reasons. Outsourcing UT to vendors is also very prevalent with mid-level e-commerce companies, but it has its own severe repercussions like very little team involvement, huge cost, misinterpretation of the findings, elongated timelines, and lack of empathy towards the customer, etc. The shortfalls of the unavailability of a UT process in place within the team and conducting UT through vendors are bad user experiences for customers while interacting with the product, badly designed products which are neither useful and nor utilitarian. As a result, companies see dipping conversions rates in apps and websites, huge bounce rates and increased uninstall rates. Thus, there was a need for a more lean UT system in place which could solve all these issues for the company. This paper highlights on optimizing the UT process with a collaborative method. The degree of optimization and structure of collaborative method is the highlight of this paper. Collaborative method of UT is one in which the centralised design team of the company takes for conducting and analysing the UT. The UT is usually a formative kind where designers take findings into account and uses in the ideation process. The success of collaborative method of UT is due to its ability to sync with the product management method employed by the company or team. The collaborative methods focus on engaging various teams (design, marketing, product, administration, IT, etc.) each with its own defined roles and responsibility in conducting a smooth UT with users In-house. The paper finally highlights the positive results of collaborative UT method after conducting more than 100 In-lab interviews with users across the different lines of businesses. Some of which are the improvement of interaction between stakeholders and the design team, empathy towards users, improved design iteration, better sanity check of design solutions, optimization of time and money, effective and efficient design solution. The future scope of collaborative UT is to make this method leaner, by reducing the number of days to complete the entire project starting from planning between teams to publishing the UT report.

Keywords: Usability testing, collaborative method, e-commerce, product management method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 656
10440 Overview of CARDIOSENSOR Project on the Development of a Nanosensor for Assessing the Risk of Cardiovascular Disease

Authors: A.C. Duarte, C.I.L. Justino, K. Duarte, A.C. Freitas, R. Pereira, P. Chaves, P. Bettencourt, S. Cardoso, T.A.P. Rocha-Santos

Abstract:

This paper aims at overviewing the topics of a research project (CARDIOSENSOR) on the field of health sciences (biomaterials and biomedical engineering). The project has focused on the development of a nanosensor for the assessment of the risk of cardiovascular diseases by the monitoring of C-reactive protein (CRP), which has been currently considered as the best validated inflammatory biomarker associated to cardiovascular diseases. The project involves tasks such as: 1) the development of sensor devices based on field effect transistors (FET): assembly, optimization and validation; 2) application of sensors to the detection of CRP in standard solutions and comparison with enzyme-linked immunosorbent assay (ELISA); and 3) application of sensors to real samples such as blood and saliva and evaluation of their ability to predict the risk of cardiovascular disease.

Keywords: Carbon nanotubes field effect transistors, cardiovascular diseases, C-reactive protein, sensor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2028
10439 Fighter Aircraft Selection Using Technique for Order Preference by Similarity to Ideal Solution with Multiple Criteria Decision Making Analysis

Authors: C. Ardil

Abstract:

This paper presents a multiple criteria decision making analysis technique for selecting fighter aircraft for the national air force. The selection of military aircraft is a process consisting of contradictory goals and objectives. When a modern air force needs to choose fighter aircraft to upgrade existing fleets, a multiple criteria decision making analysis and scenario planning for defense acquisition has been put forward. The selection of fighter aircraft for the air defense force is a strategic decision making process, since the purchase or lease of fighter jets, maintenance and operating costs and having a fleet is the biggest cost for the air force. Multiple criteria decision making analysis methods are effectively applied to facilitate decision making from various available options. The selection criteria were determined using the literature on the problem of fighter aircraft selection. The selection of fighter aircraft to be purchased for the air defense forces is handled using a multiple criteria decision making analysis technique that also determines a suitable methodological approach for the defense procurement and fleet upgrade planning process. The aim of this study is to originate an approach to evaluate fighter aircraft alternatives, Su-35, F-35, and TF-X (MMU), based on technique for order preference by similarity to ideal solution (TOPSIS).

Keywords: Fighter Aircraft, Fighter Aircraft Selection, Technique for Order Preference by Similarity to Ideal Solution, TOPSIS, Multiple Criteria Decision Making, Multiple Criteria Decision Making Analysis, MCDMA, Su-35, F-35, TF-X (MMU)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 609
10438 UEFA Super Cup: Economic Effects on Georgian Economy

Authors: Giorgi Bregadze

Abstract:

Tourism is the most viable and sustainable economic development option for Georgia and one of the main sources of foreign exchange earnings. Events are considered as one of the most effective ways to attract foreign visitors to the country, and, recently, the government of Georgia has begun investing in this sector very actively. This article stresses the necessity of research based economic policy in the tourism sector. In this regard, it is of paramount importance to measure the economic effects of the events which are subsidized by taxpayers’ money. The economic effect of events can be analyzed from two perspectives; financial perspective of the government and perspective of economic effects of the tourism administration. The article emphasizes more realistic and all-inclusive focus of the economic effect analysis of the tourism administration as it concentrates on the income of residents and local businesses, part of which generate tax revenues for the government. The public would like to know what the economic returns to investment are. In this article, the methodology used to describe the economic effects of UEFA Super Cup held in Tbilisi, will help to answer this question. Methodology is based on three main principles and covers three stages. Using the suggested methodology article estimates the direct economic effect of UEFA Super cup on Georgian economy. Although the attempt to make an economic effect analysis of the event was successful in Georgia, some obstacles and insufficiencies were identified during the survey. The article offers several recommendations that will help to refine methodology and improve the accuracy of the data. Furthermore, it is very important to receive the correct standard of measurement of events in Georgia. In this caseü non-ethical acts of measurement which are widely utilized by different research companies will not trigger others to show overestimated effects. It is worth mentioning that to author’s best knowledge, this is the first attempt to measure the economic effect of an event held in Georgia.

Keywords: Biased economic effect analysis, expenditure of local citizens, time switchers and casuals, UEFA super cup.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 688
10437 Radioactivity of the Agricultural Soil in Northern Province of Serbia, Vojvodina

Authors: I. Bikit, S. Forkapic, J. Nikolov, N. Todorovic, D. Mrdja

Abstract:

During the year 1999, Serbia (ex Yugoslavia) and their northern province, Vojvodina, has been bombarded. Because of that general public believe is that this region was contaminated by depleted uranium and that there is a potential contaminant of agricultural products due to soil radioactivity. This paper presents the repeated analysis of agricultural soil samples in Vojvodina. The same investigation was carried out during the year 2001, and it was concluded that, based on the gamma-spectrometric analysis of 50 soil samples taken from the region of Vojvodina, there haven-t been registered any increase of radioactivity that could endanger the food production. We continue with the monitoring of this region. The comparison between those two sets of results is presented.

Keywords: gamma spectrometry analysis, radioactivity of theagricultural soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1766
10436 Motion Control of a Ball Throwing Robot with a Flexible Robotic Arm

Authors: Yizhi Gai, Yukinori Kobayashi, Yohei Hoshino, Takanori Emaru

Abstract:

Motion control of flexible arms is more difficult than that of rigid arms, however utilizing its dynamics enables improved performance such as a fast motion in short operation time. This paper investigates a ball throwing robot with one rigid link and one flexible link. This robot throws a ball at a set speed with a proper control torque. A mathematical model of this ball throwing robot is derived through Hamilton’s principle. Several patterns of torque input are designed and tested through the proposed simulation models. The parameters of each torque input pattern is optimized and determined by chaos embedded vector evaluated particle swarm optimization (CEVEPSO). Then, the residual vibration of the manipulator after throwing is suppressed with input shaping technique. Finally, a real experiment is set up for the model checking.

Keywords: Motion control, flexible robotic arm, CEVEPSO, ball throwing robot.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4052
10435 Producing Outdoor Design Conditions Based on the Dependency between Meteorological Elements: Copula Approach

Authors: Zhichao Jiao, Craig Farnham, Jihui Yuan, Kazuo Emura

Abstract:

It is common to use the outdoor design weather data to select the air-conditioning capacity in the building design stage. The meteorological elements of outdoor design weather data are usually selected based on their excess frequency separately while the dependency between the elements is not well considered. It means that the simultaneous occurrence probability of these elements is smaller than the original excess frequency which may cause an overestimation of selecting air-conditioning capacity. Therefore, the copula approach which can capture the dependency between multivariate data was used to model the joint distributions of the meteorological elements, like air temperature and global solar radiation. We suggest a method based on the specific simultaneous occurrence probability of these two elements of selecting more credible outdoor design conditions. The hourly weather data at 12 noon from 2001 to 2010 in Tokyo, Japan are used to analyze the dependency structure and joint distribution, the Gaussian copula represents the dependence of data best. According to calculating the air temperature and global solar radiation in specific simultaneous occurrence probability and the common exceeding, the results show that both the air temperature and global solar radiation based on simultaneous occurrence probability are lower than these based on the conventional method in the same probability.

Keywords: Copula approach, Design weather database, energy conservation, HVAC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 332
10434 Evaluation of the Impact of Dataset Characteristics for Classification Problems in Biological Applications

Authors: Kanthida Kusonmano, Michael Netzer, Bernhard Pfeifer, Christian Baumgartner, Klaus R. Liedl, Armin Graber

Abstract:

Availability of high dimensional biological datasets such as from gene expression, proteomic, and metabolic experiments can be leveraged for the diagnosis and prognosis of diseases. Many classification methods in this area have been studied to predict disease states and separate between predefined classes such as patients with a special disease versus healthy controls. However, most of the existing research only focuses on a specific dataset. There is a lack of generic comparison between classifiers, which might provide a guideline for biologists or bioinformaticians to select the proper algorithm for new datasets. In this study, we compare the performance of popular classifiers, which are Support Vector Machine (SVM), Logistic Regression, k-Nearest Neighbor (k-NN), Naive Bayes, Decision Tree, and Random Forest based on mock datasets. We mimic common biological scenarios simulating various proportions of real discriminating biomarkers and different effect sizes thereof. The result shows that SVM performs quite stable and reaches a higher AUC compared to other methods. This may be explained due to the ability of SVM to minimize the probability of error. Moreover, Decision Tree with its good applicability for diagnosis and prognosis shows good performance in our experimental setup. Logistic Regression and Random Forest, however, strongly depend on the ratio of discriminators and perform better when having a higher number of discriminators.

Keywords: Classification, High dimensional data, Machine learning

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2371
10433 Finding Fuzzy Association Rules Using FWFP-Growth with Linguistic Supports and Confidences

Authors: Chien-Hua Wang, Chin-Tzong Pang

Abstract:

In data mining, the association rules are used to search for the relations of items of the transactions database. Following the data is collected and stored, it can find rules of value through association rules, and assist manager to proceed marketing strategy and plan market framework. In this paper, we attempt fuzzy partition methods and decide membership function of quantitative values of each transaction item. Also, by managers we can reflect the importance of items as linguistic terms, which are transformed as fuzzy sets of weights. Next, fuzzy weighted frequent pattern growth (FWFP-Growth) is used to complete the process of data mining. The method above is expected to improve Apriori algorithm for its better efficiency of the whole association rules. An example is given to clearly illustrate the proposed approach.

Keywords: Association Rule, Fuzzy Partition Methods, FWFP-Growth, Apiroir algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1641
10432 Development of a RAM Simulation Model for Acid Gas Removal System

Authors: Ainul Akmar Mokhtar, Masdi Muhammad, Hilmi Hussin, Mohd Amin Abdul Majid

Abstract:

A reliability, availability and maintainability (RAM) model has been built for acid gas removal plant for system analysis that will play an important role in any process modifications, if required, for achieving its optimum performance. Due to the complexity of the plant, the model was based on a Reliability Block Diagram (RBD) with a Monte Carlo simulation engine. The model has been validated against actual plant data as well as local expert opinions, resulting in an acceptable simulation model. The results from the model showed that the operation and maintenance can be further improved, resulting in reduction of the annual production loss.

Keywords: Acid gas removal plant, RAM model, Reliabilityblock diagram

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2324
10431 Fault Detection and Diagnosis of Broken Bar Problem in Induction Motors Base Wavelet Analysis and EMD Method: Case Study of Mobarakeh Steel Company in Iran

Authors: M. Ahmadi, M. Kafil, H. Ebrahimi

Abstract:

Nowadays, induction motors have a significant role in industries. Condition monitoring (CM) of this equipment has gained a remarkable importance during recent years due to huge production losses, substantial imposed costs and increases in vulnerability, risk, and uncertainty levels. Motor current signature analysis (MCSA) is one of the most important techniques in CM. This method can be used for rotor broken bars detection. Signal processing methods such as Fast Fourier transformation (FFT), Wavelet transformation and Empirical Mode Decomposition (EMD) are used for analyzing MCSA output data. In this study, these signal processing methods are used for broken bar problem detection of Mobarakeh steel company induction motors. Based on wavelet transformation method, an index for fault detection, CF, is introduced which is the variation of maximum to the mean of wavelet transformation coefficients. We find that, in the broken bar condition, the amount of CF factor is greater than the healthy condition. Based on EMD method, the energy of intrinsic mode functions (IMF) is calculated and finds that when motor bars become broken the energy of IMFs increases.

Keywords: Broken bar, condition monitoring, diagnostics, empirical mode decomposition, Fourier transform, wavelet transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 784