Search results for: Expert evaluations
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 416

Search results for: Expert evaluations

86 CT Medical Images Denoising Based on New Wavelet Thresholding Compared with Curvelet and Contourlet

Authors: Amir Moslemi, Amir Movafeghi, Shahab Moradi

Abstract:

One of the most important challenging factors in medical images is nominated as noise. Image denoising refers to the improvement of a digital medical image that has been infected by Additive White Gaussian Noise (AWGN). The digital medical image or video can be affected by different types of noises. They are impulse noise, Poisson noise and AWGN. Computed tomography (CT) images are subjects to low quality due to the noise. Quality of CT images is dependent on absorbed dose to patients directly in such a way that increase in absorbed radiation, consequently absorbed dose to patients (ADP), enhances the CT images quality. In this manner, noise reduction techniques on purpose of images quality enhancement exposing no excess radiation to patients is one the challenging problems for CT images processing. In this work, noise reduction in CT images was performed using two different directional 2 dimensional (2D) transformations; i.e., Curvelet and Contourlet and Discrete Wavelet Transform (DWT) thresholding methods of BayesShrink and AdaptShrink, compared to each other and we proposed a new threshold in wavelet domain for not only noise reduction but also edge retaining, consequently the proposed method retains the modified coefficients significantly that result good visual quality. Data evaluations were accomplished by using two criterions; namely, peak signal to noise ratio (PSNR) and Structure similarity (Ssim).

Keywords: Computed Tomography (CT), noise reduction, curve-let, contour-let, Signal to Noise Peak-Peak Ratio (PSNR), Structure Similarity (Ssim), Absorbed Dose to Patient (ADP).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2920
85 Basic Research on Applying Temporary Work Engineering at the Design Phase

Authors: Jin Woong Lee, Kyuman Cho, Taehoon Kim

Abstract:

The application of constructability is increasingly required not only in the construction phase but also in the whole project stage. In particular, the proper application of construction experience and knowledge during the design phase enables the minimization of inefficiencies such as design changes and improvements in constructability during the construction phase. In order to apply knowledge effectively, engineering technology efforts should be implemented with design progress. Among many engineering technologies, engineering for temporary works, including facilities, equipment, and other related construction methods, is important to improve constructability. Therefore, as basic research, this study investigates the applicability of temporary work engineering during the design phase in the building construction industry. As a result, application of temporary work engineering has a greater impact on construction cost reduction and constructability improvement. In contrast to the existing design-bid-build method, the turn-key and CM (construct management) procurement methods currently being implemented in Korea are expected to have a significant impact on the direction of temporary work engineering. To introduce temporary work engineering, expert/professional organization training is first required, and a lack of client awareness should be preferentially improved. The results of this study are expected to be useful as reference material for the development of more effective temporary work engineering tasks and work processes in the future.

Keywords: Temporary work engineering, design phase, constructability, building construction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 972
84 Implementation of Geo-knowledge Based Geographic Information System for Estimating Earthquake Hazard Potential at a Metropolitan Area, Gwangju, in Korea

Authors: Chang-Guk Sun, Jin-Soo Shin

Abstract:

In this study, an inland metropolitan area, Gwangju, in Korea was selected to assess the amplification potential of earthquake motion and provide the information for regional seismic countermeasure. A geographic information system-based expert system was implemented for reliably predicting the spatial geotechnical layers in the entire region of interesting by building a geo-knowledge database. Particularly, the database consists of the existing boring data gathered from the prior geotechnical projects and the surface geo-knowledge data acquired from the site visit. For practical application of the geo-knowledge database to estimate the earthquake hazard potential related to site amplification effects at the study area, seismic zoning maps on geotechnical parameters, such as the bedrock depth and the site period, were created within GIS framework. In addition, seismic zonation of site classification was also performed to determine the site amplification coefficients for seismic design at any site in the study area. KeywordsEarthquake hazard, geo-knowledge, geographic information system, seismic zonation, site period.

Keywords: Earthquake hazard, geo-knowledge, geographic information system, seismic zonation, site period.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1664
83 M2LGP: Mining Multiple Level Gradual Patterns

Authors: Yogi Satrya Aryadinata, Anne Laurent, Michel Sala

Abstract:

Gradual patterns have been studied for many years as they contain precious information. They have been integrated in many expert systems and rule-based systems, for instance to reason on knowledge such as “the greater the number of turns, the greater the number of car crashes”. In many cases, this knowledge has been considered as a rule “the greater the number of turns → the greater the number of car crashes” Historically, works have thus been focused on the representation of such rules, studying how implication could be defined, especially fuzzy implication. These rules were defined by experts who were in charge to describe the systems they were working on in order to turn them to operate automatically. More recently, approaches have been proposed in order to mine databases for automatically discovering such knowledge. Several approaches have been studied, the main scientific topics being: how to determine what is an relevant gradual pattern, and how to discover them as efficiently as possible (in terms of both memory and CPU usage). However, in some cases, end-users are not interested in raw level knowledge, and are rather interested in trends. Moreover, it may be the case that no relevant pattern can be discovered at a low level of granularity (e.g. city), whereas some can be discovered at a higher level (e.g. county). In this paper, we thus extend gradual pattern approaches in order to consider multiple level gradual patterns. For this purpose, we consider two aggregation policies, namely horizontal and vertical.

Keywords: Gradual Pattern.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1500
82 Image Clustering Framework for BAVM Segmentation in 3DRA Images: Performance Analysis

Authors: FH. Sarieddeen, R. El Berbari, S. Imad, J. Abdel Baki, M. Hamad, R. Blanc, A. Nakib, Y.Chenoune

Abstract:

Brain ArterioVenous Malformation (BAVM) is an abnormal tangle of brain blood vessels where arteries shunt directly into veins with no intervening capillary bed which causes high pressure and hemorrhage risk. The success of treatment by embolization in interventional neuroradiology is highly dependent on the accuracy of the vessels visualization. In this paper the performance of clustering techniques on vessel segmentation from 3- D rotational angiography (3DRA) images is investigated and a new technique of segmentation is proposed. This method consists in: preprocessing step of image enhancement, then K-Means (KM), Fuzzy C-Means (FCM) and Expectation Maximization (EM) clustering are used to separate vessel pixels from background and artery pixels from vein pixels when possible. A post processing step of removing false-alarm components is applied before constructing a three-dimensional volume of the vessels. The proposed method was tested on six datasets along with a medical assessment of an expert. Obtained results showed encouraging segmentations.

Keywords: Brain arteriovenous malformation (BAVM), 3-D rotational angiography (3DRA), K-Means (KM) clustering, Fuzzy CMeans (FCM) clustering, Expectation Maximization (EM) clustering, volume rendering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1910
81 Investigating the Formation of Nano-Hydroxyapatite on a Biocompatible and Antibacterial Cu/Mg-Substituted Bioglass

Authors: Elhamalsadat Ghaffari, Moghan Amirhosseinian, Amir Khaleghipour

Abstract:

Multifunctional bioactive glasses (BGs) are designed with a focus on the provision of bactericidal and biological properties desired for angiogenesis, osteogenesis, and ultimately potential applications in bone tissue engineering. To achieve these, six sol-gel copper/magnesium substituted derivatives of 58S-BG, i.e. a mol% series of 60SiO2-4P2O5-5CuO-(31-x) CaO/xMgO (where x=0, 1, 3, 5, 8, and 10), were synthesized. Afterwards, the effect of MgO/CaO substitution on the in vitro formation of nano-hydroxyapatite (HA), osteoblast-like cell responses and BGs antibacterial performance were studied. During the BGs synthesis, the elimination of nitrates was achieved at 700 °C that prevented the BGs crystallization and stabilized the obtained dried gels. The structural and morphological evaluations were performed with X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FTIR), and scanning electron microscopy (SEM). These characterizations revealed that Cu-substituted 58S-BG consisting of 5 mol% MgO (BG-5/5) slightly had retarded the formation of HA. In addition, Cu-substituted 58S-BGs consisting 8 mol% and 10 mol% MgO (BG-5/8 and BG-5/10) displayed lower bioactivity probably due to the lower ion release rate of Ca–Si into the simulated body fluid (SBF). The determination of 3-(4, 5 dimethylthiazol-2-yl)-2, 5-diphenyltetrazolium bromide (MTT) and alkaline phosphate (ALP) activities proved that the highest values of both differentiation and proliferation of MC3T3-E1 cells can be obtained from a 5 mol% MgO substituted BG, while the over addition of MgO (8 mol% and 10 mol%) decreased the bioactivity. Furthermore, these novel Cu/Mg-substituted 58S-BGs displayed antibacterial effect against methicillin-resistant Staphylococcus aureus bacteria. Taken together, the results suggest the equally-substituted BG-5/5 (i.e. the one consists of 5 mol% of both CuO and MgO) as a promising candidate for bone tissue engineering, among all newly designed BGs in this work, owing to its desirable cell proliferation, ALP activity and antibacterial properties.

Keywords: Apatite, bioactivity, biomedical applications sol-gel processes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 832
80 Optimization of Conditions for Xanthan Gum Production from Waste Date in Submerged Fermantation

Authors: S. Moshaf, Z. Hamidi-Esfahani, M. H. Azizi

Abstract:

Xanthan gum is one of the major commercial biopolymers. Due to its excellent rheological properties xanthan gum is used in many applications, mainly in food industry. Commercial production of xanthan gum uses glucose as the carbon substrate; consequently the price of xanthan production is high. One of the ways to decrease xanthan price, is using cheaper substrate like agricultural wastes. Iran is one of the biggest date producer countries. However approximately 50% of date production is wasted annually. The goal of this study is to produce xanthan gum from waste date using Xanthomonas campestris PTCC1473 by submerged fermentation. In this study the effect of three variables including phosphor and nitrogen amount and agitation rate in three levels using response surface methodology (RSM) has been studied. Results achieved from statistical analysis Design Expert 7.0.0 software showed that xanthan increased with increasing level of phosphor. Low level of nitrogen leaded to higher xanthan production. Xanthan amount, increasing agitation had positive influence. The statistical model identified the optimum conditions nitrogen amount=3.15g/l, phosphor amount=5.03 g/l and agitation=394.8 rpm for xanthan. To model validation, experiments in optimum conditions for xanthan gum were carried out. The mean of result for xanthan was 6.72±0.26. The result was closed to the predicted value by using RSM.

Keywords: Optimization, RSM, Waste date, Xanthan gum, Xanthomonas Campestris

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2611
79 Mining Genes Relations in Microarray Data Combined with Ontology in Colon Cancer Automated Diagnosis System

Authors: A. Gruzdz, A. Ihnatowicz, J. Siddiqi, B. Akhgar

Abstract:

MATCH project [1] entitle the development of an automatic diagnosis system that aims to support treatment of colon cancer diseases by discovering mutations that occurs to tumour suppressor genes (TSGs) and contributes to the development of cancerous tumours. The constitution of the system is based on a) colon cancer clinical data and b) biological information that will be derived by data mining techniques from genomic and proteomic sources The core mining module will consist of the popular, well tested hybrid feature extraction methods, and new combined algorithms, designed especially for the project. Elements of rough sets, evolutionary computing, cluster analysis, self-organization maps and association rules will be used to discover the annotations between genes, and their influence on tumours [2]-[11]. The methods used to process the data have to address their high complexity, potential inconsistency and problems of dealing with the missing values. They must integrate all the useful information necessary to solve the expert's question. For this purpose, the system has to learn from data, or be able to interactively specify by a domain specialist, the part of the knowledge structure it needs to answer a given query. The program should also take into account the importance/rank of the particular parts of data it analyses, and adjusts the used algorithms accordingly.

Keywords: Bioinformatics, gene expression, ontology, selforganizingmaps.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1974
78 Automatic Segmentation of Lung Areas in Magnetic Resonance Images

Authors: Alireza Osareh, Bita Shadgar

Abstract:

Segmenting the lungs in medical images is a challenging and important task for many applications. In particular, automatic segmentation of lung cavities from multiple magnetic resonance (MR) images is very useful for oncological applications such as radiotherapy treatment planning. However, distinguishing of the lung areas is not trivial due to largely changing lung shapes, low contrast and poorly defined boundaries. In this paper, we address lung segmentation problem from pulmonary magnetic resonance images and propose an automated method based on a robust regionaided geometric snake with a modified diffused region force into the standard geometric model definition. The extra region force gives the snake a global complementary view of the lung boundary information within the image which along with the local gradient flow, helps detect fuzzy boundaries. The proposed method has been successful in segmenting the lungs in every slice of 30 magnetic resonance images with 80 consecutive slices in each image. We present results by comparing our automatic method to manually segmented lung cavities provided by an expert radiologist and with those of previous works, showing encouraging results and high robustness of our approach.

Keywords: Active contours, breast cancer, fuzzy c-means segmentation, treatment planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2057
77 Diagnosis of the Abdominal Aorta Aneurysm in Magnetic Resonance Imaging Images

Authors: W. Kultangwattana, K. Somkantha, P. Phuangsuwan

Abstract:

This paper presents a technique for diagnosis of the abdominal aorta aneurysm in magnetic resonance imaging (MRI) images. First, our technique is designed to segment the aorta image in MRI images. This is a required step to determine the volume of aorta image which is the important step for diagnosis of the abdominal aorta aneurysm. Our proposed technique can detect the volume of aorta in MRI images using a new external energy for snakes model. The new external energy for snakes model is calculated from Law-s texture. The new external energy can increase the capture range of snakes model efficiently more than the old external energy of snakes models. Second, our technique is designed to diagnose the abdominal aorta aneurysm by Bayesian classifier which is classification models based on statistical theory. The feature for data classification of abdominal aorta aneurysm was derived from the contour of aorta images which was a result from segmenting of our snakes model, i.e., area, perimeter and compactness. We also compare the proposed technique with the traditional snakes model. In our experiment results, 30 images are trained, 20 images are tested and compared with expert opinion. The experimental results show that our technique is able to provide more accurate results than 95%.

Keywords: Adbominal Aorta Aneurysm, Bayesian Classifier, Snakes Model, Texture Feature.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1592
76 Measuring the Influence of Functional Proximity on Environmental Urban Performance via Integrated Modification Methodology: Four Study Cases in Milan

Authors: M. Tadi, M. Hadi Mohammad Zadeh, Ozge Ogut

Abstract:

Although how cities’ forms are structured is studied, more efforts are needed on systemic comprehensions and evaluations of the urban morphology through quantitative metrics that are able to describe the performance of a city in relation to its formal properties. More research is required in this direction in order to better describe the urban form characteristics and their impact on the environmental performance of cities and to increase their sustainability stewardship. With the aim of developing a better understanding of the built environment’s systemic structure, the intention of this paper is to present a holistic methodology for studying the behavior of the built environment and investigate the methods for measuring the effect of urban structure to the environmental performance. This goal will be pursued through an inquiry into the morphological components of the urban systems and the complex relationships between them. Particularly, this paper focuses on proximity, referring to the proximity of different land-uses, is a concept with which Integrated Modification Methodology (IMM) explains how land-use allocation might affect the choice of mobility in neighborhoods, and especially, encourage or discourage non-motived mobility. This paper uses proximity to demonstrate that the structure attributes can quantifiably relate to the performing behavior in the city. The target is to devise a mathematical pattern from the structural elements and correlate it directly with urban performance indicators concerned with environmental sustainability. The paper presents some results of this rigorous investigation of urban proximity and its correlation with performance indicators in four different areas in the city of Milan, each of them characterized by different morphological features.

Keywords: Built environment, ecology, sustainable indicators, sustainability, urban morphology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 629
75 Prophylactic Effects of Dairy Kluyveromyces marxianus YAS through Overexpression of BAX, CASP 3, CASP 8 and CASP 9 on Human Colon Cancer Cell Lines

Authors: Amir Saber Gharamaleki, Beitollah Alipour, Zeinab Faghfoori, Ahmad YariKhosroushahi

Abstract:

Colorectal cancer (CRC) is one of the most prevalent cancers and intestinal microbial community plays an important role in colorectal tumorigenesis. Probiotics have recently been assessed as effective anti-proliferative agents and thus this study was performed to examine whether CRC undergo apoptosis by treating with isolated Iranian native dairy yeast, Kluyveromyces marxianus YAS, secretion metabolites. The cytotoxicity assessments on cells (HT-29, Caco-2) were accomplished through 3-(4,5-dimethylthiazol-2-yl)-2,5-diphenyltetrazolium bromide (MTT) assay as well as qualitative DAPI (4',6-diamidino-2-phenylindole staining) and quantitative (flow cytometry assessments) evaluations of apoptosis. To evaluate the main mechanism of apoptosis, Real time PCR method was applied. Kluyveromyces marxianus YAS secretions (IC50) showed significant cytotoxicity against HT-29 and Caco-2 cancer cell lines (66.57 % and 66.34 % apoptosis) similar to 5-Fluorouracil (5-FU) while apoptosis only was developed in 27.57 % of KDR normal cells. The prophylactic effects of Kluyveromyces marxianus (PTCC 5195), as a reference yeast, was not similar to Kluyveromyces marxianus YAS indicating strain dependency of bioactivities on CRC disease prevention. Based on real time PCR results, the main cytotoxicity is related to apoptosis phenomenon and the core related mechanism is depended on the overexpression of BAX, CASP 9, CASP 8 and CASP 3 inducing apoptosis genes. However, several investigations should be conducted to precisely determine the effective compounds to be used as anticancer therapeutics in the future.

Keywords: Anticancer, anti-proliferative, apoptosis, cytotoxicity, yeast.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661
74 Development of an Omaha System-Based Remote Intervention Program for Work-Related Musculoskeletal Disorders among Front-Line Nurses

Authors: Tianqiao Zhang, Ye Tian, Yanliang Yin, Yichao Tian, Suzhai Tian, Weige Sun, Ruoliang Tang

Abstract:

Heavy biomechanical loads at workplaces may lead to high risks of work-related musculoskeletal disorders (WMSDs). However, there is a lack of investigations on the efficacy of the ergonomic interventions with theoretical frameworks. This study aimed to formulate an Omaha System based remote intervention program on the WMSDs among nurses by systematic literature review, interviews, expert consultation. After screening title and abstract, 11 articles out of the initial search results (i.e., n=1,418) were included, 12 nurses were interviewed, and 10 experts were consulted to review the initial intervention program. Modification to the draft included (1) supplementing traditional Chinese medicine practices, (2) adding the use of assistive patient handling equipment, (3) revising the on-line training method, (4) editing and proofreading the main text of the initial program, (5) adding quizzes and exercise scales, (6) it was determined that the associated coursework should be announced promptly with multiple follow-up reminders, and (7) removing bodyweight superman exercise, and peaceful/calm meditation. In the end, the final intervention program was developed.

Keywords: Omaha System, nurses, remote intervention, musculoskeletal disease.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 273
73 Toward Discovering an Architectural Typology Based on the Theory of Affordance

Authors: Falntina Ahmad Alata, Natheer Abu Obeid

Abstract:

This paper revolves around the concept of affordance. It aims to discover and develop an architectural typology based on the ecological concept of affordance. In order to achieve this aim, an analytical study is conducted and two sources were taken into account: 1- Gibson's definition of the concept of affordance and 2- The researches that are concerned on the affordance categorisation. As a result, this paper concluded 16 typologies of affordances, including the possibilities of mixing them based on both sources. To clarify these typologies and provide further understanding, a wide range of architectural examples are presented and proposed in the paper. To prove this vocabulary’s capability to diagnose and evaluate the affordance of different environments, an experimental study with two processes have been adapted: 1. Diagnostic process: the interpretation of the environments with regards to its affordance by using the new vocabulary (the developed typologies). 2. Evaluating process: the evaluation of the environments that have been interpreted and classified with regards to their affordances. By using the measures of emotional experience (the positive affect ‘PA’ and the negative affect ‘NA’) and the architectural evaluation criteria (beauty, economy and function). The experimental study proves that the typologies are capable of reading the affordance within different environments. Additionally, it explains how these different typologies reflect different interactions based on the previous processes. The data which are concluded from the evaluation of measures explain how different typologies of affordance that have already reflected different environments had different evaluations. In fact, some of them are recommended while the others are not. In other words, the paper draws a roadmap for designers to diagnose, evaluate and analyse the affordance into different architectural environments. After that, it guides them through adapting the best interaction (affordance category), which they intend to adapt into their proposed designs.

Keywords: Affordance theory, affordance categories, architectural environments, architectural evaluation criteria, emotional experience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 616
72 Bayesian Belief Networks for Test Driven Development

Authors: Vijayalakshmy Periaswamy S., Kevin McDaid

Abstract:

Testing accounts for the major percentage of technical contribution in the software development process. Typically, it consumes more than 50 percent of the total cost of developing a piece of software. The selection of software tests is a very important activity within this process to ensure the software reliability requirements are met. Generally tests are run to achieve maximum coverage of the software code and very little attention is given to the achieved reliability of the software. Using an existing methodology, this paper describes how to use Bayesian Belief Networks (BBNs) to select unit tests based on their contribution to the reliability of the module under consideration. In particular the work examines how the approach can enhance test-first development by assessing the quality of test suites resulting from this development methodology and providing insight into additional tests that can significantly reduce the achieved reliability. In this way the method can produce an optimal selection of inputs and the order in which the tests are executed to maximize the software reliability. To illustrate this approach, a belief network is constructed for a modern software system incorporating the expert opinion, expressed through probabilities of the relative quality of the elements of the software, and the potential effectiveness of the software tests. The steps involved in constructing the Bayesian Network are explained as is a method to allow for the test suite resulting from test-driven development.

Keywords: Software testing, Test Driven Development, Bayesian Belief Networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1887
71 Value Analysis Dashboard in Supply Chain Management: Real Case Study from Iran

Authors: Seyedehfatemeh Golrizgashti, Seyedali Dalil

Abstract:

The goal of this paper is proposing a supply chain value dashboard in home appliance manufacturing firms to create more value for all stakeholders via balanced scorecard approach. Balanced scorecard is an effective approach that managers have used to evaluate supply chain performance in many fields but there is a lack of enough attention to all supply chain stakeholders, improving value creation and, defining correlation between value indicators and performance measuring quantitatively. In this research the key stakeholders in home appliance supply chain, value indicators with respect to create more value for stakeholders and the most important metrics to evaluate supply chain value performance based on balanced scorecard approach have been selected via literature review. The most important indicators based on expert’s judgment acquired by in survey focused on creating more value for. Structural equation modelling has been used to disclose relations between value indicators and balanced scorecard metrics. The important result of this research is identifying effective value dashboard to create more value for all stakeholders in supply chain via balanced scorecard approach and based on an empirical study covering ten home appliance manufacturing firms in Iran. Home appliance manufacturing firms can increase their stakeholder's satisfaction by using this value dashboard.

Keywords: Supply chain management, balanced scorecard, value, Structural modeling, Stakeholders.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2078
70 On The Analysis of a Compound Neural Network for Detecting Atrio Ventricular Heart Block (AVB) in an ECG Signal

Authors: Salama Meghriche, Amer Draa, Mohammed Boulemden

Abstract:

Heart failure is the most common reason of death nowadays, but if the medical help is given directly, the patient-s life may be saved in many cases. Numerous heart diseases can be detected by means of analyzing electrocardiograms (ECG). Artificial Neural Networks (ANN) are computer-based expert systems that have proved to be useful in pattern recognition tasks. ANN can be used in different phases of the decision-making process, from classification to diagnostic procedures. This work concentrates on a review followed by a novel method. The purpose of the review is to assess the evidence of healthcare benefits involving the application of artificial neural networks to the clinical functions of diagnosis, prognosis and survival analysis, in ECG signals. The developed method is based on a compound neural network (CNN), to classify ECGs as normal or carrying an AtrioVentricular heart Block (AVB). This method uses three different feed forward multilayer neural networks. A single output unit encodes the probability of AVB occurrences. A value between 0 and 0.1 is the desired output for a normal ECG; a value between 0.1 and 1 would infer an occurrence of an AVB. The results show that this compound network has a good performance in detecting AVBs, with a sensitivity of 90.7% and a specificity of 86.05%. The accuracy value is 87.9%.

Keywords: Artificial neural networks, Electrocardiogram(ECG), Feed forward multilayer neural network, Medical diagnosis, Pattern recognitionm, Signal processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2472
69 A Review on Factors Influencing Implementation of Secure Software Development Practices

Authors: Sri Lakshmi Kanniah, Mohd Naz’ri Mahrin

Abstract:

More and more businesses and services are depending on software to run their daily operations and business services. At the same time, cyber-attacks are becoming more covert and sophisticated, posing threats to software. Vulnerabilities exist in the software due to the lack of security practices during the phases of software development. Implementation of secure software development practices can improve the resistance to attacks. Many methods, models and standards for secure software development have been developed. However, despite the efforts, they still come up against difficulties in their deployment and the processes are not institutionalized. There is a set of factors that influence the successful deployment of secure software development processes. In this study, the methodology and results from a systematic literature review of factors influencing the implementation of secure software development practices is described. A total of 44 primary studies were analysed as a result of the systematic review. As a result of the study, a list of twenty factors has been identified. Some of factors that affect implementation of secure software development practices are: Involvement of the security expert, integration between security and development team, developer’s skill and expertise, development time and communication between stakeholders. The factors were further classified into four categories which are institutional context, people and action, project content and system development process. The results obtained show that it is important to take into account organizational, technical and people issues in order to implement secure software development initiatives.

Keywords: Secure software development, software development, software security, systematic literature review.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2493
68 A Stochastic Analytic Hierarchy Process Based Weighting Model for Sustainability Measurement in an Organization

Authors: Faramarz Khosravi, Gokhan Izbirak

Abstract:

A weighted statistical stochastic based Analytical Hierarchy Process (AHP) model for modeling the potential barriers and enablers of sustainability for measuring and assessing the sustainability level is proposed. For context-dependent potential barriers and enablers, the proposed model takes the basis of the properties of the variables describing the sustainability functions and was developed into a realistic analytical model for the sustainable behavior of an organization. This thus serves as a means for measuring the sustainability of the organization. The main focus of this paper was the application of the AHP tool in a statistically-based model for measuring sustainability. Hence a strong weighted stochastic AHP based procedure was achieved. A case study scenario of a widely reported major Canadian electric utility was adopted to demonstrate the applicability of the developed model and comparatively examined its results with those of an equal-weighted model method. Variations in the sustainability of a company, as fluctuations, were figured out during the time. In the results obtained, sustainability index for successive years changed form 73.12%, 79.02%, 74.31%, 76.65%, 80.49%, 79.81%, 79.83% to more exact values 73.32%, 77.72%, 76.76%, 79.41%, 81.93%, 79.72%, and 80,45% according to priorities of factors that have found by expert views, respectively. By obtaining relatively necessary informative measurement indicators, the model can practically and effectively evaluate the sustainability extent of any organization and also to determine fluctuations in the organization over time.

Keywords: AHP, sustainability fluctuation, environmental indicators, performance measurement, environmental sustainability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 924
67 Designing an Editorialization Environment for Repeatable Self-Correcting Exercises

Authors: M. Kobylanski, D. Buskulic, P.-H. Duron, D. Revuz, F. Ruggieri, E. Sandier, C. Tijus

Abstract:

In order to design a cooperative e-learning platform, we observed teams of Teacher [T], Computer Scientist [CS] and exerciser's programmer-designer [ED] cooperating for the conception of a self-correcting exercise, but without the use of such a device in order to catch the kind of interactions a useful platform might provide. To do so, we first run a task analysis on how T, CS and ED should be cooperating in order to achieve, at best, the task of creating and implementing self-directed, self-paced, repeatable self-correcting exercises (RSE) in the context of open educational resources. The formalization of the whole process was based on the “objectives, activities and evaluations” theory of educational task analysis. Second, using the resulting frame as a “how-to-do it” guide, we run a series of three contrasted Hackathon of RSE-production to collect data about the cooperative process that could be later used to design the collaborative e-learning platform. Third, we used two complementary methods to collect, to code and to analyze the adequate survey data: the directional flow of interaction among T-CS-ED experts holding a functional role, and the Means-End Problem Solving analysis. Fourth, we listed the set of derived recommendations useful for the design of the exerciser as a cooperative e-learning platform. Final recommendations underline the necessity of building (i) an ecosystem that allows to sustain teams of T-CS-ED experts, (ii) a data safety platform although offering accessibility and open discussion about the production of exercises with their resources and (iii) a good architecture allowing the inheritance of parts of the coding of any exercise already in the data base as well as fast implementation of new kinds of exercises along with their associated learning activities.

Keywords: Distance open educational resources, pedagogical alignment, self-correcting exercises, teacher’s involvement, team roles.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 517
66 Qualitative Analysis of Current Child Custody Evaluation Practices

Authors: Carolyn J. Ortega, Stephen E. Berger

Abstract:

The role of the custody evaluator is perhaps one of the most controversial and risky endeavors in clinical practice. Complaints filed with licensing boards regarding a child-custody evaluation constitute the second most common reason for such an event. Although the evaluator is expected to answer for the family-law court what is in the “best interest of the child,” there is a lack of clarity on how to establish this in any empirically validated manner. Hence, practitioners must contend with a nebulous framework in formulating their methodological procedures that inherently places them at risk in an already litigious context. This study sought to qualitatively investigate patterns of practice among doctoral practitioners conducting child custody evaluations in the area of Southern California. Ten psychologists were interviewed who devoted between 25 and 100% of their California private practice to custody work. All held Ph.D. degrees with a range of eight to 36 years of experience in custody work. Semi-structured interviews were used to investigate assessment practices, ensure adherence to guidelines, risk management, and qualities of evaluators. Forty-three Specific Themes were identified using Interpretive Phenomenological Analysis (IPA). Seven Higher Order Themes clustered on salient factors such as use of Ethics, Law, Guidelines; Parent Variables; Child Variables; Psychologist Variables; Testing; Literature; and Trends. Evaluators were aware of the ever-present reality of a licensure complaint and thus presented idiosyncratic descriptions of risk management considerations. Ambiguity about quantifying and validly tapping parenting abilities was also reviewed. Findings from this study suggested a high reliance on unstructured and observational methods in child custody practices.

Keywords: Forensic psychology, psychological testing, assessment methodology, child custody.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827
65 Diagnostics of Existing Steel Structures of Winter Sport Halls

Authors: Marcela Karmazínová, Jindrich Melcher, Lubomír Vítek, Petr Cikrle

Abstract:

The paper deals with the diagnostics of steel roof structure of the winter sports halls built in 1970 year. The necessity of the diagnostics has been given by the requirement to the evaluation design of this structure, which has been caused by the new situation in the field of the loadings given by the validity of the European Standards in the Czech Republic from 2010 year. Due to these changes in the normative rules, in practice existing structures are gradually subjected to the evaluation design and depending on its results to the strengthening or reconstruction, respectively. Steel roof is composed of plane truss main girders, purlins and bracings and the roof structure is supported by two arch main girders with the span of L = 84 m. The in situ diagnostics of the roof structure was oriented to the following parts: (i) determination and evaluation of the actual material properties of used steel and (ii) verification of the actual dimensions of the structural members. For the solution the nondestructive methods have been used for in situ measurement. For the indicative determination of steel strengths the modified method based on the determination of Rockwell’s hardness has been used. For the verification of the member’s dimensions (thickness of hollow sections) the ultrasound method has been used. This paper presents the results obtained using these testing methods and their evaluation, from the viewpoint of the usage for the subsequent static assessment and design evaluation of the existing structure. For the comparison, the examples of the similar evaluations realized for steel structures of the stadiums in Olomouc and Jihlava cities are briefly illustrated, too.

Keywords: Diagnostics, existing steel structure, sport hall, steel strength, indirect non-destructive methods, Rockwel’s hardness, destructive methods, actual dimensions, ultrasound method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1983
64 Automatic Detection of Breast Tumors in Sonoelastographic Images Using DWT

Authors: A. Sindhuja, V. Sadasivam

Abstract:

Breast Cancer is the most common malignancy in women and the second leading cause of death for women all over the world. Earlier the detection of cancer, better the treatment. The diagnosis and treatment of the cancer rely on segmentation of Sonoelastographic images. Texture features has not considered for Sonoelastographic segmentation. Sonoelastographic images of 15 patients containing both benign and malignant tumorsare considered for experimentation.The images are enhanced to remove noise in order to improve contrast and emphasize tumor boundary. It is then decomposed into sub-bands using single level Daubechies wavelets varying from single co-efficient to six coefficients. The Grey Level Co-occurrence Matrix (GLCM), Local Binary Pattern (LBP) features are extracted and then selected by ranking it using Sequential Floating Forward Selection (SFFS) technique from each sub-band. The resultant images undergo K-Means clustering and then few post-processing steps to remove the false spots. The tumor boundary is detected from the segmented image. It is proposed that Local Binary Pattern (LBP) from the vertical coefficients of Daubechies wavelet with two coefficients is best suited for segmentation of Sonoelastographic breast images among the wavelet members using one to six coefficients for decomposition. The results are also quantified with the help of an expert radiologist. The proposed work can be used for further diagnostic process to decide if the segmented tumor is benign or malignant.

Keywords: Breast Cancer, Segmentation, Sonoelastography, Tumor Detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2207
63 Eliciting and Confirming Data, Information, Knowledge and Wisdom in a Specialist Health Care Setting: The WICKED Method

Authors: S. Impey, D. Berry, S. Furtado, M. Galvin, L. Grogan, O. Hardiman, L. Hederman, M. Heverin, V. Wade, L. Douris, D. O'Sullivan, G. Stephens

Abstract:

Healthcare is a knowledge-rich environment. This knowledge, while valuable, is not always accessible outside the borders of individual clinics. This research aims to address part of this problem (at a study site) by constructing a maximal data set (knowledge artefact) for motor neurone disease (MND). This data set is proposed as an initial knowledge base for a concurrent project to develop an MND patient data platform. It represents the domain knowledge at the study site for the duration of the research (12 months). A knowledge elicitation method was also developed from the lessons learned during this process - the WICKED method. WICKED is an anagram of the words: eliciting and confirming data, information, knowledge, wisdom. But it is also a reference to the concept of wicked problems, which are complex and challenging, as is eliciting expert knowledge. The method was evaluated at a second site, and benefits and limitations were noted. Benefits include that the method provided a systematic way to manage data, information, knowledge and wisdom (DIKW) from various sources, including healthcare specialists and existing data sets. Limitations surrounded the time required and how the data set produced only represents DIKW known during the research period. Future work is underway to address these limitations.

Keywords: Healthcare, knowledge acquisition, maximal data sets, action design science.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 543
62 A Multi-Radio Multi-Channel Unification Power Control for Wireless Mesh Networks

Authors: T. O. Olwal, K. Djouani, B. J. van Wyk, Y. Hamam, P. Siarry

Abstract:

Multi-Radio Multi-Channel Wireless Mesh Networks (MRMC-WMNs) operate at the backbone to access and route high volumes of traffic simultaneously. Such roles demand high network capacity, and long “online" time at the expense of accelerated transmission energy depletion and poor connectivity. This is the problem of transmission power control. Numerous power control methods for wireless networks are in literature. However, contributions towards MRMC configurations still face many challenges worth considering. In this paper, an energy-efficient power selection protocol called PMMUP is suggested at the Link-Layer. This protocol first divides the MRMC-WMN into a set of unified channel graphs (UCGs). A UCG consists of multiple radios interconnected to each other via a common wireless channel. In each UCG, a stochastic linear quadratic cost function is formulated. Each user minimizes this cost function consisting of trade-off between the size of unification states and the control action. Unification state variables come from independent UCGs and higher layers of the protocol stack. The PMMUP coordinates power optimizations at the network interface cards (NICs) of wireless mesh routers. The proposed PMMUP based algorithm converges fast analytically with a linear rate. Performance evaluations through simulations confirm the efficacy of the proposed dynamic power control.

Keywords: Effective band inference based power control algorithm (EBIA), Power Selection MRMC Unification Protocol (PMMUP), MRMC State unification Variable Prediction (MRSUP), Wireless Mesh Networks (WMNs).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1825
61 Resilience Assessment for Power Distribution Systems

Authors: Berna Eren Tokgoz, Mahdi Safa, Seokyon Hwang

Abstract:

Power distribution systems are essential and crucial infrastructures for the development and maintenance of a sustainable society. These systems are extremely vulnerable to various types of natural and man-made disasters. The assessment of resilience focuses on preparedness and mitigation actions under pre-disaster conditions. It also concentrates on response and recovery actions under post-disaster situations. The aim of this study is to present a methodology to assess the resilience of electric power distribution poles against wind-related events. The proposed methodology can improve the accuracy and rapidity of the evaluation of the conditions and the assessment of the resilience of poles. The methodology provides a metric for the evaluation of the resilience of poles under pre-disaster and post-disaster conditions. The metric was developed using mathematical expressions for physical forces that involve various variables, such as physical dimensions of the pole, the inclination of the pole, and wind speed. A three-dimensional imaging technology (photogrammetry) was used to determine the inclination of poles. Based on expert opinion, the proposed metric was used to define zones to visualize resilience. Visual representation of resilience is helpful for decision makers to prioritize their resources before and after experiencing a wind-related disaster. Multiple electric poles in the City of Beaumont, TX were used in a case study to evaluate the proposed methodology.  

Keywords: Photogrammetry, power distribution systems, resilience metric, system resilience, wind-related disasters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1421
60 Identifying Quality Islamic Content in Community Question Answering Sites

Authors: Rabia Bibi, Muhammad Shahzad Faisal, Khalid Iqbal, Atif Inayat

Abstract:

Internet is growing rapidly and new community-based content is added by people every second. With this fast-growing community-based content, if a user requires answers of particular questions, then reviews are required from experts or community. However, it is difficult to get quality answers. The Muslim community all over the world is seeking help to get their questions and issues discussed to get answers. Online web portals of religious schools and community-based question answering sites are two big platforms to solve the issues of users. In the case of religious schools, there are experts and qualified religious scholars (mufti) who can give the expert opinion. However, the quality of community-based content cannot be guaranteed as it may not be an answer that satisfies the question of a user. Users on CQA sites may include spammers or individual criticizing the questioner instead of providing useful answers. In this paper, we research strategies to naturally distinguish the right content. As an experiment, we concentrate on Yahoo! Answers, and Quora, popular online QA sites, where questions are asked, answered, edited, and organized by a large community of users. We present the classification of data to categorize both relevant and irrelevant answers. Specifically, we demonstrate that the proposed framework can isolate quality answers from the rest with an exactness near that of people.

Keywords: Community-based question and answering, evaluation and prediction of quality answer, answer classification, Islamic content, answer ranking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 80
59 A Novel VLSI Architecture for Image Compression Model Using Low power Discrete Cosine Transform

Authors: Vijaya Prakash.A.M, K.S.Gurumurthy

Abstract:

In Image processing the Image compression can improve the performance of the digital systems by reducing the cost and time in image storage and transmission without significant reduction of the Image quality. This paper describes hardware architecture of low complexity Discrete Cosine Transform (DCT) architecture for image compression[6]. In this DCT architecture, common computations are identified and shared to remove redundant computations in DCT matrix operation. Vector processing is a method used for implementation of DCT. This reduction in computational complexity of 2D DCT reduces power consumption. The 2D DCT is performed on 8x8 matrix using two 1-Dimensional Discrete cosine transform blocks and a transposition memory [7]. Inverse discrete cosine transform (IDCT) is performed to obtain the image matrix and reconstruct the original image. The proposed image compression algorithm is comprehended using MATLAB code. The VLSI design of the architecture is implemented Using Verilog HDL. The proposed hardware architecture for image compression employing DCT was synthesized using RTL complier and it was mapped using 180nm standard cells. . The Simulation is done using Modelsim. The simulation results from MATLAB and Verilog HDL are compared. Detailed analysis for power and area was done using RTL compiler from CADENCE. Power consumption of DCT core is reduced to 1.027mW with minimum area[1].

Keywords: Discrete Cosine Transform (DCT), Inverse DiscreteCosine Transform (IDCT), Joint Photographic Expert Group (JPEG), Low Power Design, Very Large Scale Integration (VLSI) .

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3139
58 Performing Diagnosis in Building with Partially Valid Heterogeneous Tests

Authors: Houda Najeh, Mahendra Pratap Singh, Stéphane Ploix, Antoine Caucheteux, Karim Chabir, Mohamed Naceur Abdelkrim

Abstract:

Building system is highly vulnerable to different kinds of faults and human misbehaviors. Energy efficiency and user comfort are directly targeted due to abnormalities in building operation. The available fault diagnosis tools and methodologies particularly rely on rules or pure model-based approaches. It is assumed that model or rule-based test could be applied to any situation without taking into account actual testing contexts. Contextual tests with validity domain could reduce a lot of the design of detection tests. The main objective of this paper is to consider fault validity when validate the test model considering the non-modeled events such as occupancy, weather conditions, door and window openings and the integration of the knowledge of the expert on the state of the system. The concept of heterogeneous tests is combined with test validity to generate fault diagnoses. A combination of rules, range and model-based tests known as heterogeneous tests are proposed to reduce the modeling complexity. Calculation of logical diagnoses coming from artificial intelligence provides a global explanation consistent with the test result. An application example shows the efficiency of the proposed technique: an office setting at Grenoble Institute of Technology.

Keywords: Heterogeneous tests, validity, building system, sensor grids, sensor fault, diagnosis, fault detection and isolation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 652
57 Utilizing Analytic Hierarchy Process to Analyze Consumers- Purchase Evaluation Factors of Smartphones

Authors: Yi-Chung Hu, Yu-Lin Liao

Abstract:

Due to the fast development of technology, the competition of technological products is turbulent; therefore, it is important to understand the market trend, consumers- demand and preferences. As the smartphones are prevalent, the main purpose of this paper is to utilize Analytic Hierarchy Process (AHP) to analyze consumer-s purchase evaluation factors of smartphones. Through the AHP expert questionnaire, the smartphones- main functions are classified as “user interface", “mobile commerce functions", “hardware and software specifications", “entertainment functions" and “appearance and design", five aspects to analyze the weights. Then four evaluation criteria are evaluated under each aspect to rank the weights. Based on an analysis of data shows that consumers consider when purchase factors are “hardware and software specifications", “user interface", “appearance and design", “mobile commerce functions" and “entertainment functions" in sequence. The “hardware and software specifications" aspect obtains the weight of 33.18%; it is the most important factor that consumers are taken into account. In addition, the most important evaluation criteria are central processing unit, operating system, touch screen, and battery function in sequence. The results of the study can be adopted as reference data for mobile phone manufacturers in the future on the design and marketing strategy to satisfy the voice of customer.

Keywords: Analytic Hierarchy Process (AHP), evaluation criteria, purchase evaluation factors, smartphone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3240