Search results for: semantic errors.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 718

Search results for: semantic errors.

298 Rule-Based Expert System for Headache Diagnosis and Medication Recommendation

Authors: Noura Al-Ajmi, Mohammed A. Almulla

Abstract:

With the increased utilization of technology devices around the world, healthcare and medical diagnosis are critical issues that people worry about these days. Doctors are doing their best to avoid any medical errors while diagnosing diseases and prescribing the wrong medication. Subsequently, artificial intelligence applications that can be installed on mobile devices such as rule-based expert systems facilitate the task of assisting doctors in several ways. Due to their many advantages, the usage of expert systems has increased recently in health sciences. This work presents a backward rule-based expert system that can be used for a headache diagnosis and medication recommendation system. The structure of the system consists of three main modules, namely the input unit, the processing unit, and the output unit.

Keywords: Headache diagnosis system, treatment recommender system, rule-based expert system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 681
297 Kinetic Study of Thermal Degradation of a Lignin Nanoparticle-Reinforced Phenolic Foam

Authors: Juan C. Domínguez, Belén Del Saz-Orozco, María V. Alonso, Mercedes Oliet, Francisco Rodríguez

Abstract:

In the present study, the kinetics of thermal degradation of a phenolic and lignin reinforced phenolic foams, and the lignin used as reinforcement were studied and the activation energies of their degradation processes were obtained by a DAEM model. The average values for five heating rates of the mean activation energies obtained were: 99.1, 128.2, and 144.0 kJ.mol-1 for the phenolic foam; 109.5, 113.3, and 153.0 kJ.mol-1 for the lignin reinforcement; and 82.1, 106.9, and 124.4 kJ.mol-1 for the lignin reinforced phenolic foam. The standard deviation ranges calculated for each sample were 1.27-8.85, 2.22-12.82, and 3.17-8.11 kJ.mol-1 for the phenolic foam, lignin and the reinforced foam, respectively. The DAEM model showed low mean square errors (<1x10-5), proving that is a suitable model to study the kinetics of thermal degradation of the foams and the reinforcement.

Keywords: Kinetics, lignin, phenolic foam, thermal degradation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1882
296 Effect of Channel Estimation on Capacity of MIMO System Employing Circular or Linear Receiving Array Antennas

Authors: Xia Liu, Marek E. Bialkowski

Abstract:

This paper reports on investigations into capacity of a Multiple Input Multiple Output (MIMO) wireless communication system employing a uniform linear array (ULA) at the transmitter and either a uniform linear array (ULA) or a uniform circular array (UCA) antenna at the receiver. The transmitter is assumed to be surrounded by scattering objects while the receiver is postulated to be free from scattering objects. The Laplacian distribution of angle of arrival (AOA) of a signal reaching the receiver is postulated. Calculations of the MIMO system capacity are performed for two cases without and with the channel estimation errors. For estimating the MIMO channel, the scaled least square (SLS) and minimum mean square error (MMSE) methods are considered.

Keywords: MIMO, channel capacity, channel estimation, ULA, UCA, spatial correlation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1330
295 Kinetics of Polyethylene Terephthalate (PET)and Polystyrene (PS) Dynamic Pyrolysis

Authors: S.M. Al-Salem, P. Lettieri

Abstract:

Thermo-chemical treatment (TCT) such as pyrolysis is getting recognized as a valid route for (i) materials and valuable products and petrochemicals recovery; (ii) waste recycling; and (iii) elemental characterization. Pyrolysis is also receiving renewed attention for its operational, economical and environmental advantages. In this study, samples of polyethylene terephthalate (PET) and polystyrene (PS) were pyrolysed in a microthermobalance reactor (using a thermogravimetric-TGA setup). Both polymers were prepared and conditioned prior to experimentation. The main objective was to determine the kinetic parameters of the depolymerization reactions that occur within the thermal degradation process. Overall kinetic rate constants (ko) and activation energies (Eo) were determined using the general kinetics theory (GKT) method previously used by a number of authors. Fitted correlations were found and validated using the GKT, errors were within ± 5%. This study represents a fundamental step to pave the way towards the development of scaling relationship for the investigation of larger scale reactors relevant to industry.

Keywords: Kinetics, PET, PS, Pyrolysis, Recycling, Petrochemicals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3684
294 Computer - based Systems for High Speed Vessels Navigators – Engineers Training

Authors: D. E. Gourgoulis, C. G. Yakinthos, M. G. Vassiliadou

Abstract:

With high speed vessels getting ever more sophisti-cated, travelling at higher and higher speeds and operating in With high speed vessels getting ever more sophisticated, travelling at higher and higher speeds and operating in areas of high maritime traffic density, training becomes of the highest priority to ensure that safety levels are maintained, and risks are adequately mitigated. Training onboard the actual craft on the actual route still remains the most effective way for crews to gain experience. However, operational experience and incidents during the last 10 years demonstrate the need for supplementary training whether in the area of simulation or man to man, man/ machine interaction. Training and familiarisation of the crew is the most important aspect in preventing incidents. The use of simulator, computer and web based training systems in conjunction with onboard training focusing on critical situations will improve the man machine interaction and thereby reduce the risk of accidents. Today, both ship simulator and bridge teamwork courses are now becoming the norm in order to improve further emergency response and crisis management skills. One of the main causes of accidents is the human factor. An efficient way to reduce human errors is to provide high-quality training to the personnel and to select the navigators carefully.areas of high maritime traffic density, training becomes of the highest priority to ensure that safety levels are maintained, and risks are adequately mitigated. Training onboard the actual craft on the actual route still remains the most effective way for crews to gain experience. How-ever, operational experience and incidents during the last 10 years demonstrate the need for supplementary training whether in the area of simulation or man to man, man/ machine interaction. Training and familiarisation of the crew is the most important aspect in preventing incidents. The use of simulator, computer and web based training systems in conjunction with onboard training focusing on critical situations will improve the man machine interaction and thereby reduce the risk of accidents. Today, both ship simulator and bridge teamwork courses are now becoming the norm in order to improve further emergency response and crisis management skills. One of the main causes of accidents is the human factor. An efficient way to reduce human errors is to provide high-quality training to the person-nel and to select the navigators carefully. KeywordsCBT - WBT systems, Human factors.

Keywords: CBT - WBT systems, Human factors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
293 Optimizing the Capacity of a Convolutional Neural Network for Image Segmentation and Pattern Recognition

Authors: Yalong Jiang, Zheru Chi

Abstract:

In this paper, we study the factors which determine the capacity of a Convolutional Neural Network (CNN) model and propose the ways to evaluate and adjust the capacity of a CNN model for best matching to a specific pattern recognition task. Firstly, a scheme is proposed to adjust the number of independent functional units within a CNN model to make it be better fitted to a task. Secondly, the number of independent functional units in the capsule network is adjusted to fit it to the training dataset. Thirdly, a method based on Bayesian GAN is proposed to enrich the variances in the current dataset to increase its complexity. Experimental results on the PASCAL VOC 2010 Person Part dataset and the MNIST dataset show that, in both conventional CNN models and capsule networks, the number of independent functional units is an important factor that determines the capacity of a network model. By adjusting the number of functional units, the capacity of a model can better match the complexity of a dataset.

Keywords: CNN, capsule network, capacity optimization, character recognition, data augmentation; semantic segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 654
292 Talent in Autism: Cognitive Style based on Weak Central Coherence and Special Sensory Characteristics in State of Kuwait: Case Study

Authors: Mariam Abdulaziz Y.Esmaeel

Abstract:

The study aimed to identify the nature of autistic talent, the manifestations of their weak central coherence, and their sensory characteristics. The case study consisted of four talented autistic males. Two of them in drawing, one in clay formation and one in jigsaw puzzle. Tools of data collection were Group Embedded Figures Test, Block Design Test, Sensory Profile Checklist Revised, Interview forms and direct observation. Results indicated that talent among autistics emerges in limited domain and being extraordinary for each case. Also overlapping construction properties. Indeed, they show three perceptual aspects of weak central coherence: The weak in visual spatial-constructional coherence, the weak in perceptual coherence and the weak in verbal – semantic coherence. Moreover, the majority of the study cases used the three strategies of weak central coherence (segmentation, obliqueness and rotation). As for the sensory characteristics, all study cases have numbers of that characteristics that especially emerges in the visual system.

Keywords: Autism, Central Coherence, Savant, Sensory characteristics, Talent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2634
291 Traffic Density Estimation for Multiple Segment Freeways

Authors: Karandeep Singh, Baibing Li

Abstract:

Traffic density, an indicator of traffic conditions, is one of the most critical characteristics to Intelligent Transport Systems (ITS). This paper investigates recursive traffic density estimation using the information provided from inductive loop detectors. On the basis of the phenomenological relationship between speed and density, the existing studies incorporate a state space model and update the density estimate using vehicular speed observations via the extended Kalman filter, where an approximation is made because of the linearization of the nonlinear observation equation. In practice, this may lead to substantial estimation errors. This paper incorporates a suitable transformation to deal with the nonlinear observation equation so that the approximation is avoided when using Kalman filter to estimate the traffic density. A numerical study is conducted. It is shown that the developed method outperforms the existing methods for traffic density estimation.

Keywords: Density estimation, Kalman filter, speed-densityrelationship, Traffic surveillance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1799
290 UD Covariance Factorization for Unscented Kalman Filter using Sequential Measurements Update

Authors: H. Ghanbarpour Asl, S. H. Pourtakdoust

Abstract:

Extended Kalman Filter (EKF) is probably the most widely used estimation algorithm for nonlinear systems. However, not only it has difficulties arising from linearization but also many times it becomes numerically unstable because of computer round off errors that occur in the process of its implementation. To overcome linearization limitations, the unscented transformation (UT) was developed as a method to propagate mean and covariance information through nonlinear transformations. Kalman filter that uses UT for calculation of the first two statistical moments is called Unscented Kalman Filter (UKF). Square-root form of UKF (SRUKF) developed by Rudolph van der Merwe and Eric Wan to achieve numerical stability and guarantee positive semi-definiteness of the Kalman filter covariances. This paper develops another implementation of SR-UKF for sequential update measurement equation, and also derives a new UD covariance factorization filter for the implementation of UKF. This filter is equivalent to UKF but is computationally more efficient.

Keywords: Unscented Kalman filter, Square-root unscentedKalman filter, UD covariance factorization, Target tracking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4788
289 Mechanical Quadrature Methods for Solving First Kind Boundary Integral Equations of Stationary Stokes Problem

Authors: Xin Luo, Jin Huang, Pan Cheng

Abstract:

By means of Sidi-Israeli’s quadrature rules, mechanical quadrature methods (MQMs) for solving the first kind boundary integral equations (BIEs) of steady state Stokes problem are presented. The convergence of numerical solutions by MQMs is proved based on Anselone’s collective compact and asymptotical compact theory, and the asymptotic expansions with the odd powers of the errors are provided, which implies that the accuracy of the approximations by MQMs possesses high accuracy order O (h3). Finally, the numerical examples show the efficiency of our methods.

Keywords: Stokes problem, boundary integral equation, mechanical quadrature methods, asymptotic expansions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1358
288 Quality Control of Automotive Gearbox Based On Vibration Signal Analysis

Authors: Nilson Barbieri, Bruno Matos Martins, Gabriel de Sant'Anna Vitor Barbieri

Abstract:

In more complex systems, such as automotive gearbox, a rigorous treatment of the data is necessary because there are several moving parts (gears, bearings, shafts, etc.), and in this way, there are several possible sources of errors and also noise. The basic objective of this work is the detection of damage in automotive gearbox. The detection methods used are the wavelet method, the bispectrum; advanced filtering techniques (selective filtering) of vibrational signals and mathematical morphology. Gearbox vibration tests were performed (gearboxes in good condition and with defects) of a production line of a large vehicle assembler. The vibration signals are obtained using five accelerometers in different positions of the sample. The results obtained using the kurtosis, bispectrum, wavelet and mathematical morphology showed that it is possible to identify the existence of defects in automotive gearboxes.

Keywords: Automotive gearbox, mathematical morphology, wavelet, bispectrum.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2276
287 DocPro: A Framework for Processing Semantic and Layout Information in Business Documents

Authors: Ming-Jen Huang, Chun-Fang Huang, Chiching Wei

Abstract:

With the recent advance of the deep neural network, we observe new applications of NLP (natural language processing) and CV (computer vision) powered by deep neural networks for processing business documents. However, creating a real-world document processing system needs to integrate several NLP and CV tasks, rather than treating them separately. There is a need to have a unified approach for processing documents containing textual and graphical elements with rich formats, diverse layout arrangement, and distinct semantics. In this paper, a framework that fulfills this unified approach is presented. The framework includes a representation model definition for holding the information generated by various tasks and specifications defining the coordination between these tasks. The framework is a blueprint for building a system that can process documents with rich formats, styles, and multiple types of elements. The flexible and lightweight design of the framework can help build a system for diverse business scenarios, such as contract monitoring and reviewing.

Keywords: Document processing, framework, formal definition, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 583
286 Sliding Mode Control of Autonomous Underwater Vehicles

Authors: Ahmad Forouzan Tabar, Mohammad Azadi, Alireza Alesaadi

Abstract:

This paper describes a sliding mode controller for autonomous underwater vehicles (AUVs). The dynamic of AUV model is highly nonlinear because of many factors, such as hydrodynamic drag, damping, and lift forces, Coriolis and centripetal forces, gravity and buoyancy forces, as well as forces from thruster. To address these difficulties, a nonlinear sliding mode controller is designed to approximate the nonlinear dynamics of AUV and improve trajectory tracking. Moreover, the proposed controller can profoundly attenuate the effects of uncertainties and external disturbances in the closed-loop system. Using the Lyapunov theory the boundedness of AUV tracking errors and the stability of the proposed control system are also guaranteed. Numerical simulation studies of an AUV are included to illustrate the effectiveness of the presented approach.

Keywords: Lyapunov stability, autonomous underwater vehicle (AUV), sliding mode controller, electronics engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2562
285 Learning Process Enhancement for Robot Behaviors

Authors: Saeed Mohammed Baneamoon, Rosalina Abdul Salam, Abdullah Zawawi Hj. Talib

Abstract:

Designing a simulated system and training it to optimize its tasks in simulated environment helps the designers to avoid problems that may appear when designing the system directly in real world. These problems are: time consuming, high cost, high errors percentage and low efficiency and accuracy of the system. The proposed system will investigate and improve the efficiency and accuracy of a simulated robot to choose correct behavior to perform its task. In this paper, machine learning, which uses genetic algorithm, is adopted. This type of machine learning is called genetic-based machine learning in which a distributed classifier system is used to improve the efficiency and accuracy of the robot. Consequently, it helps the robot to achieve optimal action.

Keywords: Machine Learning, Genetic-Based MachineLearning, Learning Classifier System, Behaviors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
284 A Consideration on the Offset Frontal Impact Modeling Using Spring-Mass Model

Authors: Jaemoon Lim

Abstract:

To construct the lumped spring-mass model considering the occupants for the offset frontal crash, the SISAME software and the NHTSA test data were used. The data on 56 kph 40% offset frontal vehicle to deformable barrier crash test of a MY2007 Mazda 6 4-door sedan were obtained from NHTSA test database. The overall behaviors of B-pillar and engine of simulation models agreed very well with the test data. The trends of accelerations at the driver and passenger head were similar but big differences in peak values. The differences of peak values caused the large errors of the HIC36 and 3 ms chest g’s. To predict well the behaviors of dummies, the spring-mass model for the offset frontal crash needs to be improved.

Keywords: Chest g’s, HIC36, lumped spring-mass model, offset frontal impact, SISAME.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2636
283 Smartphones for In-home Diagnostics in Telemedicine

Authors: Nálevka Petr

Abstract:

Many contemporary telemedical applications rely on regular consultations over the phone or video conferencing which consumes valuable resources such as the time of the doctors. Some applications or treatments allow automated diagnostics on the patient side which only notifies the doctors in case a significant worsening of patient’s condition is measured. Such programs can save valuable resources but an important implementation issue is how to ensure effective and cheap diagnostics on the patient side. First, specific diagnostic devices on patient side are expensive and second, they need to be user-˜friendly to encourage patient’s cooperation and reduce errors in usage which may cause noise in diagnostic data. This article proposes the use of modern smartphones and various build-in or attachable sensors as universal diagnostic devices applicable in a wider range of telemedical programs and demonstrates their application on a case-study – a program for schizophrenic relapse prevention.

Keywords: Smartphones, Actigraphy, Telemedicine, In-home Diagnostics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1838
282 Security Architecture for At-Home Medical Care Using Sensor Network

Authors: S.S.Mohanavalli, Sheila Anand

Abstract:

This paper proposes a novel architecture for At- Home medical care which enables senior citizens, patients with chronic ailments and patients requiring post- operative care to be remotely monitored in the comfort of their homes. This architecture is implemented using sensors and wireless networking for transmitting patient data to the hospitals, health- care centers for monitoring by medical professionals. Patients are equipped with sensors to measure their physiological parameters, like blood pressure, pulse rate etc. and a Wearable Data Acquisition Unit is used to transmit the patient sensor data. Medical professionals can be alerted to any abnormal variations in these values for diagnosis and suitable treatment. Security threats and challenges inherent to wireless communication and sensor network have been discussed and a security mechanism to ensure data confidentiality and source authentication has been proposed. Symmetric key algorithm AES has been used for encrypting the data and a patent-free, two-pass block cipher mode CCFB has been used for implementing semantic security.

Keywords: data confidentiality, integrity, remotemonitoring, source authentication

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1694
281 Exp-Function Method for Finding Some Exact Solutions of Rosenau Kawahara and Rosenau Korteweg-de Vries Equations

Authors: Ehsan Mahdavi

Abstract:

In this paper, we apply the Exp-function method to Rosenau-Kawahara and Rosenau-KdV equations. Rosenau-Kawahara equation is the combination of the Rosenau and standard Kawahara equations and Rosenau-KdV equation is the combination of the Rosenau and standard KdV equations. These equations are nonlinear partial differential equations (NPDE) which play an important role in mathematical physics. Exp-function method is easy, succinct and powerful to implement to nonlinear partial differential equations arising in mathematical physics. We mainly try to present an application of Exp-function method and offer solutions for common errors wich occur during some of the recent works.

Keywords: Exp-function method, Rosenau Kawahara equation, Rosenau Korteweg-de Vries equation, nonlinear partial differential equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2013
280 The Design of the HL7 RIM-based Sharing Components for Clinical Information Systems

Authors: Wei-Yi Yang, Li-Hui Lee, Hsiao-Li Gien, Hsing-Yi Chu, Yi-Ting Chou, Der-Ming Liou

Abstract:

The American Health Level Seven (HL7) Reference Information Model (RIM) consists of six back-bone classes that have different specialized attributes. Furthermore, for the purpose of enforcing the semantic expression, there are some specific mandatory vocabulary domains have been defined for representing the content values of some attributes. In the light of the fact that it is a duplicated effort on spending a lot of time and human cost to develop and modify Clinical Information Systems (CIS) for most hospitals due to the variety of workflows. This study attempts to design and develop sharing RIM-based components of the CIS for the different business processes. Therefore, the CIS contains data of a consistent format and type. The programmers can do transactions with the RIM-based clinical repository by the sharing RIM-based components. And when developing functions of the CIS, the sharing components also can be adopted in the system. These components not only satisfy physicians- needs in using a CIS but also reduce the time of developing new components of a system. All in all, this study provides a new viewpoint that integrating the data and functions with the business processes, it is an easy and flexible approach to build a new CIS.

Keywords: HL7, Reference Information Model (RIM), web service, process management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1845
279 Hybrid Fuzzy Selecting-Control-by- Range Controllers of a Servopneumatic Fatigue System

Authors: Marco Soares dos Santos, Jorge Augusto Ferreira, Camila Nicola Boeri, Fernando Neto da Silva

Abstract:

The present paper proposes high performance nonlinear force controllers for a servopneumatic real-time fatigue test machine. A CompactRIO® controller was used, being fully programmed using LabVIEW language. Fuzzy logic control algorithms were evaluated to tune the integral and derivative components in the development of hybrid controllers, namely a FLC P and a hybrid FLC PID real-time-based controllers. Their behaviours were described by using state diagrams. The main contribution is to ensure a smooth transition between control states, avoiding discrete transitions in controller outputs. Steady-state errors lower than 1.5 N were reached, without retuning the controllers. Good results were also obtained for sinusoidal tracking tasks from 1/¤Ç to 8/¤Ç Hz.

Keywords: Hybrid Fuzzy Selecting, Control, Range Controllers, Servopneumatic Fatigue System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957
278 Video Quality Control Using a ROI and Two- Component Weighted Metrics

Authors: Petra Heribanová, Jaroslav Polec, Michal Martinovič

Abstract:

In this paper we propose a new content-weighted method for full reference (FR) video quality control using a region of interest (ROI) and wherein two-component weighted metrics for Deaf People Video Communication. In our approach, an image is partitioned into region of interest and into region "dry-as-dust", then region of interest is partitioned into two parts: edges and background (smooth regions), while the another methods (metrics) combined and weighted three or more parts as edges, edges errors, texture, smooth regions, blur, block distance etc. as we proposed. Using another idea that different image regions from deaf people video communication have different perceptual significance relative to quality. Intensity edges certainly contain considerable image information and are perceptually significant.

Keywords: Video quality assessment, weighted MSE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1941
277 Analysis of Simple Mechanisms to Continuously Vary Mach Number in a Supersonic Wind Tunnel Facility

Authors: Prateek Kishore, T. M. Muruganandam

Abstract:

Supersonic wind tunnel nozzles are generally capable of producing a constant Mach number flow in the test section of the wind tunnel. As a result, most of the supersonic vehicles are widely designed using steady state flow characteristics which may have errors while facing unsteady situations. This study aims to explore the possibility of varying the Mach number of the flow during wind tunnel operation. The nozzle walls are restricted to be inflexible for cooling near the throat due to high stagnation temperature requirement of the flow to simulate the conditions as experienced by the vehicle. Two simple independent mechanisms, rotation and translation of nozzle walls have been analyzed and the nozzle ranges have been optimized to vary the Mach number from Mach 2 to Mach 5 using minimum number of nozzles in the wind tunnel.

Keywords: Method of characteristics, Nozzle, supersonic wind tunnel, variable Mach number.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 974
276 A New Fuzzy Decision Support Method for Analysis of Economic Factors of Turkey's Construction Industry

Authors: R. Tur, A. Yardımcı

Abstract:

Imperfect knowledge cannot be avoided all the time. Imperfections may have several forms; uncertainties, imprecision and incompleteness. When we look to classification of methods for the management of imperfect knowledge we see fuzzy set-based techniques. The choice of a method to process data is linked to the choice of knowledge representation, which can be numerical, symbolic, logical or semantic and it depends on the nature of the problem to be solved for example decision support, which will be mentioned in our study. Fuzzy Logic is used for its ability to manage imprecise knowledge, but it can take advantage of the ability of neural networks to learn coefficients or functions. Such an association of methods is typical of so-called soft computing. In this study a new method was used for the management of imprecision for collected knowledge which related to economic analysis of construction industry in Turkey. Because of sudden changes occurring in economic factors decrease competition strength of construction companies. The better evaluation of these changes in economical factors in view of construction industry will made positive influence on company-s decisions which are dealing construction.

Keywords: Fuzzy logic, decision support systems, construction industry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1602
275 Preliminary Study on Fixture Layout Optimization Using Element Strain Energy

Authors: Zeshan Ahmad, Matteo Zoppi, Rezia Molfino

Abstract:

The objective of positioning the fixture elements in the fixture is to make the workpiece stiff, so that geometric errors in the manufacturing process can be reduced. Most of the work for optimal fixture layout used the minimization of the sum of the nodal deflection normal to the surface as objective function. All deflections in other direction have been neglected. We propose a new method for fixture layout optimization in this paper, which uses the element strain energy. The deformations in all the directions have been considered in this way. The objective function in this method is to minimize the sum of square of element strain energy. Strain energy and stiffness are inversely proportional to each other. The optimization problem is solved by the sequential quadratic programming method. Three different kinds of case studies are presented, and results are compared with the method using nodal deflections as objective function to verify the propose method.

Keywords: Fixture layout, optimization, strain energy, quadratic programming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1514
274 Ontology Population via NLP Techniques in Risk Management

Authors: Jawad Makki, Anne-Marie Alquier, Violaine Prince

Abstract:

In this paper we propose an NLP-based method for Ontology Population from texts and apply it to semi automatic instantiate a Generic Knowledge Base (Generic Domain Ontology) in the risk management domain. The approach is semi-automatic and uses a domain expert intervention for validation. The proposed approach relies on a set of Instances Recognition Rules based on syntactic structures, and on the predicative power of verbs in the instantiation process. It is not domain dependent since it heavily relies on linguistic knowledge. A description of an experiment performed on a part of the ontology of the PRIMA1 project (supported by the European community) is given. A first validation of the method is done by populating this ontology with Chemical Fact Sheets from Environmental Protection Agency2. The results of this experiment complete the paper and support the hypothesis that relying on the predicative power of verbs in the instantiation process improves the performance.

Keywords: Information Extraction, Instance Recognition Rules, Ontology Population, Risk Management, Semantic analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1504
273 Average Switching Thresholds and Average Throughput for Adaptive Modulation using Markov Model

Authors: Essam S. Altubaishi

Abstract:

The motivation for adaptive modulation and coding is to adjust the method of transmission to ensure that the maximum efficiency is achieved over the link at all times. The receiver estimates the channel quality and reports it back to the transmitter. The transmitter then maps the reported quality into a link mode. This mapping however, is not a one-to-one mapping. In this paper we investigate a method for selecting the proper modulation scheme. This method can dynamically adapt the mapping of the Signal-to- Noise Ratio (SNR) into a link mode. It enables the use of the right modulation scheme irrespective of changes in the channel conditions by incorporating errors in the received data. We propose a Markov model for this method, and use it to derive the average switching thresholds and the average throughput. We show that the average throughput of this method outperforms the conventional threshold method.

Keywords: Adaptive modulation and coding, CDMA, Markov model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697
272 GPU-Accelerated Triangle Mesh Simplification Using Parallel Vertex Removal

Authors: Thomas Odaker, Dieter Kranzlmueller, Jens Volkert

Abstract:

We present an approach to triangle mesh simplification designed to be executed on the GPU. We use a quadric error metric to calculate an error value for each vertex of the mesh and order all vertices based on this value. This step is followed by the parallel removal of a number of vertices with the lowest calculated error values. To allow for the parallel removal of multiple vertices we use a set of per-vertex boundaries that prevent mesh foldovers even when simplification operations are performed on neighbouring vertices. We execute multiple iterations of the calculation of the vertex errors, ordering of the error values and removal of vertices until either a desired number of vertices remains in the mesh or a minimum error value is reached. This parallel approach is used to speed up the simplification process while maintaining mesh topology and avoiding foldovers at every step of the simplification.

Keywords: Computer graphics, half edge collapse, mesh simplification, precomputed simplification, topology preserving.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2749
271 Feature-Based Summarizing and Ranking from Customer Reviews

Authors: Dim En Nyaung, Thin Lai Lai Thein

Abstract:

Due to the rapid increase of Internet, web opinion sources dynamically emerge which is useful for both potential customers and product manufacturers for prediction and decision purposes. These are the user generated contents written in natural languages and are unstructured-free-texts scheme. Therefore, opinion mining techniques become popular to automatically process customer reviews for extracting product features and user opinions expressed over them. Since customer reviews may contain both opinionated and factual sentences, a supervised machine learning technique applies for subjectivity classification to improve the mining performance. In this paper, we dedicate our work is the task of opinion summarization. Therefore, product feature and opinion extraction is critical to opinion summarization, because its effectiveness significantly affects the identification of semantic relationships. The polarity and numeric score of all the features are determined by Senti-WordNet Lexicon. The problem of opinion summarization refers how to relate the opinion words with respect to a certain feature. Probabilistic based model of supervised learning will improve the result that is more flexible and effective.

Keywords: Opinion Mining, Opinion Summarization, Sentiment Analysis, Text Mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2898
270 A Universal Model for Content-Based Image Retrieval

Authors: S. Nandagopalan, Dr. B. S. Adiga, N. Deepak

Abstract:

In this paper a novel approach for generalized image retrieval based on semantic contents is presented. A combination of three feature extraction methods namely color, texture, and edge histogram descriptor. There is a provision to add new features in future for better retrieval efficiency. Any combination of these methods, which is more appropriate for the application, can be used for retrieval. This is provided through User Interface (UI) in the form of relevance feedback. The image properties analyzed in this work are by using computer vision and image processing algorithms. For color the histogram of images are computed, for texture cooccurrence matrix based entropy, energy, etc, are calculated and for edge density it is Edge Histogram Descriptor (EHD) that is found. For retrieval of images, a novel idea is developed based on greedy strategy to reduce the computational complexity. The entire system was developed using AForge.Imaging (an open source product), MATLAB .NET Builder, C#, and Oracle 10g. The system was tested with Coral Image database containing 1000 natural images and achieved better results.

Keywords: Content Based Image Retrieval (CBIR), Cooccurrencematrix, Feature vector, Edge Histogram Descriptor(EHD), Greedy strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2893
269 Liver Lesion Extraction with Fuzzy Thresholding in Contrast Enhanced Ultrasound Images

Authors: Abder-Rahman Ali, Adélaïde Albouy-Kissi, Manuel Grand-Brochier, Viviane Ladan-Marcus, Christine Hoeffl, Claude Marcus, Antoine Vacavant, Jean-Yves Boire

Abstract:

In this paper, we present a new segmentation approach for focal liver lesions in contrast enhanced ultrasound imaging. This approach, based on a two-cluster Fuzzy C-Means methodology, considers type-II fuzzy sets to handle uncertainty due to the image modality (presence of speckle noise, low contrast, etc.), and to calculate the optimum inter-cluster threshold. Fine boundaries are detected by a local recursive merging of ambiguous pixels. The method has been tested on a representative database. Compared to both Otsu and type-I Fuzzy C-Means techniques, the proposed method significantly reduces the segmentation errors.

Keywords: Defuzzification, fuzzy clustering, image segmentation, type-II fuzzy sets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2255