Search results for: uncertainty of processing times
2140 Fighter Aircraft Selection Using Neutrosophic Multiple Criteria Decision Making Analysis
Authors: C. Ardil
Abstract:
Fuzzy set and intuitionistic fuzzy set are dealing with the imprecision and uncertainty inherent in a complex decision problem. However, sometimes these theories are not sufficient to model indeterminate and inconsistent information encountered in real-life problems. To overcome this insufficiency, the neutrosophic set, which is useful in practical applications, is proposed, triangular neutrosophic numbers and trapezoidal neutrosophic numbers are examined, their definitions and applications are discussed. In this study, a decision making algorithm is developed using neutrosophic set processes and an application is given in fighter aircraft selection as an example of a decision making problem. The estimation of the fighter aircraft selection with the neutrosophic multiple criteria decision analysis method is examined.
Keywords: neutrosophic set, multiple criteria decision making analysis, fighter aircraft selection, MCDMA, neutrosophic numbers
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9332139 An Optimization Model for Natural Gas Supply Chain through a Cost Approach under Uncertainty
Abstract:
Natural gas, as one of the most important sources of energy for many of the industrial and domestic users all over the world, has a complex, huge supply chain which is in need of heavy investments in all the phases of exploration, extraction, production, transportation, storage and distribution. The main purpose of supply chain is to meet customers’ need efficiently and with minimum cost. In this study, with the aim of minimizing economic costs, different levels of natural gas supply chain in the form of a multi-echelon, multi-period fuzzy linear programming have been modeled. In this model, different constraints including constraints on demand satisfaction, capacity, input/output balance and presence/absence of a path have been defined. The obtained results suggest efficiency of the recommended model in optimal allocation and reduction of supply chain costs.
Keywords: Cost Approach, Fuzzy Theory, Linear Programming, Natural Gas Supply Chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25212138 Decision Making with Dempster-Shafer Theory of Evidence Using Geometric Operators
Authors: José M. Merigó, Montserrat Casanovas
Abstract:
We study the problem of decision making with Dempster-Shafer belief structure. We analyze the previous work developed by Yager about using the ordered weighted averaging (OWA) operator in the aggregation of the Dempster-Shafer decision process. We discuss the possibility of aggregating with an ascending order in the OWA operator for the cases where the smallest value is the best result. We suggest the introduction of the ordered weighted geometric (OWG) operator in the Dempster-Shafer framework. In this case, we also discuss the possibility of aggregating with an ascending order and we find that it is completely necessary as the OWG operator cannot aggregate negative numbers. Finally, we give an illustrative example where we can see the different results obtained by using the OWA, the Ascending OWA (AOWA), the OWG and the Ascending OWG (AOWG) operator.
Keywords: Decision making, aggregation operators, Dempster- Shafer theory of evidence, Uncertainty, OWA operator, OWG operator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15292137 Influence of Canola Oil and Lysine Supplementation Diets on Growth Performance and Fatty Acid Composition of Meat in Broiler Chicks
Authors: Ali Kiani, Seyed Davod. Sharifi, Shokoufeh Ghazanfari
Abstract:
A study was conducted to evaluate the effects of diets containing different levels of lysine and canola oil on growth performance and fatty acid composition of meat of broilers chicks. 240-day old Ross broiler chicks were used in a 3×2 factorial arrangement with canola oil (1, 3, and 5%) and lysine (recommended, and 25% more than recommended by Ross broiler manual) in completely randomized design with four replicates and 10 birds per each. The experimental diets were iso-caloric and iso-nitrogenous. Feed intake and body weight gain were recorded at the end of starter (10 d), grower (24 d) and finisher (42 d) periods, and feed conversion ratio was calculated. The results showed that the weight gain of chickens fed diets containing 5% canola oil were greater than those of birds fed on other diets (P<0.05). The dietary lysine had significant effect on feed intake and diets with 25% more than recommended, increased feed intake significantly (P<0.05). The canola oil×lysine interaction effects on performance were not significant. Among all treatment birds, those fed diets containing 5% canola oil had the highest meristic acid and oleic acid content in their meat. Broilers fed diets containing 3 or 5% canola oil possessed the higher content of linolenic acid and lower content of arachidonic acid in their meat (P<0.05). The results of the present experiment indicated that the diets containing canola oil (5%) and lysine at 25% higher than requirement, improve the growth performance, carcass and breast yield of broiler, and increase the accumulation of Omega-3 fatty acids in breast meat.Keywords: Broiler, canola oil, lysine, fatty acid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11602136 Numerical Investigation of the Chilling of Food Products by Air-Mist Spray
Authors: Roy J. Issa
Abstract:
Spray chilling using air-mist nozzles has received much attention in the food processing industry because of the benefits it has shown over forced air convection. These benefits include an increase in the heat transfer coefficient and a reduction in the water loss by the product during cooling. However, few studies have simulated the heat transfer and aerodynamics phenomena of the air-mist chilling process for optimal operating conditions. The study provides insight into the optimal conditions for spray impaction, heat transfer efficiency and control of surface flooding. A computational fluid dynamics model using a two-phase flow composed of water droplets injected with air is developed to simulate the air-mist chilling of food products. The model takes into consideration droplet-to-surface interaction, water-film accumulation and surface runoff. The results of this study lead to a better understanding of the heat transfer enhancement, water conservation, and to a clear direction for the optimal design of air-mist chilling systems that can be used in commercial applications in the food and meat processing industries.Keywords: Droplets impaction efficiency, Droplet size, Heat transfer enhancement factor, Water runoff.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19532135 A New Measurable Definition of Knowledge in New Growth Theory
Authors: Mohammad Ali Molaei
Abstract:
New Growth Theory helps us make sense of the ongoing shift from a resource-based economy to a knowledge-based economy. It underscores the point that the economic processes which create and diffuse new knowledge are critical to shaping the growth of nations, communities and individual firms. In all too many contributions to New (Endogenous) Growth Theory – though not in all – central reference is made to 'a stock of knowledge', a 'stock of ideas', etc., this variable featuring centre-stage in the analysis. Yet it is immediately apparent that this is far from being a crystal clear concept. The difficulty and uncertainty of being able to capture the value associated with knowledge is a real problem. The intent of this paper is introducing new thinking and theorizing about the knowledge and its measurability in new growth theory. Moreover the study aims to synthesize various strain of the literature with a practical bearing on knowledge concept. By contribution of institution framework which is found within NGT, we can indirectly measure the knowledge concept. Institutions matter because they shape the environment for production and employment of new knowledgeKeywords: Institution Framework, Knowledge, New GrowthTheory (NGT)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15442134 Suitability of Black Box Approaches for the Reliability Assessment of Component-Based Software
Authors: Anjushi Verma, Tirthankar Gayen
Abstract:
Although, reliability is an important attribute of quality, especially for mission critical systems, yet, there does not exist any versatile model even today for the reliability assessment of component-based software. The existing Black Box models are found to make various assumptions which may not always be realistic and may be quite contrary to the actual behaviour of software. They focus on observing the manner in which the system behaves without considering the structure of the system, the components composing the system, their interconnections, dependencies, usage frequencies, etc.As a result, the entropy (uncertainty) in assessment using these models is much high.Though, there are some models based on operation profile yet sometimes it becomes extremely difficult to obtain the exact operation profile concerned with a given operation. This paper discusses the drawbacks, deficiencies and limitations of Black Box approaches from the perspective of various authors and finally proposes a conceptual model for the reliability assessment of software.
Keywords: Black Box, faults, failure, software reliability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13842133 Controller Design for Active Suspension System of ¼ Car with Unknown Mass and Time-Delay
Authors: Ali Al-Zughaibi, Huw Davies
Abstract:
The purpose of this paper is to present a modeling and control of a quarter-car active suspension system with unknown mass, unknown time-delay and road disturbance. The objective of designing the controller is to derive a control law to achieve stability of the system and convergence that can considerably improve ride comfort and road disturbance handling. This is accomplished by using Routh-Hurwitz criterion based on defined parameters. Mathematical proof is given to show the ability of the designed controller to ensure the target of design, implementation with the active suspension system and enhancement dispersion oscillation of the system despite these problems. Simulations were also performed to control quarter car suspension, where the results obtained from these simulations verify the validity of the proposed design.Keywords: Active suspension system, disturbance rejection, dynamic uncertainty, time-delay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52302132 Discovery of Fuzzy Censored Production Rules from Large Set of Discovered Fuzzy if then Rules
Authors: Tamanna Siddiqui, M. Afshar Alam
Abstract:
Censored Production Rule is an extension of standard production rule, which is concerned with problems of reasoning with incomplete information, subject to resource constraints and problem of reasoning efficiently with exceptions. A CPR has a form: IF A (Condition) THEN B (Action) UNLESS C (Censor), Where C is the exception condition. Fuzzy CPR are obtained by augmenting ordinary fuzzy production rule “If X is A then Y is B with an exception condition and are written in the form “If X is A then Y is B Unless Z is C. Such rules are employed in situation in which the fuzzy conditional statement “If X is A then Y is B" holds frequently and the exception condition “Z is C" holds rarely. Thus “If X is A then Y is B" part of the fuzzy CPR express important information while the unless part acts only as a switch that changes the polarity of “Y is B" to “Y is not B" when the assertion “Z is C" holds. The proposed approach is an attempt to discover fuzzy censored production rules from set of discovered fuzzy if then rules in the form: A(X) ÔçÆ B(Y) || C(Z).Keywords: Uncertainty Quantification, Fuzzy if then rules, Fuzzy Censored Production Rules, Learning algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14852131 GPI Observer-based Tracking Control and Synchronization of Chaotic Systems
Authors: Dangjun Zhao, Yongji Wang, Lei Liu
Abstract:
Based on general proportional integral (GPI) observers and sliding mode control technique, a robust control method is proposed for the master-slave synchronization of chaotic systems in the presence of parameter uncertainty and with partially measurable output signal. By using GPI observer, the master dynamics are reconstructed by the observations from a measurable output under the differential algebraic framework. Driven by the signals provided by GPI observer, a sliding mode control technique is used for the tracking control and synchronization of the master-slave dynamics. The convincing numerical results reveal the proposed method is effective, and successfully accommodate the system uncertainties, disturbances, and noisy corruptions.
Keywords: GPI observer, sliding mode control, master-slave synchronization, chaotic systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19952130 Use of Vegetation and Geo-Jute in Erosion Control of Slopes in a Sub-Tropical Climate
Authors: Mohammad Shariful Islam, Shamima Nasrin, Md. Shahidul Islam, Farzana Rahman Moury
Abstract:
Protection of slope and embankment from erosion has become an important issue in Bangladesh. The constructions of strong structures require large capital, integrated designing, high maintenance cost. Strong structure methods have negative impact on the environment and sometimes not function for the design period. Plantation of vetiver system along the slopes is an alternative solution. Vetiver not only serves the purpose of slope protection but also adds green environment reducing pollution. Vetiver is available in almost all the districts of Bangladesh. This paper presents the application of vetiver system with geo-jute, for slope protection and erosion control of embankments and slopes. In-situ shear tests have been conducted on vetiver rooted soil system to find the shear strength. The shear strength and effective soil cohesion of vetiver rooted soil matrix are respectively 2.0 times and 2.1 times higher than that of the bared soil. Similar trends have been found in direct shear tests conducted on laboratory reconstituted samples. Field trials have been conducted in road embankment and slope protection with vetiver at different sites. During the time of vetiver root growth the soil protection has been accomplished by geo-jute. As the geo-jute degrades with time, vetiver roots grow and take over the function of geo-jutes. Slope stability analyses showed that vegetation increase the factor of safety significantly.Keywords: Erosion, geo-jute, green technology, vegetation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41642129 On the Robust Stability of Homogeneous Perturbed Large-Scale Bilinear Systems with Time Delays and Constrained Inputs
Authors: Chien-Hua Lee, Cheng-Yi Chen
Abstract:
The stability test problem for homogeneous large-scale perturbed bilinear time-delay systems subjected to constrained inputs is considered in this paper. Both nonlinear uncertainties and interval systems are discussed. By utilizing the Lyapunove equation approach associated with linear algebraic techniques, several delay-independent criteria are presented to guarantee the robust stability of the overall systems. The main feature of the presented results is that although the Lyapunov stability theorem is used, they do not involve any Lyapunov equation which may be unsolvable. Furthermore, it is seen the proposed schemes can be applied to solve the stability analysis problem of large-scale time-delay systems.
Keywords: homogeneous bilinear system, constrained input, time-delay, uncertainty, transient response, decay rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16092128 Medical Image Fusion Based On Redundant Wavelet Transform and Morphological Processing
Authors: P. S. Gomathi, B. Kalaavathi
Abstract:
The process in which the complementary information from multiple images is integrated to provide composite image that contains more information than the original input images is called image fusion. Medical image fusion provides useful information from multimodality medical images that provides additional information to the doctor for diagnosis of diseases in a better way. This paper represents the wavelet based medical image fusion algorithm on different multimodality medical images. In order to fuse the medical images, images are decomposed using Redundant Wavelet Transform (RWT). The high frequency coefficients are convolved with morphological operator followed by the maximum-selection (MS) rule. The low frequency coefficients are processed by MS rule. The reconstructed image is obtained by inverse RWT. The quantitative measures which includes Mean, Standard Deviation, Average Gradient, Spatial frequency, Edge based Similarity Measures are considered for evaluating the fused images. The performance of this proposed method is compared with Pixel averaging, PCA, and DWT fusion methods. When compared with conventional methods, the proposed framework provides better performance for analysis of multimodality medical images.
Keywords: Discrete Wavelet Transform (DWT), Image Fusion, Morphological Processing, Redundant Wavelet Transform (RWT).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21572127 Order Partitioning in Hybrid MTS/MTO Contexts using Fuzzy ANP
Authors: H. Rafiei, M. Rabbani
Abstract:
A novel concept to balance and tradeoff between make-to-stock and make-to-order has been hybrid MTS/MTO production context. One of the most important decisions involved in the hybrid MTS/MTO environment is determining whether a product is manufactured to stock, to order, or hybrid MTS/MTO strategy. In this paper, a model based on analytic network process is developed to tackle the addressed decision. Since the regarded decision deals with the uncertainty and ambiguity of data as well as experts- and managers- linguistic judgments, the proposed model is equipped with fuzzy sets theory. An important attribute of the model is its generality due to diverse decision factors which are elicited from the literature and developed by the authors. Finally, the model is validated by applying to a real case study to reveal how the proposed model can actually be implemented.Keywords: Fuzzy analytic network process, Hybrid make-tostock/ make-to-order, Order partitioning, Production planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21762126 Verification and Proposal of Information Processing Model Using EEG-Based Brain Activity Monitoring
Authors: Toshitaka Higashino, Naoki Wakamiya
Abstract:
Human beings perform a task by perceiving information from outside, recognizing them, and responding them. There have been various attempts to analyze and understand internal processes behind the reaction to a given stimulus by conducting psychological experiments and analysis from multiple perspectives. Among these, we focused on Model Human Processor (MHP). However, it was built based on psychological experiments and thus the relation with brain activity was unclear so far. To verify the validity of the MHP and propose our model from a viewpoint of neuroscience, EEG (Electroencephalography) measurements are performed during experiments in this study. More specifically, first, experiments were conducted where Latin alphabet characters were used as visual stimuli. In addition to response time, ERPs (event-related potentials) such as N100 and P300 were measured by using EEG. By comparing cycle time predicted by the MHP and latency of ERPs, it was found that N100, related to perception of stimuli, appeared at the end of the perceptual processor. Furthermore, by conducting an additional experiment, it was revealed that P300, related to decision making, appeared during the response decision process, not at the end. Second, by experiments using Japanese Hiragana characters, i.e. Japan's own phonetic symbols, those findings were confirmed. Finally, Japanese Kanji characters were used as more complicated visual stimuli. A Kanji character usually has several readings and several meanings. Despite the difference, a reading-related task and a meaning-related task exhibited similar results, meaning that they involved similar information processing processes of the brain. Based on those results, our model was proposed which reflects response time and ERP latency. It consists of three processors: the perception processor from an input of a stimulus to appearance of N100, the cognitive processor from N100 to P300, and the decision-action processor from P300 to response. Using our model, an application system which reflects brain activity can be established.
Keywords: Brain activity, EEG, information processing model, model human processor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6912125 Using Dempster-Shafer Theory in XML Information Retrieval
Authors: F. Raja, M. Rahgozar, F. Oroumchian
Abstract:
XML is a markup language which is becoming the standard format for information representation and data exchange. A major purpose of XML is the explicit representation of the logical structure of a document. Much research has been performed to exploit logical structure of documents in information retrieval in order to precisely extract user information need from large collections of XML documents. In this paper, we describe an XML information retrieval weighting scheme that tries to find the most relevant elements in XML documents in response to a user query. We present this weighting model for information retrieval systems that utilize plausible inferences to infer the relevance of elements in XML documents. We also add to this model the Dempster-Shafer theory of evidence to express the uncertainty in plausible inferences and Dempster-Shafer rule of combination to combine evidences derived from different inferences.Keywords: Dempster-Shafer theory, plausible inferences, XMLinformation retrieval.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15302124 Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data
Authors: J. D. D. Daniel, K. N. Goh, S. M. Yusop
Abstract:
Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.Keywords: Data Transformation Services (DTS), ObjectLinking and Embedding Database (OLEDB), Data Mart, OnlineAnalytical Processing (OLAP), Online Transactional Processing(OLTP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20382123 'Drought Proofing' Australian Cities: Implications for Climate Change Adaptation and Sustainability
Authors: Phoenix Lawhon Isler, John Merson, David Roser
Abstract:
Urban water management in Australia faces increasing pressure to deal with the challenges of droughts, growing population and the climate change uncertainty. Addressing these challenges is an opportunity to incorporate the parallel goals of sustainable water management and climate change adaptation through holistic, non-technical means. This paper presents case studies from Perth and Sydney which show how despite robust adaptation plans and experience, recent efforts to 'drought proof' cities have focused on supply-side measures (i.e. desalination), rather than rethinking how water is used and managing demand. The trend towards desalination as a climate adaptation measure raises questions about the sustainability of urban water futures in Australia.
Keywords: Climate change adaptation, desalination, drought management, sustainable urban water management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22542122 Digital Forensics Compute Cluster: A High Speed Distributed Computing Capability for Digital Forensics
Authors: Daniel Gonzales, Zev Winkelman, Trung Tran, Ricardo Sanchez, Dulani Woods, John Hollywood
Abstract:
We have developed a distributed computing capability, Digital Forensics Compute Cluster (DFORC2) to speed up the ingestion and processing of digital evidence that is resident on computer hard drives. DFORC2 parallelizes evidence ingestion and file processing steps. It can be run on a standalone computer cluster or in the Amazon Web Services (AWS) cloud. When running in a virtualized computing environment, its cluster resources can be dynamically scaled up or down using Kubernetes. DFORC2 is an open source project that uses Autopsy, Apache Spark and Kafka, and other open source software packages. It extends the proven open source digital forensics capabilities of Autopsy to compute clusters and cloud architectures, so digital forensics tasks can be accomplished efficiently by a scalable array of cluster compute nodes. In this paper, we describe DFORC2 and compare it with a standalone version of Autopsy when both are used to process evidence from hard drives of different sizes.Keywords: Cloud computing, cybersecurity, digital forensics, Kafka, Kubernetes, Spark.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16582121 An Intelligent Nondestructive Testing System of Ultrasonic Infrared Thermal Imaging Based on Embedded Linux
Authors: Hao Mi, Ming Yang, Tian-yue Yang
Abstract:
Ultrasonic infrared nondestructive testing is a kind of testing method with high speed, accuracy and localization. However, there are still some problems, such as the detection requires manual real-time field judgment, the methods of result storage and viewing are still primitive. An intelligent non-destructive detection system based on embedded linux is put forward in this paper. The hardware part of the detection system is based on the ARM (Advanced Reduced Instruction Set Computer Machine) core and an embedded linux system is built to realize image processing and defect detection of thermal images. The CLAHE algorithm and the Butterworth filter are used to process the thermal image, and then the boa server and CGI (Common Gateway Interface) technology are used to transmit the test results to the display terminal through the network for real-time monitoring and remote monitoring. The system also liberates labor and eliminates the obstacle of manual judgment. According to the experiment result, the system provides a convenient and quick solution for industrial non-destructive testing.Keywords: Remote monitoring, non-destructive testing, embedded linux system, image processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9672120 An Implementation of MacMahon's Partition Analysis in Ordering the Lower Bound of Processing Elements for the Algorithm of LU Decomposition
Authors: Halil Snopce, Ilir Spahiu, Lavdrim Elmazi
Abstract:
A lot of Scientific and Engineering problems require the solution of large systems of linear equations of the form bAx in an effective manner. LU-Decomposition offers good choices for solving this problem. Our approach is to find the lower bound of processing elements needed for this purpose. Here is used the so called Omega calculus, as a computational method for solving problems via their corresponding Diophantine relation. From the corresponding algorithm is formed a system of linear diophantine equalities using the domain of computation which is given by the set of lattice points inside the polyhedron. Then is run the Mathematica program DiophantineGF.m. This program calculates the generating function from which is possible to find the number of solutions to the system of Diophantine equalities, which in fact gives the lower bound for the number of processors needed for the corresponding algorithm. There is given a mathematical explanation of the problem as well. Keywordsgenerating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equationsand : calculus.
Keywords: generating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equations and calculus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14732119 Predictor Factors for Treatment Failure among Patients on Second Line Antiretroviral Therapy
Authors: Mohd. A. M. Rahim, Yahaya Hassan, Mathumalar L. Fahrni
Abstract:
Second line antiretroviral therapy (ART) regimen is used when patients fail their first line regimen. There are many factors such as non-adherence, drug resistance as well as virological and immunological failure that lead to second line highly active antiretroviral therapy (HAART) regimen treatment failure. This study was aimed at determining predictor factors to treatment failure with second line HAART and analyzing median survival time. An observational, retrospective study was conducted in Sungai Buloh Hospital (HSB) to assess current status of HIV patients treated with second line HAART regimen. Convenience sampling was used and 104 patients were included based on the study’s inclusion and exclusion criteria. Data was collected for six months i.e. from July until December 2013. Data was then analysed using SPSS version 18. Kaplan-Meier and Cox regression analyses were used to measure median survival times and predictor factors for treatment failure. The study population consisted mainly of male subjects, aged 30- 45 years, who were heterosexual, and had HIV infection for less than 6 years. The most common second line HAART regimen given was lopinavir/ritonavir (LPV/r)-based combination. Kaplan-Meier analysis showed that patients on LPV/r demonstrated longer median survival times than patients on indinavir/ritonavir (IDV/r) based combination (p<0.001). The commonest reason for a treatment to fail with second line HAART was non-adherence. Based on Cox regression analysis, other predictor factors for treatment failure with second line HAART regimen were age and mode of HIV transmission.
Keywords: Adherence, antiretroviral therapy, second line, treatment failure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27172118 Using Electrical Impedance Tomography to Control a Robot
Authors: Shayan Rezvanigilkolaei, Shayesteh Vefaghnematollahi
Abstract:
Electrical impedance tomography is a non-invasive medical imaging technique suitable for medical applications. This paper describes an electrical impedance tomography device with the ability to navigate a robotic arm to manipulate a target object. The design of the device includes various hardware and software sections to perform medical imaging and control the robotic arm. In its hardware section an image is formed by 16 electrodes which are located around a container. This image is used to navigate a 3DOF robotic arm to reach the exact location of the target object. The data set to form the impedance imaging is obtained by having repeated current injections and voltage measurements between all electrode pairs. After performing the necessary calculations to obtain the impedance, information is transmitted to the computer. This data is fed and then executed in MATLAB which is interfaced with EIDORS (Electrical Impedance Tomography Reconstruction Software) to reconstruct the image based on the acquired data. In the next step, the coordinates of the center of the target object are calculated by image processing toolbox of MATLAB (IPT). Finally, these coordinates are used to calculate the angles of each joint of the robotic arm. The robotic arm moves to the desired tissue with the user command.Keywords: Electrical impedance tomography, EIT, Surgeon robot, image processing of Electrical impedance tomography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23332117 Orchestra/Percussion Classification Algorithm for United Speech Audio Coding System
Authors: Yueming Wang, Rendong Ying, Sumxin Jiang, Peilin Liu
Abstract:
Unified Speech Audio Coding (USAC), the latest MPEG standardization for unified speech and audio coding, uses a speech/audio classification algorithm to distinguish speech and audio segments of the input signal. The quality of the recovered audio can be increased by well-designed orchestra/percussion classification and subsequent processing. However, owing to the shortcoming of the system, introducing an orchestra/percussion classification and modifying subsequent processing can enormously increase the quality of the recovered audio. This paper proposes an orchestra/percussion classification algorithm for the USAC system which only extracts 3 scales of Mel-Frequency Cepstral Coefficients (MFCCs) rather than traditional 13 scales of MFCCs and use Iterative Dichotomiser 3 (ID3) Decision Tree rather than other complex learning method, thus the proposed algorithm has lower computing complexity than most existing algorithms. Considering that frequent changing of attributes may lead to quality loss of the recovered audio signal, this paper also design a modified subsequent process to help the whole classification system reach an accurate rate as high as 97% which is comparable to classical 99%.
Keywords: ID3 Decision Tree, MFCC, Orchestra/Percussion Classification, USAC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16732116 Network Intrusion Detection Design Using Feature Selection of Soft Computing Paradigms
Authors: T. S. Chou, K. K. Yen, J. Luo
Abstract:
The network traffic data provided for the design of intrusion detection always are large with ineffective information and enclose limited and ambiguous information about users- activities. We study the problems and propose a two phases approach in our intrusion detection design. In the first phase, we develop a correlation-based feature selection algorithm to remove the worthless information from the original high dimensional database. Next, we design an intrusion detection method to solve the problems of uncertainty caused by limited and ambiguous information. In the experiments, we choose six UCI databases and DARPA KDD99 intrusion detection data set as our evaluation tools. Empirical studies indicate that our feature selection algorithm is capable of reducing the size of data set. Our intrusion detection method achieves a better performance than those of participating intrusion detectors.Keywords: Intrusion detection, feature selection, k-nearest neighbors, fuzzy clustering, Dempster-Shafer theory
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19332115 Moderation in Temperature Dependence on Counter Frictional Coefficient and Prevention of Wear of C/C Composites by Synthesizing SiC around Surface and Internal Vacancies
Authors: Noboru Wakamoto, Kiyotaka Obunai, Kazuya Okubo, Toru Fujii
Abstract:
The aim of this study is to moderate the dependence of counter frictional coefficient on temperature between counter surfaces and to reduce the wear of C/C composites at low temperature. To modify the C/C composites, Silica (SiO2) powders were added into phenolic resin for carbon precursor. The preform plate of the precursor of C/C composites was prepared by conventional filament winding method. The C/C composites plates were obtained by carbonizing preform plate at 2200 °C under an argon atmosphere. At that time, the silicon carbides (SiC) were synthesized around the surfaces and the internal vacancies of the C/C composites. The frictional coefficient on the counter surfaces and specific wear volumes of the C/C composites were measured by our developed frictional test machine like pin-on disk type. The XRD indicated that SiC was synthesized in the body of C/C composite fabricated by current method. The results of friction test showed that coefficient of friction of unmodified C/C composites have temperature dependence when the test condition was changed. In contrast, frictional coefficient of the C/C composite modified with SiO2 powders was almost constant at about 0.27 when the temperature condition was changed from Room Temperature (RT) to 300 °C. The specific wear rate decreased from 25×10-6 mm2/N to 0.1×10-6 mm2/N. The observations of the surfaces after friction tests showed that the frictional surface of the modified C/C composites was covered with a film produced by the friction. This study found that synthesizing SiC around surface and internal vacancies of C/C composites was effective to moderate the dependence on the frictional coefficient and reduce to the abrasion of C/C composites.
Keywords: C/C composites, frictional coefficient, SiC, wear.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8292114 Distributional Semantics Approach to Thai Word Sense Disambiguation
Authors: Sunee Pongpinigpinyo, Wanchai Rivepiboon
Abstract:
Word sense disambiguation is one of the most important open problems in natural language processing applications such as information retrieval and machine translation. Many approach strategies can be employed to resolve word ambiguity with a reasonable degree of accuracy. These strategies are: knowledgebased, corpus-based, and hybrid-based. This paper pays attention to the corpus-based strategy that employs an unsupervised learning method for disambiguation. We report our investigation of Latent Semantic Indexing (LSI), an information retrieval technique and unsupervised learning, to the task of Thai noun and verbal word sense disambiguation. The Latent Semantic Indexing has been shown to be efficient and effective for Information Retrieval. For the purposes of this research, we report experiments on two Thai polysemous words, namely /hua4/ and /kep1/ that are used as a representative of Thai nouns and verbs respectively. The results of these experiments demonstrate the effectiveness and indicate the potential of applying vector-based distributional information measures to semantic disambiguation.
Keywords: Distributional semantics, Latent Semantic Indexing, natural language processing, Polysemous words, unsupervisedlearning, Word Sense Disambiguation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18142113 Medical Negligence Disputes in Malaysia: Resolving through Hazards of Litigation or through Community Responsibilities?
Authors: Puteri Nemie Jahn Kassim, Khadijah Mohd Najid
Abstract:
Medical negligence disputes in Malaysia are mainly resolved through litigation by using the tort system. The tort system, being adversarial in nature has subjected parties to litigation hazards such as delay, excessive costs and uncertainty of outcome. The dissatisfaction of the tort system in compensating medically injured victims has created various alternatives to litigation. Amongst them is the implementation of a no-fault compensation system which would allow compensation to be given without the need of proving fault on the medical personnel. Instead, the community now bears the burden of compensating and at the end, promotes collective responsibility. For Malaysia, introducing a no-fault system would provide a tempting solution and may ultimately, achieve justice for the medical injured victims. Nevertheless, such drastic change requires a great deal of consideration to determine the suitability of the system and whether or not it will eventually cater for the needs of the Malaysian population
Keywords: Medical Disputes, Litigation, Malaysia, No-Fault Compensation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47752112 Achieving Performance in an Organization through Marketing Innovation
Authors: Andreea Maier, Diana Nicoară, Dorin Maier, Maria-Mihaela Suărăsan, Alexandra Anastasiu
Abstract:
Innovation is becoming more and more important in modern society. There are a lot of researches on different kinds of innovation but marketing innovation is one kind of innovation that has not been studied frequently before. Marketing innovation is defined as a new way in which companies can market themselves to potential or existing customers. The study shows some key elements for marketing innovation that are worth paying attention to when implementing marketing innovation projects. Examples of such key elements are: paying attention to the neglected market, suitable market segmentatio reliable market information, public relationship, increased customer value, combination of market factors, explore different marketing channels and the use of technology in combination with what? Beside the key elements for marketing innovation, we also present some risks that may occur, such as cost, market uncertainty, information leakage, imitation and overdependence on experience. By proposing a set of indicators to measure marketing innovation, the article offers solutions for marketing innovation implementation so that any organization can achieve optimal results.Keywords: Innovation, marketing innovation, performance in an organization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15892111 Liver Lesion Extraction with Fuzzy Thresholding in Contrast Enhanced Ultrasound Images
Authors: Abder-Rahman Ali, Adélaïde Albouy-Kissi, Manuel Grand-Brochier, Viviane Ladan-Marcus, Christine Hoeffl, Claude Marcus, Antoine Vacavant, Jean-Yves Boire
Abstract:
In this paper, we present a new segmentation approach for focal liver lesions in contrast enhanced ultrasound imaging. This approach, based on a two-cluster Fuzzy C-Means methodology, considers type-II fuzzy sets to handle uncertainty due to the image modality (presence of speckle noise, low contrast, etc.), and to calculate the optimum inter-cluster threshold. Fine boundaries are detected by a local recursive merging of ambiguous pixels. The method has been tested on a representative database. Compared to both Otsu and type-I Fuzzy C-Means techniques, the proposed method significantly reduces the segmentation errors.Keywords: Defuzzification, fuzzy clustering, image segmentation, type-II fuzzy sets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2290