Search results for: Linear Programming Problem.
1215 Multiple-Level Sequential Pattern Discovery from Customer Transaction Databases
Abstract:
Mining sequential patterns from large customer transaction databases has been recognized as a key research topic in database systems. However, the previous works more focused on mining sequential patterns at a single concept level. In this study, we introduced concept hierarchies into this problem and present several algorithms for discovering multiple-level sequential patterns based on the hierarchies. An experiment was conducted to assess the performance of the proposed algorithms. The performances of the algorithms were measured by the relative time spent on completing the mining tasks on two different datasets. The experimental results showed that the performance depends on the characteristics of the datasets and the pre-defined threshold of minimal support for each level of the concept hierarchy. Based on the experimental results, some suggestions were also given for how to select appropriate algorithm for a certain datasets.Keywords: Data Mining, Multiple-Level Sequential Pattern, Concept Hierarchy, Customer Transaction Database.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14541214 Segmentation of Gray Scale Images of Dropwise Condensation on Textured Surfaces
Authors: Helene Martin, Solmaz Boroomandi Barati, Jean-Charles Pinoli, Stephane Valette, Yann Gavet
Abstract:
In the present work we developed an image processing algorithm to measure water droplets characteristics during dropwise condensation on pillared surfaces. The main problem in this process is the similarity between shape and size of water droplets and the pillars. The developed method divides droplets into four main groups based on their size and applies the corresponding algorithm to segment each group. These algorithms generate binary images of droplets based on both their geometrical and intensity properties. The information related to droplets evolution during time including mean radius and drops number per unit area are then extracted from the binary images. The developed image processing algorithm is verified using manual detection and applied to two different sets of images corresponding to two kinds of pillared surfaces.Keywords: Dropwise condensation, textured surface, image processing, watershed.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6911213 Environmental Performance of the United States Energy Sector: A DEA Model with Non-Discretionary Factors and Perfect Object
Authors: Alexander Y. Vaninsky
Abstract:
It is suggested to evaluate environmental performance of energy sector using Data Envelopment Analysis with nondiscretionary factors (DEA-ND) with relative indicators as inputs and outputs. The latter allows for comparison of the objects essentially different in size. Inclusion of non-discretionary factors serves separation of the indicators that are beyond the control of the objects. A virtual perfect object comprised of maximal outputs and minimal inputs was added to the group of actual ones. In this setting, explicit solution of the DEA-ND problem was obtained. Energy sector of the United States was analyzed using suggested approach for the period of 1980 – 2006 with expected values of economic indicators for 2030 used for forming the perfect object. It was obtained that environmental performance has been increasing steadily for the period from 7.7% through 50.0% but still remains well below the prospected levelKeywords: DEA with Non Discretionary Factors, Environmental Performance, Energy Sector, Explicit Solution, Perfect Object.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15291212 The Rank-scaled Mutation Rate for Genetic Algorithms
Authors: Mike Sewell, Jagath Samarabandu, Ranga Rodrigo, Kenneth McIsaac
Abstract:
A novel method of individual level adaptive mutation rate control called the rank-scaled mutation rate for genetic algorithms is introduced. The rank-scaled mutation rate controlled genetic algorithm varies the mutation parameters based on the rank of each individual within the population. Thereby the distribution of the fitness of the papulation is taken into consideration in forming the new mutation rates. The best fit mutate at the lowest rate and the least fit mutate at the highest rate. The complexity of the algorithm is of the order of an individual adaptation scheme and is lower than that of a self-adaptation scheme. The proposed algorithm is tested on two common problems, namely, numerical optimization of a function and the traveling salesman problem. The results show that the proposed algorithm outperforms both the fixed and deterministic mutation rate schemes. It is best suited for problems with several local optimum solutions without a high demand for excessive mutation rates.
Keywords: Genetic algorithms, mutation rate control, adaptive mutation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26691211 A General Framework for Modeling Replicated Real-Time Database
Authors: Hala Abdel hameed, Hazem M. El-Bakry, Torky Sultan
Abstract:
There are many issues that affect modeling and designing real-time databases. One of those issues is maintaining consistency between the actual state of the real-time object of the external environment and its images as reflected by all its replicas distributed over multiple nodes. The need to improve the scalability is another important issue. In this paper, we present a general framework to design a replicated real-time database for small to medium scale systems and maintain all timing constrains. In order to extend the idea for modeling a large scale database, we present a general outline that consider improving the scalability by using an existing static segmentation algorithm applied on the whole database, with the intent to lower the degree of replication, enables segments to have individual degrees of replication with the purpose of avoiding excessive resource usage, which all together contribute in solving the scalability problem for DRTDBS.
Keywords: Database modeling, Distributed database, Real time databases, Replication
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13681210 Synthesis of Highly Sensitive Molecular Imprinted Sensor for Selective Determination of Doxycycline in Honey Samples
Authors: Nadia El Alami El Hassani, Soukaina Motia, Benachir Bouchikhi, Nezha El Bari
Abstract:
Doxycycline (DXy) is a cycline antibiotic, most frequently prescribed to treat bacterial infections in veterinary medicine. However, its broad antimicrobial activity and low cost, lead to an intensive use, which can seriously affect human health. Therefore, its spread in the food products has to be monitored. The scope of this work was to synthetize a sensitive and very selective molecularly imprinted polymer (MIP) for DXy detection in honey samples. Firstly, the synthesis of this biosensor was performed by casting a layer of carboxylate polyvinyl chloride (PVC-COOH) on the working surface of a gold screen-printed electrode (Au-SPE) in order to bind covalently the analyte under mild conditions. Secondly, DXy as a template molecule was bounded to the activated carboxylic groups, and the formation of MIP was performed by a biocompatible polymer by the mean of polyacrylamide matrix. Then, DXy was detected by measurements of differential pulse voltammetry (DPV). A non-imprinted polymer (NIP) prepared in the same conditions and without the use of template molecule was also performed. We have noticed that the elaborated biosensor exhibits a high sensitivity and a linear behavior between the regenerated current and the logarithmic concentrations of DXy from 0.1 pg.mL−1 to 1000 pg.mL−1. This technic was successfully applied to determine DXy residues in honey samples with a limit of detection (LOD) of 0.1 pg.mL−1 and an excellent selectivity when compared to the results of oxytetracycline (OXy) as analogous interfering compound. The proposed method is cheap, sensitive, selective, simple, and is applied successfully to detect DXy in honey with the recoveries of 87% and 95%. Considering these advantages, this system provides a further perspective for food quality control in industrial fields.Keywords: Electrochemical sensor, molecular imprinted polymer, doxycycline, food control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11751209 Graph-Based Text Similarity Measurement by Exploiting Wikipedia as Background Knowledge
Authors: Lu Zhang, Chunping Li, Jun Liu, Hui Wang
Abstract:
Text similarity measurement is a fundamental issue in many textual applications such as document clustering, classification, summarization and question answering. However, prevailing approaches based on Vector Space Model (VSM) more or less suffer from the limitation of Bag of Words (BOW), which ignores the semantic relationship among words. Enriching document representation with background knowledge from Wikipedia is proven to be an effective way to solve this problem, but most existing methods still cannot avoid similar flaws of BOW in a new vector space. In this paper, we propose a novel text similarity measurement which goes beyond VSM and can find semantic affinity between documents. Specifically, it is a unified graph model that exploits Wikipedia as background knowledge and synthesizes both document representation and similarity computation. The experimental results on two different datasets show that our approach significantly improves VSM-based methods in both text clustering and classification.Keywords: Text classification, Text clustering, Text similarity, Wikipedia
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21171208 SimplexIS: Evaluating the Impact of e-Gov Simplification Measures in the Information System Architecure
Authors: Bruno Félix, André Vasconcelos, José Tribolet
Abstract:
Nowadays increasingly the population makes use of Information Technology (IT). As such, in recent year the Portuguese government increased its focus on using the IT for improving people-s life and began to develop a set of measures to enable the modernization of the Public Administration, and so reducing the gap between Public Administration and citizens.Thus the Portuguese Government launched the Simplex Program. However these SIMPLEX eGov measures, which have been implemented over the years, present a serious challenge: how to forecast its impact on existing Information Systems Architecture (ISA). Thus, this research is focus in addressing the problem of automating the evaluation of the actual impact of implementation an eGovSimplification and Modernization measures in the Information Systems Architecture. To realize the evaluation we proposes a Framework, which is supported by some key concepts as: Quality Factors, ISA modeling, Multicriteria Approach, Polarity Profile and Quality MetricsKeywords: Information System Architecture, Evaluation, eGov Simplification measure, Multicriteria Evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14451207 Speed Optimization Model for Reducing Fuel Consumption Based on Shipping Log Data
Authors: Ayudhia P. Gusti, Semin
Abstract:
It is known that total operating cost of a vessel is dominated by the cost of fuel consumption. How to reduce the fuel cost of ship so that the operational costs of fuel can be minimized is the question that arises. As the basis of these kinds of problem, sailing speed determination is an important factor to be considered by a shipping company. Optimal speed determination will give a significant influence on the route and berth schedule of ships, which also affect vessel operating costs. The purpose of this paper is to clarify some important issues about ship speed optimization. Sailing speed, displacement, sailing time, and specific fuel consumption were obtained from shipping log data to be further analyzed for modeling the speed optimization. The presented speed optimization model is expected to affect the fuel consumption and to reduce the cost of fuel consumption.
Keywords: Maritime transportation, reducing fuel, shipping log data, speed optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17371206 Application of Whole Genome Amplification Technique for Genotype Analysis of Bovine Embryos
Authors: S. Moghaddaszadeh-Ahrabi, S. Farajnia, Gh. Rahimi-Mianji, A. Nejati-Javaremi
Abstract:
In recent years, there has been an increasing interest toward the use of bovine genotyped embryos for commercial embryo transfer programs. Biopsy of a few cells in morulla stage is essential for preimplantation genetic diagnosis (PGD). Low amount of DNA have limited performing the several molecular analyses within PGD analyses. Whole genome amplification (WGA) promises to eliminate this problem. We evaluated the possibility and performance of an improved primer extension preamplification (I-PEP) method with a range of starting bovine genomic DNA from 1-8 cells into the WGA reaction. We optimized a short and simple I-PEP (ssI-PEP) procedure (~3h). This optimized WGA method was assessed by 6 loci specific polymerase chain reactions (PCRs), included restriction fragments length polymorphism (RFLP). Optimized WGA procedure possesses enough sensitivity for molecular genetic analyses through the few input cells. This is a new era for generating characterized bovine embryos in preimplantation stage.Keywords: Whole genome amplification (WGA), Genotyping, Bovine, Preimplantation genetic diagnosis (PGD)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16691205 Heterogeneous Attribute Reduction in Noisy System based on a Generalized Neighborhood Rough Sets Model
Authors: Siyuan Jing, Kun She
Abstract:
Neighborhood Rough Sets (NRS) has been proven to be an efficient tool for heterogeneous attribute reduction. However, most of researches are focused on dealing with complete and noiseless data. Factually, most of the information systems are noisy, namely, filled with incomplete data and inconsistent data. In this paper, we introduce a generalized neighborhood rough sets model, called VPTNRS, to deal with the problem of heterogeneous attribute reduction in noisy system. We generalize classical NRS model with tolerance neighborhood relation and the probabilistic theory. Furthermore, we use the neighborhood dependency to evaluate the significance of a subset of heterogeneous attributes and construct a forward greedy algorithm for attribute reduction based on it. Experimental results show that the model is efficient to deal with noisy data.Keywords: attribute reduction, incomplete data, inconsistent data, tolerance neighborhood relation, rough sets
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15881204 Artificial Intelligence-Based Detection of Individuals Suffering from Vestibular Disorder
Authors: D. Hişam, S. İkizoğlu
Abstract:
Identifying the problem behind balance disorder is one of the most interesting topics in medical literature. This study has considerably enhanced the development of artificial intelligence (AI) algorithms applying multiple machine learning (ML) models to sensory data on gait collected from humans to classify between normal people and those suffering from Vestibular System (VS) problems. Although AI is widely utilized as a diagnostic tool in medicine, AI models have not been used to perform feature extraction and identify VS disorders through training on raw data. In this study, three ML models, the Random Forest Classifier (RF), Extreme Gradient Boosting (XGB), and K-Nearest Neighbor (KNN), have been trained to detect VS disorder, and the performance comparison of the algorithms has been made using accuracy, recall, precision, and f1-score. With an accuracy of 95.28 %, Random Forest (RF) Classifier was the most accurate model.
Keywords: Vestibular disorder, machine learning, random forest classifier, k-nearest neighbor, extreme gradient boosting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1691203 Multidimensional Performance Management
Authors: David Wiese
Abstract:
In order to maximize efficiency of an information management platform and to assist in decision making, the collection, storage and analysis of performance-relevant data has become of fundamental importance. This paper addresses the merits and drawbacks provided by the OLAP paradigm for efficiently navigating large volumes of performance measurement data hierarchically. The system managers or database administrators navigate through adequately (re)structured measurement data aiming to detect performance bottlenecks, identify causes for performance problems or assessing the impact of configuration changes on the system and its representative metrics. Of particular importance is finding the root cause of an imminent problem, threatening availability and performance of an information system. Leveraging OLAP techniques, in contrast to traditional static reporting, this is supposed to be accomplished within moderate amount of time and little processing complexity. It is shown how OLAP techniques can help improve understandability and manageability of measurement data and, hence, improve the whole Performance Analysis process.
Keywords: Data Warehousing, OLAP, Multidimensional Navigation, Performance Diagnosis, Performance Management, Performance Tuning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21351202 Human Pose Estimation using Active Shape Models
Authors: Changhyuk Jang, Keechul Jung
Abstract:
Human pose estimation can be executed using Active Shape Models. The existing techniques for applying to human-body research using Active Shape Models, such as human detection, primarily take the form of silhouette of human body. This technique is not able to estimate accurately for human pose to concern two arms and legs, as the silhouette of human body represents the shape as out of round. To solve this problem, we applied the human body model as stick-figure, “skeleton". The skeleton model of human body can give consideration to various shapes of human pose. To obtain effective estimation result, we applied background subtraction and deformed matching algorithm of primary Active Shape Models in the fitting process. The images which were used to make the model were 600 human bodies, and the model has 17 landmark points which indicate body junction and key features of human pose. The maximum iteration for the fitting process was 30 times and the execution time was less than .03 sec.
Keywords: Active shape models, skeleton, pose estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24161201 Legal Basis for Water Resources Management in Brazil: Case Study of the Rio Grande Basin
Authors: Janaína F. Guidolini, Jean P. H. B. Ometto, Angélica Giarolla, Peter M. Toledo, Carlos A. Valera
Abstract:
The water crisis, a major problem of the 21st century, occurs mainly due to poor management. The central issue that should govern the management is the integration of the various aspects that interfere with the use of water resources and their protection, supported by legal basis. A watershed is a unit of water interacting with the physical, biotic, social, economic and cultural variables. The Brazilian law recognized river basin as the territorial management unit. Based on the diagnosis of the current situation of the water resources of the Rio Grande Basin, a discussion informed in the Brazilian legal basis was made to propose measures to fight or mitigate damages and environmental degradation in the Basin. To manage water resources more efficiently, conserve water and optimize their multiple uses, the integration of acquired scientific knowledge and management is essential. Moreover, it is necessary to monitor compliance with environmental legislation.
Keywords: Conservation of soil and water, river basin, sustainability, water governance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9851200 Human Facial Expression Recognition using MANFIS Model
Authors: V. Gomathi, Dr. K. Ramar, A. Santhiyaku Jeevakumar
Abstract:
Facial expression analysis plays a significant role for human computer interaction. Automatic analysis of human facial expression is still a challenging problem with many applications. In this paper, we propose neuro-fuzzy based automatic facial expression recognition system to recognize the human facial expressions like happy, fear, sad, angry, disgust and surprise. Initially facial image is segmented into three regions from which the uniform Local Binary Pattern (LBP) texture features distributions are extracted and represented as a histogram descriptor. The facial expressions are recognized using Multiple Adaptive Neuro Fuzzy Inference System (MANFIS). The proposed system designed and tested with JAFFE face database. The proposed model reports 94.29% of classification accuracy.Keywords: Adaptive neuro-fuzzy inference system, Facialexpression, Local binary pattern, Uniform Histogram
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21031199 Personalization of Web Search Using Web Page Clustering Technique
Authors: Amol Bapuso Rajmane, Pradeep M. Patil, Prakash J. Kulkarni
Abstract:
The Information Retrieval community is facing the problem of effective representation of Web search results. When we organize web search results into clusters it becomes easy to the users to quickly browse through search results. The traditional search engines organize search results into clusters for ambiguous queries, representing each cluster for each meaning of the query. The clusters are obtained according to the topical similarity of the retrieved search results, but it is possible for results to be totally dissimilar and still correspond to the same meaning of the query. People search is also one of the most common tasks on the Web nowadays, but when a particular person’s name is queried the search engines return web pages which are related to different persons who have the same queried name. By placing the burden on the user of disambiguating and collecting pages relevant to a particular person, in this paper, we have developed an approach that clusters web pages based on the association of the web pages to the different people and clusters that are based on generic entity search.
Keywords: Entity resolution, information retrieval, graph based disambiguation, web people search, clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15041198 Reduction of Energy Consumption of Distillation Process by Recovering the Heat from Exit Streams
Authors: Apichit Svang-Ariyaskul, Thanapat Chaireongsirikul, Pawit Tangviroon
Abstract:
Distillation consumes enormous quantity of energy. This work proposed a process to recover the energy from exit streams during the distillation process of three consecutive columns. There are several novel techniques to recover the heat with the distillation system; however, a complex control system is required. This work proposed a simpler technique by exchanging the heat between streams without interrupting the internal distillation process that might cause a serious control problem. The proposed process is executed by using heat exchanger network with pinch analysis to maximize the process heat recovery. The test model is the distillation of butane, pentane, hexane, and heptanes, which is a common mixture in the petroleum refinery. This proposed process saved the energy consumption for hot and cold utilities of 29 and 27%, which is considered significant. Therefore, the recovery of heat from exit streams from distillation process is proved to be effective for energy saving.
Keywords: Distillation, Heat Exchanger, Network Pinch Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32181197 Nuclear Fuel Safety Threshold Determined by Logistic Regression Plus Uncertainty
Authors: D. S. Gomes, A. T. Silva
Abstract:
Analysis of the uncertainty quantification related to nuclear safety margins applied to the nuclear reactor is an important concept to prevent future radioactive accidents. The nuclear fuel performance code may involve the tolerance level determined by traditional deterministic models producing acceptable results at burn cycles under 62 GWd/MTU. The behavior of nuclear fuel can simulate applying a series of material properties under irradiation and physics models to calculate the safety limits. In this study, theoretical predictions of nuclear fuel failure under transient conditions investigate extended radiation cycles at 75 GWd/MTU, considering the behavior of fuel rods in light-water reactors under reactivity accident conditions. The fuel pellet can melt due to the quick increase of reactivity during a transient. Large power excursions in the reactor are the subject of interest bringing to a treatment that is known as the Fuchs-Hansen model. The point kinetic neutron equations show similar characteristics of non-linear differential equations. In this investigation, the multivariate logistic regression is employed to a probabilistic forecast of fuel failure. A comparison of computational simulation and experimental results was acceptable. The experiments carried out use the pre-irradiated fuels rods subjected to a rapid energy pulse which exhibits the same behavior during a nuclear accident. The propagation of uncertainty utilizes the Wilk's formulation. The variables chosen as essential to failure prediction were the fuel burnup, the applied peak power, the pulse width, the oxidation layer thickness, and the cladding type.Keywords: Logistic regression, reactivity-initiated accident, safety margins, uncertainty propagation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10181196 Unsteady Aerodynamics of Multiple Airfoils in Configuration
Authors: Hossain Aziz, Rinku Mukherjee
Abstract:
A potential flow model is used to study the unsteady flow past two airfoils in configuration, each of which is suddenly set into motion. The airfoil bound vortices are modeled using lumped vortex elements and the wake behind the airfoil is modeled by discrete vortices. This consists of solving a steady state flow problem at each time-step where unsteadiness is incorporated through the “zero normal flow on a solid surface" boundary condition at every time instant. Additionally, along with the “zero normal flow on a solid surface" boundary condition Kelvin-s condition is used to compute the strength of the latest wake vortex shed from the trailing edge of the airfoil. Location of the wake vortices is updated at each time-step to get the wake shape at each time instant. Results are presented to show the effect of airfoil-airfoil interaction and airfoil-wake interaction on the aerodynamic characteristics of each airfoil.Keywords: Aerodynamics, Airfoils, Configuration, Unsteady.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20591195 WEMax: Virtual Manned Assembly Line Generation
Authors: Won Kyung Ham, Kang Hoon Cho, Yongho Chung, Sang C. Park
Abstract:
Presented in this paper is a framework of a software ‘WEMax’. The WEMax is invented for analysis and simulation for manned assembly lines to sustain and improve performance of manufacturing systems. In a manufacturing system, performance, such as productivity, is a key of competitiveness for output products. However, the manned assembly lines are difficult to forecast performance, because human labors are not expectable factors by computer simulation models or mathematical models. Existing approaches to performance forecasting of the manned assembly lines are limited to matters of the human itself, such as ergonomic and workload design, and non-human-factor-relevant simulation. Consequently, an approach for the forecasting and improvement of manned assembly line performance is needed to research. As a solution of the current problem, this study proposes a framework that is for generation and simulation of virtual manned assembly lines, and the framework has been implemented as a software.
Keywords: Performance Forecasting, Simulation, Virtual Manned Assembly Line.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18981194 Wood Species Recognition System
Authors: Bremananth R, Nithya B, Saipriya R
Abstract:
The proposed system identifies the species of the wood using the textural features present in its barks. Each species of a wood has its own unique patterns in its bark, which enabled the proposed system to identify it accurately. Automatic wood recognition system has not yet been well established mainly due to lack of research in this area and the difficulty in obtaining the wood database. In our work, a wood recognition system has been designed based on pre-processing techniques, feature extraction and by correlating the features of those wood species for their classification. Texture classification is a problem that has been studied and tested using different methods due to its valuable usage in various pattern recognition problems, such as wood recognition, rock classification. The most popular technique used for the textural classification is Gray-level Co-occurrence Matrices (GLCM). The features from the enhanced images are thus extracted using the GLCM is correlated, which determines the classification between the various wood species. The result thus obtained shows a high rate of recognition accuracy proving that the techniques used in suitable to be implemented for commercial purposes.Keywords: Correlation, Grey Level Co-Occurrence Matrix, ProbabilityDensity Function, Wood Recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24621193 A Study of Liver Checkup in Patients with Hepatitis C in the Region of Batna
Authors: A. Zidani, M. Yahia K. Belhadi, S. Benbia
Abstract:
Hepatitis C is an infectious disease transmitted by blood and due to hepatitis C virus (HCV), which attacks the liver. The infection is characterized by liver inflammation (hepatitis) that is often asymptomatic but can progress to chronic hepatitis and later cirrhosis and liver cancer. Our problem tends to highlight on the one hand the prevalence of infectious disease in the population of the region of Batna and on other hand the biological characteristics of this disease by a screening and a specific diagnosis based on serological tests, liver checkup (measurement of haematological and biochemical parameters). The results showed: The serology of hepatitis C establishes the diagnosis of infection with hepatitis C. In this study and with the serological test, 24 cases of the disease of hepatitis C were found in 1000 suspected cases (7 cases with normal transaminases and 17 cases with elevated transaminases). The prevalence of this disease in this study population was 2.4%. The presence of hepatitis C disrupts liver function including the onset of cytolysis, cholestasis, jaundice, thrombocytopenia, and coagulation disorders.Keywords: Disease hepatitis C, serology, liver checkup
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16521192 A New Heuristic Approach to Solving U-shape Assembly Line Balancing Problems Type-1
Authors: M. Fathi, M. J. Alvarez, V. Rodríguez
Abstract:
Assembly line balancing is a very important issue in mass production systems due to production cost. Although many studies have been done on this topic, but because assembly line balancing problems are so complex they are categorized as NP-hard problems and researchers strongly recommend using heuristic methods. This paper presents a new heuristic approach called the critical task method (CTM) for solving U-shape assembly line balancing problems. The performance of the proposed heuristic method is tested by solving a number of test problems and comparing them with 12 other heuristics available in the literature to confirm the superior performance of the proposed heuristic. Furthermore, to prove the efficiency of the proposed CTM, the objectives are increased to minimize the number of workstation (or equivalently maximize line efficiency), and minimizing the smoothness index. Finally, it is proven that the proposed heuristic is more efficient than the others to solve the U-shape assembly line balancing problem.Keywords: Critical task method, Heuristic, Line balancingproblem, U-shape
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24731191 A Context-Sensitive Algorithm for Media Similarity Search
Authors: Guang-Ho Cha
Abstract:
This paper presents a context-sensitive media similarity search algorithm. One of the central problems regarding media search is the semantic gap between the low-level features computed automatically from media data and the human interpretation of them. This is because the notion of similarity is usually based on high-level abstraction but the low-level features do not sometimes reflect the human perception. Many media search algorithms have used the Minkowski metric to measure similarity between image pairs. However those functions cannot adequately capture the aspects of the characteristics of the human visual system as well as the nonlinear relationships in contextual information given by images in a collection. Our search algorithm tackles this problem by employing a similarity measure and a ranking strategy that reflect the nonlinearity of human perception and contextual information in a dataset. Similarity search in an image database based on this contextual information shows encouraging experimental results.
Keywords: Context-sensitive search, image search, media search, similarity ranking, similarity search.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6411190 Generalized Rough Sets Applied to Graphs Related to Urban Problems
Authors: Mihai Rebenciuc, Simona Mihaela Bibic
Abstract:
Branch of modern mathematics, graphs represent instruments for optimization and solving practical applications in various fields such as economic networks, engineering, network optimization, the geometry of social action, generally, complex systems including contemporary urban problems (path or transport efficiencies, biourbanism, & c.). In this paper is studied the interconnection of some urban network, which can lead to a simulation problem of a digraph through another digraph. The simulation is made univoc or more general multivoc. The concepts of fragment and atom are very useful in the study of connectivity in the digraph that is simulation - including an alternative evaluation of k- connectivity. Rough set approach in (bi)digraph which is proposed in premier in this paper contribute to improved significantly the evaluation of k-connectivity. This rough set approach is based on generalized rough sets - basic facts are presented in this paper.Keywords: (Bi)digraphs, rough set theory, systems of interacting agents, complex systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11901189 Formal Specification of Web Services Applications for Digital Reference Services of Library Information System
Authors: Zainab M. Musa, Nordin M. A. Rahman, Julaily A. Jusoh
Abstract:
Digital reference service is when a traditional library reference service is provided electronically. In most cases users do not get full satisfaction from using digital reference service due to variety of reasons. This paper discusses the formal specification of web services applications for digital reference services (WSDRS). WSDRS is an informal model that claims to reduce the problems of digital reference services in libraries. It uses web services technology to provide efficient digital way of satisfying users’ need in the reference section of libraries. Informal model is in natural language which is inconsistent and ambiguous that may cause difficulties to the developers of the system. In order to solve this problem we decided to convert the informal specifications into formal specifications. This is supposed to reduce the overall development time and cost. We use Z language to develop the formal model and verify it with Z/EVES theorem prover tool.Keywords: Formal, specifications, web services, digital reference services.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16151188 Totally Integrated Smart Energy System through Data Acquisition via Remote Location
Authors: Muhammad Tahir Qadri, M. Irfan Anis, M. Nawaz Irshad Khan
Abstract:
This paper discusses the approach of real-time controlling of the energy management system using the data acquisition tool of LabVIEW. The main idea of this inspiration was to interface the Station (PC) with the system and publish the data on internet using LabVIEW. In this venture, controlling and switching of 3 phase AC loads are effectively and efficiently done. The phases are also sensed through devices. In case of any failure the attached generator starts functioning automatically. The computer sends command to the system and system respond to the request. The modern feature is to access and control the system world-wide using world wide web (internet). This controlling can be done at any time from anywhere to effectively use the energy especially in developing countries where energy management is a big problem. In this system totally integrated devices are used to operate via remote location.Keywords: VI-server, Remote Access, Telemetry, Data Acquisition, web server.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18791187 Development of Web-Based Remote Desktop to Provide Adaptive User Interfaces in Cloud Platform
Authors: Shuen-Tai Wang, Hsi-Ya Chang
Abstract:
Cloud virtualization technologies are becoming more and more prevalent, cloud users usually encounter the problem of how to access to the virtualized remote desktops easily over the web without requiring the installation of special clients. To resolve this issue, we took advantage of the HTML5 technology and developed web-based remote desktop. It permits users to access the terminal which running in our cloud platform from anywhere. We implemented a sketch of web interface following the cloud computing concept that seeks to enable collaboration and communication among users for high performance computing. Given the development of remote desktop virtualization, it allows to shift the user’s desktop from the traditional PC environment to the cloud platform, which is stored on a remote virtual machine rather than locally. This proposed effort has the potential to positively provide an efficient, resilience and elastic environment for online cloud service. This is also made possible by the low administrative costs as well as relatively inexpensive end-user terminals and reduced energy expenses.
Keywords: Virtualization, Remote Desktop, HTML5, Cloud Computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32531186 How to Improve Teaching and Learning Strategies through Educational Research: An Experience of Peer Observation in Legal Education
Authors: L. Mortari, A. Bevilacqua, R. Silva
Abstract:
The experience presented in this paper aims to understand how educational research can support the introduction and optimization of teaching innovations in legal education. In this increasingly complex context, a strong need to introduce paths aimed at acquiring not only professional knowledge and skills but also reflective, critical and problem-solving skills emerges. Through a peer observation intertwined with an analysis of discursive practices, researchers and the teacher worked together through a process of participatory and transformative accompaniment whose objective was to promote the active participation and engagement of students in learning processes, an element indispensable to work in the more specific direction of strengthening key competences. This reflective faculty development path led the teacher to activate metacognitive processes, becoming thus aware of the strengths and areas of improvement of his teaching innovation.
Keywords: Discursive analysis, faculty development, legal education, peer observation, teaching innovation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 356