Search results for: computer applications
886 Volterra Filter for Color Image Segmentation
Authors: M. B. Meenavathi, K. Rajesh
Abstract:
Color image segmentation plays an important role in computer vision and image processing areas. In this paper, the features of Volterra filter are utilized for color image segmentation. The discrete Volterra filter exhibits both linear and nonlinear characteristics. The linear part smoothes the image features in uniform gray zones and is used for getting a gross representation of objects of interest. The nonlinear term compensates for the blurring due to the linear term and preserves the edges which are mainly used to distinguish the various objects. The truncated quadratic Volterra filters are mainly used for edge preserving along with Gaussian noise cancellation. In our approach, the segmentation is based on K-means clustering algorithm in HSI space. Both the hue and the intensity components are fully utilized. For hue clustering, the special cyclic property of the hue component is taken into consideration. The experimental results show that the proposed technique segments the color image while preserving significant features and removing noise effects.Keywords: Color image segmentation, HSI space, K–means clustering, Volterra filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1857885 Apoptosis Inspired Intrusion Detection System
Authors: R. Sridevi, G. Jagajothi
Abstract:
Artificial Immune Systems (AIS), inspired by the human immune system, are algorithms and mechanisms which are self-adaptive and self-learning classifiers capable of recognizing and classifying by learning, long-term memory and association. Unlike other human system inspired techniques like genetic algorithms and neural networks, AIS includes a range of algorithms modeling on different immune mechanism of the body. In this paper, a mechanism of a human immune system based on apoptosis is adopted to build an Intrusion Detection System (IDS) to protect computer networks. Features are selected from network traffic using Fisher Score. Based on the selected features, the record/connection is classified as either an attack or normal traffic by the proposed methodology. Simulation results demonstrates that the proposed AIS based on apoptosis performs better than existing AIS for intrusion detection.
Keywords: Apoptosis, Artificial Immune System (AIS), Fisher Score, KDD dataset, Network intrusion detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2191884 GPS INS Integration Application in Flight Management System
Authors: Othman Maklouf, Abdurazag Ghila, Saleh Gashoot, Ahmed Abdulla
Abstract:
Flight management system (FMS) is a specialized computer system that automates a wide variety of in-flight tasks, reducing the workload on the flight crew to the point that modern aircraft no longer carry flight engineers or navigators. The primary function of FMS is to perform the in-flight management of the flight plan using various sensors (such as GPS and INS often backed up by radio navigation) to determine the aircraft's position. From the cockpit FMS is normally controlled through a Control Display Unit (CDU) which incorporates a small screen and keyboard or touch screen. This paper investigates the performance of GPS/ INS integration techniques in which the data fusion process is done using Kalman filtering. This will include the importance of sensors calibration as well as the alignment of the strap down inertial navigation system. The limitations of the inertial navigation systems are investigated in order to understand why INS sometimes is integrated with other navigation aids and not just operating in standalone mode. Finally, both the loosely coupled and tightly coupled configurations are analyzed for several types of situations and operational conditions.Keywords: GPS, INS, Kalman Filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2494883 MGAUM—Towards a Mobile Government Adoption and Utilization Model: The Case of Saudi Arabia
Authors: Mohammed Alonazi, Natalia Beloff, Martin White
Abstract:
This paper presents a proposal for a mobile government adoption and utilization model (MGAUM), which is a framework designed to increase the adoption rate of m-government services in Saudi Arabia. Recent advances in mobile technologies such are Mobile compatibilities, The development of wireless communication, mobile applications and devices are enabling governments to deliver services in new ways to citizens more efficiently and economically. In the last decade, many governments around the globe are utilizing these advances effectively to develop their next generation of e-government services. However, a low adoption rate of m-government services by citizens is a common problem in Arabian countries, including Saudi Arabia. Yet, to our knowledge, very little research has been conducted focused on understanding the factors that influence citizen adoption of these m-government services in this part of the world. A set of social, cultural and technological factors have been identified in the literature, which has led to the formulation of associated research questions and hypotheses. These hypotheses will be tested on Saudi citizens using questionnaires and interview methods based around the technology acceptance model. A key objective of the MGAUM framework is to investigate and understand Saudi citizens perception towards adoption and utilization of m-government services.Keywords: E-Government, M-Government, citizen services quality, technology acceptance model, Saudi Arabia, adoption framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1644882 Fault and Theft Recognition Using Toro Dial Sensor in Programmable Current Relay for Feeder Security
Authors: R. Kamalakannan, N. Ravi Kumar
Abstract:
Feeder protection is important in transmission and distribution side because if any fault occurs in any feeder or transformer, man power is needed to identify the problem and it will take more time. In the existing system, directional overcurrent elements with load further secured by a load encroachment function can be used to provide necessary security and sensitivity for faults on remote points in a circuit. It is validated only in renewable plant collector circuit protection applications over a wide range of operating conditions. In this method, the directional overcurrent feeder protection is developed by using monitoring of feeder section through internet. In this web based monitoring, the fault and power theft are identified by using Toro dial sensor and its information is received by SCADA (Supervisory Control and Data Acquisition) and controlled by ARM microcontroller. This web based monitoring is also used to monitor the feeder management, directional current detection, demand side management, overload fault. This monitoring system is capable of monitoring the distribution feeder over a large area depending upon the cost. It is also used to reduce the power theft, time and man power. The simulation is done by MATLAB software.
Keywords: Current sensor, distribution feeder protection, directional overcurrent, power theft, protective relay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 793881 Formal Verification of a Multicast Protocol in Mobile Networks
Authors: M. Matash Borujerdi, S.M. Mirzababaei
Abstract:
As computer network technology becomes increasingly complex, it becomes necessary to place greater requirements on the validity of developing standards and the resulting technology. Communication networks are based on large amounts of protocols. The validity of these protocols have to be proved either individually or in an integral fashion. One strategy for achieving this is to apply the growing field of formal methods. Formal methods research defines systems in high order logic so that automated reasoning can be applied for verification. In this research we represent and implement a formerly announced multicast protocol in Prolog language so that certain properties of the protocol can be verified. It is shown that by using this approach some minor faults in the protocol were found and repaired. Describing the protocol as facts and rules also have other benefits i.e. leads to a process-able knowledge. This knowledge can be transferred as ontology between systems in KQML format. Since the Prolog language can increase its knowledge base every time, this method can also be used to learn an intelligent network.Keywords: Formal methods, MobiCast, Mobile Network, Multicast.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1380880 An Intelligent System Framework for Generating Activity List of a Project Using WBS Mind map and Semantic Network
Authors: H. Iranmanesh, M. Madadi
Abstract:
Work Breakdown Structure (WBS) is one of the most vital planning processes of the project management since it is considered to be the fundamental of other processes like scheduling, controlling, assigning responsibilities, etc. In fact WBS or activity list is the heart of a project and omission of a simple task can lead to an irrecoverable result. There are some tools in order to generate a project WBS. One of the most powerful tools is mind mapping which is the basis of this article. Mind map is a method for thinking together and helps a project manager to stimulate the mind of project team members to generate project WBS. Here we try to generate a WBS of a sample project involving with the building construction using the aid of mind map and the artificial intelligence (AI) programming language. Since mind map structure can not represent data in a computerized way, we convert it to a semantic network which can be used by the computer and then extract the final WBS from the semantic network by the prolog programming language. This method will result a comprehensive WBS and decrease the probability of omitting project tasks.Keywords: Expert System, Mind map, Semantic network, Work breakdown structure,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2608879 Contamination in Industrial Areas and Environmental Management in Latvia
Authors: Juris Burlakovs, Maris Klavins, Raimonds Ernsteins, Armands Ruskulis
Abstract:
Environmental contamination is a common problem in ex-industrial and industrial sites. This article gives a brief description of general applied environmental investigation methodologies and possible remediation applications in Latvia. Most of contaminated areas are situated in former and active industrial, military areas and ports. Industrial and logistic activities very often have been with great impact for more than hundred years thus the contamination level with heavy metals, hydrocarbons, pesticides, persistent organic pollutants is high and is threatening health and environment in general. 242 territories now are numbered as contaminated and fixed in the National Register of contaminated territories in Latvia. Research and remediation of contamination in densely populated areas are of important environmental policy domain. Four different investigation case studies of contaminated areas are given describing the history of use, environmental quality assessment as well as planned environmental management actions. All four case study locations are situated in Riga - the capital of the Republic of Latvia. The aim of this paper is to analyze the situation and problems with management of contaminated areas in Latvia, give description of field research methods and recommendations for remediation industry based on scientific data and innovations.
Keywords: Remediation technology, environmental quality assessment, heavy metals, hydrocarbon contamination, environmental management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2301878 Efficient Copy-Move Forgery Detection for Digital Images
Authors: Somayeh Sadeghi, Hamid A. Jalab, Sajjad Dadkhah
Abstract:
Due to availability of powerful image processing software and improvement of human computer knowledge, it becomes easy to tamper images. Manipulation of digital images in different fields like court of law and medical imaging create a serious problem nowadays. Copy-move forgery is one of the most common types of forgery which copies some part of the image and pastes it to another part of the same image to cover an important scene. In this paper, a copy-move forgery detection method proposed based on Fourier transform to detect forgeries. Firstly, image is divided to same size blocks and Fourier transform is performed on each block. Similarity in the Fourier transform between different blocks provides an indication of the copy-move operation. The experimental results prove that the proposed method works on reasonable time and works well for gray scale and colour images. Computational complexity reduced by using Fourier transform in this method.Keywords: Copy-Move forgery, Digital Forensics, Image Forgery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2785877 An Economical Operation Analysis Optimization Model for Heavy Equipment Selection
Authors: A. Jrade, N. Markiz, N. Albelwi
Abstract:
Optimizing equipment selection in heavy earthwork operations is a critical key in the success of any construction project. The objective of this research incentive was geared towards developing a computer model to assist contractors and construction managers in estimating the cost of heavy earthwork operations. Economical operation analysis was conducted for an equipment fleet taking into consideration the owning and operating costs involved in earthwork operations. The model is being developed in a Microsoft environment and is capable of being integrated with other estimating and optimization models. In this study, Caterpillar® Performance Handbook [5] was the main resource used to obtain specifications of selected equipment. The implementation of the model shall give optimum selection of equipment fleet not only based on cost effectiveness but also in terms of versatility. To validate the model, a case study of an actual dam construction project was selected to quantify its degree of accuracy.Keywords: Operation analysis, optimization model, equipment economics, equipment selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4258876 CFD Simulations to Validate Two and Three Phase Up-flow in Bubble Columns
Authors: Shyam Kumar, Nannuri Srinivasulu, Ashok Khanna
Abstract:
Bubble columns have a variety of applications in absorption, bio-reactions, catalytic slurry reactions, and coal liquefaction; because they are simple to operate, provide good heat and mass transfer, having less operational cost. The use of Computational Fluid Dynamics (CFD) for bubble column becomes important, since it can describe the fluid hydrodynamics on both local and global scale. Euler- Euler two-phase fluid model has been used to simulate two-phase (air and water) transient up-flow in bubble column (15cm diameter) using FLUENT6.3. These simulations and experiments were operated over a range of superficial gas velocities in the bubbly flow and churn turbulent regime (1 to16 cm/s) at ambient conditions. Liquid velocity was varied from 0 to 16cm/s. The turbulence in the liquid phase is described using the standard k-ε model. The interactions between the two phases are described through drag coefficient formulations (Schiller Neumann). The objectives are to validate CFD simulations with experimental data, and to obtain grid-independent numerical solutions. Quantitatively good agreements are obtained between experimental data for hold-up and simulation values. Axial liquid velocity profiles and gas holdup profiles were also obtained for the simulation.Keywords: Bubble column, Computational fluid dynamics, Gas holdup profile, k-ε model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2719875 Production of Composite Materials by Mixing Chromium-Rich Ash and Soda-Lime Glass Powder: Mechanical Properties and Microstructure
Authors: Savvas Varitis, Panagiotis Kavouras, George Vourlias, Eleni Pavlidou, Theodoros Karakostas, Philomela Komninou
Abstract:
A chromium-loaded ash originating from incineration of tannery sludge under anoxic conditions was mixed with low grade soda-lime glass powder coming from commercial glass bottles. The relative weight proportions of ash over glass powder tested were 30/70, 40/60 and 50/50. The solid mixtures, formed in green state compacts, were sintered at the temperature range of 800o C up to 1200o C. The resulting products were characterized by X-ray diffraction (XRD), scanning electron microscopy (SEM), energy dispersive X-ray spectrometry (EDXS) and micro-indentation. The above methods were employed to characterize the various phases, microstructure and hardness of the produced materials. Thermal treatment at 800o C and 1000o C produced opaque ceramic products composed of a variety of chromium-containing and chromium-free crystalline phases. Thermal treatment at 1200o C gave rise to composite products, where only chromium-containing crystalline phases were detected. Hardness results suggest that specific products are serious candidates for structural applications.
Keywords: Chromium-rich tannery residues, glass-ceramic materials, mechanical properties, microstructure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1866874 Comparative Study of Transformed and Concealed Data in Experimental Designs and Analyses
Authors: K. Chinda, P. Luangpaiboon
Abstract:
This paper presents the comparative study of coded data methods for finding the benefit of concealing the natural data which is the mercantile secret. Influential parameters of the number of replicates (rep), treatment effects (τ) and standard deviation (σ) against the efficiency of each transformation method are investigated. The experimental data are generated via computer simulations under the specified condition of the process with the completely randomized design (CRD). Three ways of data transformation consist of Box-Cox, arcsine and logit methods. The difference values of F statistic between coded data and natural data (Fc-Fn) and hypothesis testing results were determined. The experimental results indicate that the Box-Cox results are significantly different from natural data in cases of smaller levels of replicates and seem to be improper when the parameter of minus lambda has been assigned. On the other hand, arcsine and logit transformations are more robust and obviously, provide more precise numerical results. In addition, the alternate ways to select the lambda in the power transformation are also offered to achieve much more appropriate outcomes.Keywords: Experimental Designs, Box-Cox, Arcsine, Logit Transformations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1622873 Node Pair Selection Scheme in Relay-Aided Communication Based On Stable Marriage Problem
Authors: Tetsuki Taniguchi, Yoshio Karasawa
Abstract:
This paper describes a node pair selection scheme in relay-aided multiple source multiple destination communication system based on stable marriage problem. A general case is assumed in which all of source, relay and destination nodes are equipped with multiantenna and carry out multistream transmission. Based on several metrics introduced from inter-node channel condition, the preference order is determined about all source-relay and relay-destination relations, and then the node pairs are determined using Gale-Shapley algorithm. The computer simulations show that the effectiveness of node pair selection is larger in multihop communication. Some additional aspects which are different from relay-less case are also investigated.
Keywords: Relay, multiple input multiple output (MIMO), multiuser, amplify and forward, stable marriage problem, Gale-Shapley algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1935872 A Modified Run Length Coding Technique for Test Data Compression Based on Multi-Level Selective Huffman Coding
Authors: C. Kalamani, K. Paramasivam
Abstract:
Test data compression is an efficient method for reducing the test application cost. The problem of reducing test data has been addressed by researchers in three different aspects: Test Data Compression, Built-in-Self-Test (BIST) and Test set compaction. The latter two methods are capable of enhancing fault coverage with cost of hardware overhead. The drawback of the conventional methods is that they are capable of reducing the test storage and test power but when test data have redundant length of runs, no additional compression method is followed. This paper presents a modified Run Length Coding (RLC) technique with Multilevel Selective Huffman Coding (MLSHC) technique to reduce test data volume, test pattern delivery time and power dissipation in scan test applications where redundant length of runs is encountered then the preceding run symbol is replaced with tiny codeword. Experimental results show that the presented method not only improves the test data compression but also reduces the overall test data volume compared to recent schemes. Experiments for the six largest ISCAS-98 benchmarks show that our method outperforms most known techniques.
Keywords: Modified run length coding, multilevel selective Huffman coding, built-in-self-test modified selective Huffman coding, automatic test equipment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1274871 On the Exact Solution of Non-Uniform Torsion for Beams with Asymmetric Cross-Section
Authors: A.Campanile, M. Mandarino, V. Piscopo
Abstract:
This paper deals with the problem of non-uniform torsion in thin-walled elastic beams with asymmetric cross-section, removing the basic concept of a fixed center of twist, necessary in the Vlasov-s and Benscoter-s theories to obtain a warping stress field equivalent to zero. In this new torsion/flexure theory, despite of the classical ones, the warping function will punctually satisfy the first indefinite equilibrium equation along the beam axis and it wont- be necessary to introduce the classical congruence condition, to take into account the effect of the beam restraints. The solution, based on the Fourier development of the displacement field, is obtained assuming that the applied external torque is constant along the beam axis and on both beam ends the unit twist angle and the warping axial displacement functions are totally restrained. Finally, in order to verify the feasibility of the proposed method and to compare it with the classical theories, two applications are carried out. The first one, relative to an open profile, is necessary to test the numerical method adopted to find the solution; the second one, instead, is relative to a simplified containership section, considered as full restrained in correspondence of two adjacent transverse bulkheads.Keywords: Non-uniform torsion, Asymmetric cross-section, Fourier series, Helmholtz equation, FE method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942870 A Study on the Developing Method of the BIM (Building Information Modeling) Software Based On Cloud Computing Environment
Authors: Byung-Kon Kim
Abstract:
According as the Architecture, Engineering and Construction (AEC) Industry projects have grown more complex and larger, the number of utilization of BIM for 3D design and simulation is increasing significantly. Therefore, typical applications of BIM such as clash detection and alternative measures based on 3-dimenstional planning are expanded to process management, cost and quantity management, structural analysis, check for regulation, and various domains for virtual design and construction. Presently, commercial BIM software is operated on single-user environment, so initial cost is so high and the investment may be wasted frequently. Cloud computing that is a next-generation internet technology enables simple internet devices (such as PC, Tablet, Smart phone etc) to use services and resources of BIM software. In this paper, we suggested developing method of the BIM software based on cloud computing environment in order to expand utilization of BIM and reduce cost of BIM software. First, for the benchmarking, we surveyed successful case of BIM and cloud computing. And we analyzed needs and opportunities of BIM and cloud computing in AEC Industry. Finally, we suggested main functions of BIM software based on cloud computing environment and developed a simple prototype of cloud computing BIM software for basic BIM model viewing.
Keywords: Construction IT, BIM(Building Information Modeling), Cloud Computing, BIM Service Based Cloud Computing, Viewer Based BIM Server, 3D Design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4101869 Hull Separation Optimization of Catamaran Unmanned Surface Vehicle Powered with Hydrogen Fuel Cell
Authors: Seok-In Sohn, Dae-Hwan Park, Yeon-Seung Lee, Il-Kwon Oh
Abstract:
This paper presents an optimization of the hull separation, i.e. transverse clearance. The main objective is to identify the feasible speed ranges and find the optimum transverse clearance considering the minimum wave-making resistance. The dimensions and the weight of hardware systems installed in the catamaran structured fuel cell powered USV (Unmanned Surface Vehicle) were considered as constraints. As the CAE (Computer Aided Engineering) platform FRIENDSHIP-Framework was used. The hull surface modeling, DoE (Design of Experiment), Tangent search optimization, tool integration and the process automation were performed by FRIENDSHIP-Framework. The hydrodynamic result was evaluated by XPAN the potential solver of SHIPFLOW.Keywords: Full parametric modeling, Hull Separation, Wave-making resistance, Design Of Experiment, Tangent search method
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2960868 Image Indexing Using a Color Similarity Metric based on the Human Visual System
Authors: Angelo Nodari, Ignazio Gallo
Abstract:
The novelty proposed in this study is twofold and consists in the developing of a new color similarity metric based on the human visual system and a new color indexing based on a textual approach. The new color similarity metric proposed is based on the color perception of the human visual system. Consequently the results returned by the indexing system can fulfill as much as possibile the user expectations. We developed a web application to collect the users judgments about the similarities between colors, whose results are used to estimate the metric proposed in this study. In order to index the image's colors, we used a text indexing engine to facilitate the integration of visual features in a database of text documents. The textual signature is build by weighting the image's colors in according to their occurrence in the image. The use of a textual indexing engine, provide us a simple, fast and robust solution to index images. A typical usage of the system proposed in this study, is the development of applications whose data type is both visual and textual. In order to evaluate the proposed method we chose a price comparison engine as a case of study, collecting a series of commercial offers containing the textual description and the image representing a specific commercial offer.
Keywords: Color Extraction, Content-Based Image Retrieval, Indexing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3027867 Neural Network based Texture Analysis of Liver Tumor from Computed Tomography Images
Authors: K.Mala, V.Sadasivam, S.Alagappan
Abstract:
Advances in clinical medical imaging have brought about the routine production of vast numbers of medical images that need to be analyzed. As a result an enormous amount of computer vision research effort has been targeted at achieving automated medical image analysis. Computed Tomography (CT) is highly accurate for diagnosing liver tumors. This study aimed to evaluate the potential role of the wavelet and the neural network in the differential diagnosis of liver tumors in CT images. The tumors considered in this study are hepatocellular carcinoma, cholangio carcinoma, hemangeoma and hepatoadenoma. Each suspicious tumor region was automatically extracted from the CT abdominal images and the textural information obtained was used to train the Probabilistic Neural Network (PNN) to classify the tumors. Results obtained were evaluated with the help of radiologists. The system differentiates the tumor with relatively high accuracy and is therefore clinically useful.
Keywords: Fuzzy c means clustering, texture analysis, probabilistic neural network, LVQ neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2988866 The Internet of Healthcare Things: A European Perspective and a Review of Ethical Concerns
Authors: M. Emmanouilidou
Abstract:
The Internet of Things (IoT) is a disruptive technological paradigm that is at the center of the digital evolution by integrating physical and virtual worlds leading to the creation of extended interconnected ecosystems that are characterized as smart environments. The concept of the IoT has a broad range of applications in different industries including the healthcare sector. The Internet of Healthcare Things (IoHT), a branch of the IoT, is expected to bring promising benefits to all involved stakeholders and accelerate the revolution of the healthcare sector through a transition towards preventive and personalized medicine. The socio-economic challenges that the healthcare sector is facing further emphasize the need for a radical transformation of healthcare systems in both developed and developing countries with the role of pervasive technological innovations, such as IoHT, recognized as key to counteract the relevant challenges. Besides the number of potential opportunities that IoHT presents, there are fundamental ethical concerns that need to be considered and addressed in relation to the application of IoHT. This paper contributes to the discussion of the emerging topic of IoHT by providing an overview of the role and potential of IoHT, highlighting the characteristics of the current and future healthcare landscape, reporting on the up-to-date status of IoHT in Europe and reflecting upon existing research in the ethics of IoHT by incorporating additional ethical dimensions that have been ignored which can provide pathways for future research in the field.Keywords: Ethics, Europe, healthcare, internet of things.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1759865 The Energy Impacts of Using Top-Light Daylighting Systems for Academic Buildings in Tropical Climate
Authors: M. S. Alrubaih, M. F. M. Zain, N. L. N. Ibrahim, M.A. Alghoul, K. I. Ben Sauod
Abstract:
Careful design and selection of daylighting systems can greatly help in reducing not only artificial lighting use, but also decrease cooling energy consumption and, therefore, potential for downsizing air-conditioning systems. This paper aims to evaluate the energy performance of two types of top-light daylighting systems due to the integration of daylight together with artificial lighting in an existing examinaton hall in University Kebangsaan Malaysia, based on a hot and humid climate. Computer simulation models have been created for building case study (base case) and the two types of toplight daylighting designs for building energy performance evaluation using the VisualDOE 4.0 building energy simulation program. The finding revealed that daylighting through top-light systems is a very beneficial design strategy in reducing annual lighting energy consumption and the overall total annual energy consumption.
Keywords: Academic buildings, Daylighting, Top-lighting, Energy savings, Tropical Climate
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1946864 In Cognitive Radio the Analysis of Bit-Error- Rate (BER) by using PSO Algorithm
Authors: Shrikrishan Yadav, Akhilesh Saini, Krishna Chandra Roy
Abstract:
The electromagnetic spectrum is a natural resource and hence well-organized usage of the limited natural resources is the necessities for better communication. The present static frequency allocation schemes cannot accommodate demands of the rapidly increasing number of higher data rate services. Therefore, dynamic usage of the spectrum must be distinguished from the static usage to increase the availability of frequency spectrum. Cognitive radio is not a single piece of apparatus but it is a technology that can incorporate components spread across a network. It offers great promise for improving system efficiency, spectrum utilization, more effective applications, reduction in interference and reduced complexity of usage for users. Cognitive radio is aware of its environmental, internal state, and location, and autonomously adjusts its operations to achieve designed objectives. It first senses its spectral environment over a wide frequency band, and then adapts the parameters to maximize spectrum efficiency with high performance. This paper only focuses on the analysis of Bit-Error-Rate in cognitive radio by using Particle Swarm Optimization Algorithm. It is theoretically as well as practically analyzed and interpreted in the sense of advantages and drawbacks and how BER affects the efficiency and performance of the communication system.Keywords: BER, Cognitive Radio, Environmental Parameters, PSO, Radio spectrum, Transmission Parameters
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2155863 The Role of Velocity Map Quality in Estimation of Intravascular Pressure Distribution
Authors: Ali Pashaee, Parisa Shooshtari, Gholamreza Atae, Nasser Fatouraee
Abstract:
Phase-Contrast MR imaging methods are widely used for measurement of blood flow velocity components. Also there are some other tools such as CT and Ultrasound for velocity map detection in intravascular studies. These data are used in deriving flow characteristics. Some clinical applications are investigated which use pressure distribution in diagnosis of intravascular disorders such as vascular stenosis. In this paper an approach to the problem of measurement of intravascular pressure field by using velocity field obtained from flow images is proposed. The method presented in this paper uses an algorithm to calculate nonlinear equations of Navier- Stokes, assuming blood as an incompressible and Newtonian fluid. Flow images usually suffer the lack of spatial resolution. Our attempt is to consider the effect of spatial resolution on the pressure distribution estimated from this method. In order to achieve this aim, velocity map of a numerical phantom is derived at six different spatial resolutions. To determine the effects of vascular stenoses on pressure distribution, a stenotic phantom geometry is considered. A comparison between the pressure distribution obtained from the phantom and the pressure resulted from the algorithm is presented. In this regard we also compared the effects of collocated and staggered computational grids on the pressure distribution resulted from this algorithm.Keywords: Flow imaging, pressure distribution estimation, phantom, resolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1682862 Exploring the Application of Knowledge Management Factors in Esfahan University's Medical College
Authors: Alireza Shirvani, Shadi Ebrahimi Mehrabani
Abstract:
In this competitive age, one of the key tools of most successful organizations is knowledge management. Today some organizations measure their current knowledge and use it as an indicator for rating the organization on their reports. Noting that the universities and colleges of medical science have a great role in public health of societies, their access to newest scientific research and the establishment of organizational knowledge management systems is very important. In order to explore the Application of Knowledge Management Factors, a national study was undertaken. The main purpose of this study was to find the rate of the application of knowledge management factors and some ways to establish more application of knowledge management system in Esfahan University-s Medical College (EUMC). Esfahan is the second largest city after Tehran, the capital city of Iran, and the EUMC is the biggest medical college in Esfahan. To rate the application of knowledge management, this study uses a quantitative research methodology based on Probst, Raub and Romhardt model of knowledge management. A group of 267 faculty members and staff of the EUMC were asked via questionnaire. Finding showed that the rate of the application of knowledge management factors in EUMC have been lower than average. As a result, an interview with ten faculty members conducted to find the guidelines to establish more applications of knowledge management system in EUMC.
Keywords: Knowledge, knowledge management, knowledge management factors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1468861 Towards Clustering of Web-based Document Structures
Authors: Matthias Dehmer, Frank Emmert Streib, Jürgen Kilian, Andreas Zulauf
Abstract:
Methods for organizing web data into groups in order to analyze web-based hypertext data and facilitate data availability are very important in terms of the number of documents available online. Thereby, the task of clustering web-based document structures has many applications, e.g., improving information retrieval on the web, better understanding of user navigation behavior, improving web users requests servicing, and increasing web information accessibility. In this paper we investigate a new approach for clustering web-based hypertexts on the basis of their graph structures. The hypertexts will be represented as so called generalized trees which are more general than usual directed rooted trees, e.g., DOM-Trees. As a important preprocessing step we measure the structural similarity between the generalized trees on the basis of a similarity measure d. Then, we apply agglomerative clustering to the obtained similarity matrix in order to create clusters of hypertext graph patterns representing navigation structures. In the present paper we will run our approach on a data set of hypertext structures and obtain good results in Web Structure Mining. Furthermore we outline the application of our approach in Web Usage Mining as future work.Keywords: Clustering methods, graph-based patterns, graph similarity, hypertext structures, web structure mining
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1506860 Object Motion Tracking Based On Color Detection for Android Devices
Authors: Zacharenia I. Garofalaki, John T. Amorginos, John N. Ellinas
Abstract:
This paper presents the development of a robot car that can track the motion of an object by detecting its color through an Android device. The employed computer vision algorithm uses the OpenCV library, which is embedded into an Android application of a smartphone, for manipulating the captured image of the object. The captured image of the object is subjected to color conversion and is transformed to a binary image for further processing after color filtering. The desired object is clearly determined after removing pixel noise by applying image morphology operations and contour definition. Finally, the area and the center of the object are determined so that object’s motion to be tracked. The smartphone application has been placed on a robot car and transmits by Bluetooth to an Arduino assembly the motion directives so that to follow objects of a specified color. The experimental evaluation of the proposed algorithm shows reliable color detection and smooth tracking characteristics.Keywords: Android, Arduino Uno, Image processing, Object motion detection, OpenCV library.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4564859 Electroencephalography-Based Intention Recognition and Consensus Assessment during Emergency Response
Abstract:
After natural and man-made disasters, robots can bypass the danger, expedite the search, and acquire unprecedented situational awareness to design rescue plans. Brain-computer interface is a promising option to overcome the limitations of tedious manual control and operation of robots in the urgent search-and-rescue tasks. This study aims to test the feasibility of using electroencephalography (EEG) signals to decode human intentions and detect the level of consensus on robot-provided information. EEG signals were classified using machine-learning and deep-learning methods to discriminate search intentions and agreement perceptions. The results show that the average classification accuracy for intention recognition and consensus assessment is 67% and 72%, respectively, proving the potential of incorporating recognizable users’ bioelectrical responses into advanced robot-assisted systems for emergency response.
Keywords: Consensus assessment, electroencephalogram, EEG, emergency response, human-robot collaboration, intention recognition, search and rescue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 344858 Face Recognition Based On Vector Quantization Using Fuzzy Neuro Clustering
Authors: Elizabeth B. Varghese, M. Wilscy
Abstract:
A face recognition system is a computer application for automatically identifying or verifying a person from a digital image or a video frame. A lot of algorithms have been proposed for face recognition. Vector Quantization (VQ) based face recognition is a novel approach for face recognition. Here a new codebook generation for VQ based face recognition using Integrated Adaptive Fuzzy Clustering (IAFC) is proposed. IAFC is a fuzzy neural network which incorporates a fuzzy learning rule into a competitive neural network. The performance of proposed algorithm is demonstrated by using publicly available AT&T database, Yale database, Indian Face database and a small face database, DCSKU database created in our lab. In all the databases the proposed approach got a higher recognition rate than most of the existing methods. In terms of Equal Error Rate (ERR) also the proposed codebook is better than the existing methods.
Keywords: Face Recognition, Vector Quantization, Integrated Adaptive Fuzzy Clustering, Self Organization Map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2241857 Capacitor Placement in Radial Distribution System for Loss Reduction Using Artificial Bee Colony Algorithm
Authors: R. Srinivasa Rao
Abstract:
This paper presents a new method which applies an artificial bee colony algorithm (ABC) for capacitor placement in distribution systems with an objective of improving the voltage profile and reduction of power loss. The ABC algorithm is a new population based meta heuristic approach inspired by intelligent foraging behavior of honeybee swarm. The advantage of ABC algorithm is that it does not require external parameters such as cross over rate and mutation rate as in case of genetic algorithm and differential evolution and it is hard to determine these parameters in prior. The other advantage is that the global search ability in the algorithm is implemented by introducing neighborhood source production mechanism which is a similar to mutation process. To demonstrate the validity of the proposed algorithm, computer simulations are carried out on 69-bus system and compared the results with the other approach available in the literature. The proposed method has outperformed the other methods in terms of the quality of solution and computational efficiency.Keywords: Distribution system, Capacitor Placement, Loss reduction, Artificial Bee Colony Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2817