Search results for: Web based System
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16167

Search results for: Web based System

9477 Assessment of the Adaptive Pushover Analysis Using Displacement-based Loading in Prediction the Seismic Behaviour of the Unsymmetric-Plan Buildings

Authors: M.O. Makhmalbaf, F. Mohajeri Nav, M. Zabihi Samani

Abstract:

The recent drive for use of performance-based methodologies in design and assessment of structures in seismic areas has significantly increased the demand for the development of reliable nonlinear inelastic static pushover analysis tools. As a result, the adaptive pushover methods have been developed during the last decade, which unlike their conventional pushover counterparts, feature the ability to account for the effect that higher modes of vibration and progressive stiffness degradation might have on the distribution of seismic storey forces. Even in advanced pushover methods, little attention has been paid to the Unsymmetric structures. This study evaluates the seismic demands for three dimensional Unsymmetric-Plan buildings determined by the Displacement-based Adaptive Pushover (DAP) analysis, which has been introduced by Antoniou and Pinho [2004]. The capability of DAP procedure in capturing the torsional effects due to the irregularities of the structures, is investigated by comparing its estimates to the exact results, obtained from Incremental Dynamic Analysis (IDA). Also the capability of the procedure in prediction the seismic behaviour of the structure is discussed.

Keywords: Nonlinear static procedures, Unsymmetric-PlanBuildings, Torsional effects, IDA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2758
9476 General Purpose Pulse Width Modulation Based Sliding Mode Controller for Buck DC-DC

Authors: M.Bensaada , A.Boudghene Stambouli , M.Bekhti, A. Bellar, L. Boukhris

Abstract:

This paper is a simple and systematic approaches to the design and analysis a pulse width modulation (PWM) based sliding mode controller for buck DC-DC Converters. Various aspects of the design, including the practical problems and the proposed solutions, are detailed. However, these control strategies can't compensate for large load current and input voltage variations. In this paper, a new control strategy by compromising both schemes advantages and avoiding their drawbacks is proposed, analyzed and simulated.

Keywords: Buck, DC/DC converters, sliding mode control, pulse width modulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2675
9475 A Noble Flow Rate Control based on Leaky Bucket Method for Multi-Media OBS Networks

Authors: Kentaro Miyoko, Yoshihiko Mori, Yugo Ikeda, Yoshihiro Nishino, Yong-Bok Choi, Hiromi Okada

Abstract:

Optical burst switching (OBS) has been proposed to realize the next generation Internet based on the wavelength division multiplexing (WDM) network technologies. In the OBS, the burst contention is one of the major problems. The deflection routing has been designed for resolving the problem. However, the deflection routing becomes difficult to prevent from the burst contentions as the network load becomes high. In this paper, we introduce a flow rate control methods to reduce burst contentions. We propose new flow rate control methods based on the leaky bucket algorithm and deflection routing, i.e. separate leaky bucket deflection method, and dynamic leaky bucket deflection method. In proposed methods, edge nodes which generate data bursts carry out the flow rate control protocols. In order to verify the effectiveness of the flow rate control in OBS networks, we show that the proposed methods improve the network utilization and reduce the burst loss probability through computer simulations.

Keywords: Optical burst switching, OBS, flow rate control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703
9474 Confidence Intervals for the Normal Mean with Known Coefficient of Variation

Authors: Suparat Niwitpong

Abstract:

In this paper we proposed two new confidence intervals for the normal population mean with known coefficient of variation. This situation occurs normally in environment and agriculture experiments where the scientist knows the coefficient of variation of their experiments. We propose two new confidence intervals for this problem based on the recent work of Searls [5] and the new method proposed in this paper for the first time. We derive analytic expressions for the coverage probability and the expected length of each confidence interval. Monte Carlo simulation will be used to assess the performance of these intervals based on their expected lengths.

Keywords: confidence interval, coverage probability, expected length, known coefficient of variation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1643
9473 Image Classification and Accuracy Assessment Using the Confusion Matrix, Contingency Matrix, and Kappa Coefficient

Authors: F. F. Howard, C. B. Boye, I. Yakubu, J. S. Y. Kuma

Abstract:

One of the ways that could be used for the production of land use and land cover maps by a procedure known as image classification is the use of the remote sensing technique. Numerous elements ought to be taken into consideration, including the availability of highly satisfactory Landsat imagery, secondary data and a precise classification process. The goal of this study was to classify and map the land use and land cover of the study area using remote sensing and Geospatial Information System (GIS) analysis. The classification was done using Landsat 8 satellite images acquired in December 2020 covering the study area. The Landsat image was downloaded from the USGS. The Landsat image with 30 m resolution was geo-referenced to the WGS_84 datum and Universal Transverse Mercator (UTM) Zone 30N coordinate projection system. A radiometric correction was applied to the image to reduce the noise in the image. This study consists of two sections: the Land Use/Land Cover (LULC) and Accuracy Assessments using the confusion and contingency matrix and the Kappa coefficient. The LULC classifications were vegetation (agriculture) (67.87%), water bodies (0.01%), mining areas (5.24%), forest (26.02%), and settlement (0.88%). The overall accuracy of 97.87% and the kappa coefficient (K) of 97.3% were obtained for the confusion matrix. While an overall accuracy of 95.7% and a Kappa coefficient of 0.947 were obtained for the contingency matrix, the kappa coefficients were rated as substantial; hence, the classified image is fit for further research.

Keywords: Confusion Matrix, contingency matrix, kappa coefficient, land used/ land cover, accuracy assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 225
9472 An Advanced Technology for Renovation of Extruding Shafts

Authors: Dimitar Karastoyanov, Vladimir Monov

Abstract:

The paper is concerned with the technological process of renovation of shafts used in industrial manufacturing for extruding of sheet material. In the classical renovation technologies, a chrome based coating is applied to the working surface of the shaft in galvanic baths. The process, however, is known to be exclusively harmful due to the waste cyanide products. In this work, we present an advanced nanotechnology based on nonelectric chemical laying of a nickel coating with included nanoparticles. The technology is environmentally harmless and the new coating features an increased hardness and wear resistance. Results from experimental tests of the nanostructured nickel coating are presented and discussed.

Keywords: Materials processing, nanoparticles, nickel coating, shafts renovation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1861
9471 Using Rao-Blackwellised Particle Filter Track 3D Arm Motion based on Hierarchical Limb Model

Authors: XueSong Yu, JiaFeng Liu, XiangLong Tang, JianHua Huang

Abstract:

For improving the efficiency of human 3D tracking, we present an algorithm to track 3D Arm Motion. First, the Hierarchy Limb Model (HLM) is proposed based on the human 3D skeleton model. Second, via graph decomposition, the arm motion state space, modeled by HLM, can be discomposed into two low dimension subspaces: root nodes and leaf nodes. Finally, Rao-Blackwellised Particle Filter is used to estimate the 3D arm motion. The result of experiment shows that our algorithm can advance the computation efficiency.

Keywords: Hierarchy Limb Model; Rao-Blackwellised Particle Filter; 3D tracking

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581
9470 Information Fusion for Identity Verification

Authors: Girija Chetty, Monica Singh

Abstract:

In this paper we propose a novel approach for ascertaining human identity based on fusion of profile face and gait biometric cues The identification approach based on feature learning in PCA-LDA subspace, and classification using multivariate Bayesian classifiers allows significant improvement in recognition accuracy for low resolution surveillance video scenarios. The experimental evaluation of the proposed identification scheme on a publicly available database [2] showed that the fusion of face and gait cues in joint PCA-LDA space turns out to be a powerful method for capturing the inherent multimodality in walking gait patterns, and at the same time discriminating the person identity..

Keywords: Biometrics, gait recognition, PCA, LDA, Eigenface, Fisherface, Multivariate Gaussian Classifier

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772
9469 A Block World Problem Based Sudoku Solver

Authors: Luciana Abednego, Cecilia Nugraheni

Abstract:

There are many approaches proposed for solving Sudoku puzzles. One of them is by modelling the puzzles as block world problems. There have been three model for Sudoku solvers based on this approach. Each model expresses Sudoku solver as a parameterized multi agent systems. In this work, we propose a new model which is an improvement over the existing models. This paper presents the development of a Sudoku solver that implements all the proposed models. Some experiments have been conducted to determine the performance of each model.

Keywords: Sudoku puzzle, Sudoku solver, block world problem, parameterized multi agent systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2232
9468 Classification Algorithms in Human Activity Recognition using Smartphones

Authors: Mohd Fikri Azli bin Abdullah, Ali Fahmi Perwira Negara, Md. Shohel Sayeed, Deok-Jai Choi, Kalaiarasi Sonai Muthu

Abstract:

Rapid advancement in computing technology brings computers and humans to be seamlessly integrated in future. The emergence of smartphone has driven computing era towards ubiquitous and pervasive computing. Recognizing human activity has garnered a lot of interest and has raised significant researches- concerns in identifying contextual information useful to human activity recognition. Not only unobtrusive to users in daily life, smartphone has embedded built-in sensors that capable to sense contextual information of its users supported with wide range capability of network connections. In this paper, we will discuss the classification algorithms used in smartphone-based human activity. Existing technologies pertaining to smartphone-based researches in human activity recognition will be highlighted and discussed. Our paper will also present our findings and opinions to formulate improvement ideas in current researches- trends. Understanding research trends will enable researchers to have clearer research direction and common vision on latest smartphone-based human activity recognition area.

Keywords: Classification algorithms, Human Activity Recognition (HAR), Smartphones

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6293
9467 A Nonlinear Parabolic Partial Differential Equation Model for Image Enhancement

Authors: Tudor Barbu

Abstract:

We present a robust nonlinear parabolic partial differential equation (PDE)-based denoising scheme in this article. Our approach is based on a second-order anisotropic diffusion model that is described first. Then, a consistent and explicit numerical approximation algorithm is constructed for this continuous model by using the finite-difference method. Finally, our restoration experiments and method comparison, which prove the effectiveness of this proposed technique, are discussed in this paper.

Keywords: Image denoising and restoration, nonlinear PDE model, anisotropic diffusion, numerical approximation scheme, finite differences.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1296
9466 Medical Image Fusion Based On Redundant Wavelet Transform and Morphological Processing

Authors: P. S. Gomathi, B. Kalaavathi

Abstract:

The process in which the complementary information from multiple images is integrated to provide composite image that contains more information than the original input images is called image fusion. Medical image fusion provides useful information from multimodality medical images that provides additional information to the doctor for diagnosis of diseases in a better way. This paper represents the wavelet based medical image fusion algorithm on different multimodality medical images. In order to fuse the medical images, images are decomposed using Redundant Wavelet Transform (RWT). The high frequency coefficients are convolved with morphological operator followed by the maximum-selection (MS) rule. The low frequency coefficients are processed by MS rule. The reconstructed image is obtained by inverse RWT. The quantitative measures which includes Mean, Standard Deviation, Average Gradient, Spatial frequency, Edge based Similarity Measures are considered for evaluating the fused images. The performance of this proposed method is compared with Pixel averaging, PCA, and DWT fusion methods. When compared with conventional methods, the proposed framework provides better performance for analysis of multimodality medical images.

Keywords: Discrete Wavelet Transform (DWT), Image Fusion, Morphological Processing, Redundant Wavelet Transform (RWT).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2151
9465 A Survey of Field Programmable Gate Array-Based Convolutional Neural Network Accelerators

Authors: Wei Zhang

Abstract:

With the rapid development of deep learning, neural network and deep learning algorithms play a significant role in various practical applications. Due to the high accuracy and good performance, Convolutional Neural Networks (CNNs) especially have become a research hot spot in the past few years. However, the size of the networks becomes increasingly large scale due to the demands of the practical applications, which poses a significant challenge to construct a high-performance implementation of deep learning neural networks. Meanwhile, many of these application scenarios also have strict requirements on the performance and low-power consumption of hardware devices. Therefore, it is particularly critical to choose a moderate computing platform for hardware acceleration of CNNs. This article aimed to survey the recent advance in Field Programmable Gate Array (FPGA)-based acceleration of CNNs. Various designs and implementations of the accelerator based on FPGA under different devices and network models are overviewed, and the versions of Graphic Processing Units (GPUs), Application Specific Integrated Circuits (ASICs) and Digital Signal Processors (DSPs) are compared to present our own critical analysis and comments. Finally, we give a discussion on different perspectives of these acceleration and optimization methods on FPGA platforms to further explore the opportunities and challenges for future research. More helpfully, we give a prospect for future development of the FPGA-based accelerator.

Keywords: Deep learning, field programmable gate array, FPGA, hardware acceleration, convolutional neural networks, CNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 882
9464 Collaboration in Palliative Care Networks in Urban and Rural Regions of Switzerland

Authors: R. Schweighoffer, N. Nagy, E. Reeves, B. Liebig

Abstract:

Due to aging populations, the need for seamless palliative care provision is of central interest for western societies. An essential aspect of palliative care delivery is the quality of collaboration amongst palliative care providers. Therefore, the current research is based on Bainbridge’s conceptual framework, which provides an outline for the evaluation of palliative care provision. This study is the first one to investigate the predictive validity of spatial distribution on the quantity of interaction amongst various palliative care providers. Furthermore, based on the familiarity principle, we examine whether the extent of collaboration influences the perceived quality of collaboration among palliative care providers in urban versus rural areas of Switzerland. Based on a population-representative survey of Swiss palliative care providers, the results of the current study show that professionals in densely populated areas report higher absolute numbers of interactions and are more satisfied with their collaborative practice. This indicates that palliative care providers who work in urban areas are better embedded into networks than their counterparts in more rural areas. The findings are especially important, considering that efficient collaboration is a prerequisite to achieve satisfactory patient outcomes. Conclusively, measures should be taken to foster collaboration in weakly interconnected palliative care networks.

Keywords: Collaboration, healthcare networks, palliative care, Switzerland.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1452
9463 Evaluating Content Based Image Retrieval Techniques with the One Million Images CLIC Test Bed

Authors: Pierre-Alain Moëllic, Patrick Hède, Gr egory Grefenstette, Christophe Millet

Abstract:

Pattern recognition and image recognition methods are commonly developed and tested using testbeds, which contain known responses to a query set. Until now, testbeds available for image analysis and content-based image retrieval (CBIR) have been scarce and small-scale. Here we present the one million images CEA-List Image Collection (CLIC) testbed that we have produced, and report on our use of this testbed to evaluate image analysis merging techniques. This testbed will soon be made publicly available through the EU MUSCLE Network of Excellence.

Keywords: CBIR, CLIC, evaluation, image indexing and retrieval, testbed.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1380
9462 The Relationship between Absorptive Capacity and Green Innovation

Authors: R. Hashim, A. J. Bock, S. Cooper

Abstract:

Absorptive capacity generally facilitates the adoption of innovation. How does this relationship change when economic return is not the sole driver of innovation uptake? We investigate whether absorptive capacity facilitates the adoption of green innovation based on a survey of 79 construction companies in Scotland. Based on the results of multiple regression analyses, we confirm that existing knowledge utilisation (EKU), knowledge building (KB) and external knowledge acquisition (EKA) are significant predictors of green process GP), green administrative (GA) and green technical innovation (GT), respectively. We discuss the implications for theories of innovation adoption and knowledge enhancement associated with environmentally-friendly practices.

Keywords: Absorptive capacity, construction industry, environmental, green innovation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3162
9461 Confidence Interval for the Inverse of a Normal Mean with a Known Coefficient of Variation

Authors: Arunee Wongkha, Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

In this paper, we propose two new confidence intervals for the inverse of a normal mean with a known coefficient of variation. One of new confidence intervals for the inverse of a normal mean with a known coefficient of variation is constructed based on the pivotal statistic Z where Z is a standard normal distribution and another confidence interval is constructed based on the generalized confidence interval, presented by Weerahandi. We examine the performance of these confidence intervals in terms of coverage probabilities and average lengths via Monte Carlo simulation.

Keywords: The inverse of a normal mean, confidence interval, generalized confidence intervals, known coefficient of variation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2578
9460 Methodologies for Crack Initiation in Welded Joints Applied to Inspection Planning

Authors: Guang Zou, Kian Banisoleiman, Arturo González

Abstract:

Crack initiation and propagation threatens structural integrity of welded joints and normally inspections are assigned based on crack propagation models. However, the approach based on crack propagation models may not be applicable for some high-quality welded joints, because the initial flaws in them may be so small that it may take long time for the flaws to develop into a detectable size. This raises a concern regarding the inspection planning of high-quality welded joins, as there is no generally acceptable approach for modeling the whole fatigue process that includes the crack initiation period. In order to address the issue, this paper reviews treatment methods for crack initiation period and initial crack size in crack propagation models applied to inspection planning. Generally, there are four approaches, by: 1) Neglecting the crack initiation period and fitting a probabilistic distribution for initial crack size based on statistical data; 2) Extrapolating the crack propagation stage to a very small fictitious initial crack size, so that the whole fatigue process can be modeled by crack propagation models; 3) Assuming a fixed detectable initial crack size and fitting a probabilistic distribution for crack initiation time based on specimen tests; and, 4) Modeling the crack initiation and propagation stage separately using small crack growth theories and Paris law or similar models. The conclusion is that in view of trade-off between accuracy and computation efforts, calibration of a small fictitious initial crack size to S-N curves is the most efficient approach.

Keywords: Crack initiation, fatigue reliability, inspection planning, welded joints.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1392
9459 Development and in vitro Characterization of Self-nanoemulsifying Drug Delivery Systems of Valsartan

Authors: P. S. Rajinikanth, Yeoh Suyu, Sanjay Garg

Abstract:

The present study is aim to prepare and evaluate the selfnanoemulsifying drug delivery (SNEDDS) system of a poorly water soluble drug valsartan in order to achieve a better dissolution rate which would further help in enhancing oral bioavailability. The present research work describes a SNEDDS of valsartan using labrafil M 1944 CS, Tween 80 and Transcutol HP. The pseudoternary phase diagrams with presence and absence of drug were plotted to check for the emulsification range and also to evaluate the effect of valsartan on the emulsification behavior of the phases. The mixtures consisting of oil (labrafil M 1944 CS) with surfactant (tween 80), co-surfactant (Transcutol HP) were found to be optimum formulations. Prepared formulations were evaluated for its particle size distribution, nanoemulsifying properties, robustness to dilution, self emulsication time, turbidity measurement, drug content and invitro dissolution. The optimized formulations are further evaluated for heating cooling cycle, centrifugation studies, freeze thaw cycling, particle size distribution and zeta potential were carried out to confirm the stability of the formed SNEDDS formulations. The prepared formulation revealed t a significant improvement in terms of the drug solubility as compared with marketed tablet and pure drug.

Keywords: Self Emulsifying Drug Delivery System, Valsartan, Bioavailability, poorly soluble drug.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2671
9458 Towards a Suitable and Systematic Approach for Component Based Software Development

Authors: Kuljit Kaur, Parminder Kaur, Jaspreet Bedi, Hardeep Singh

Abstract:

Software crisis refers to the situation in which the developers are not able to complete the projects within time and budget constraints and moreover these overscheduled and over budget projects are of low quality as well. Several methodologies have been adopted form time to time to overcome this situation and now in the focus is component based software engineering. In this approach, emphasis is on reuse of already existing software artifacts. But the results can not be achieved just by preaching the principles; they need to be practiced as well. This paper highlights some of the very basic elements of this approach, which has to be in place to get the desired goals of high quality, low cost with shorter time-to-market software products.

Keywords: Component Model, Software Components, SoftwareRepository, Process Models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1757
9457 A New Model for Economic Optimization of Water Diversion System during Dam Construction using PSO Algorithm

Authors: Saeed Sedighizadeh, Abbas Mansoori, Mohammad Reza Pirestani, Davoud Sedighizadeh

Abstract:

The usual method of river flow diversion involves construction of tunnels and cofferdams. Given the fact that the cost of diversion works could be as high as 10-20% of the total dam construction cost, due attention should be paid to optimum design of the diversion works. The cost of diversion works depends, on factors, such as: the tunnel dimensions and the intended tunneling support measures during and after excavation; quality and characterizes of the rock through which the tunnel should be excavated; the dimensions of the upstream (and downstream) cofferdams; and the magnitude of river flood the system is designed to divert. In this paper by use of the cost of unit prices for tunnel excavation, tunnel lining, tunnel support (rock bolt + shotcrete) and cofferdam fill the cost function was determined. The function is then minimized by the aid of PSO Algorithm (particle swarm optimization). It is found that the optimum diameter and the total diversion cost are directly related to the river flood discharge (Q). It has also shown that in addition to optimum diameter design discharge (Q), river length, tunnel length, is mainly a function of the ratios (not the absolute values) of the unit prices and does not depend on the overall price levels in the respective country. The results of optimization use in some of the case study lead us to significant changes in the cost.

Keywords: Diversion Tunnel, Optimization, PSO Algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2721
9456 A Character Detection Method for Ancient Yi Books Based on Connected Components and Regressive Character Segmentation

Authors: Xu Han, Shanxiong Chen, Shiyu Zhu, Xiaoyu Lin, Fujia Zhao, Dingwang Wang

Abstract:

Character detection is an important issue for character recognition of ancient Yi books. The accuracy of detection directly affects the recognition effect of ancient Yi books. Considering the complex layout, the lack of standard typesetting and the mixed arrangement between images and texts, we propose a character detection method for ancient Yi books based on connected components and regressive character segmentation. First, the scanned images of ancient Yi books are preprocessed with nonlocal mean filtering, and then a modified local adaptive threshold binarization algorithm is used to obtain the binary images to segment the foreground and background for the images. Second, the non-text areas are removed by the method based on connected components. Finally, the single character in the ancient Yi books is segmented by our method. The experimental results show that the method can effectively separate the text areas and non-text areas for ancient Yi books and achieve higher accuracy and recall rate in the experiment of character detection, and effectively solve the problem of character detection and segmentation in character recognition of ancient books.

Keywords: Computing methodologies, interest point, salient region detections, image segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 853
9455 Wavelet-Based Classification of Myocardial Ischemia, Arrhythmia, Congestive Heart Failure and Sleep Apnea

Authors: Santanu Chattopadhyay, Gautam Sarkar, Arabinda Das

Abstract:

This paper presents wavelet based classification of various heart diseases. Electrocardiogram signals of different heart patients have been studied. Statistical natures of electrocardiogram signals for different heart diseases have been compared with the statistical nature of electrocardiograms for normal persons. Under this study four different heart diseases have been considered as follows: Myocardial Ischemia (MI), Congestive Heart Failure (CHF), Arrhythmia and Sleep Apnea. Statistical nature of electrocardiograms for each case has been considered in terms of kurtosis values of two types of wavelet coefficients: approximate and detail. Nine wavelet decomposition levels have been considered in each case. Kurtosis corresponding to both approximate and detail coefficients has been considered for decomposition level one to decomposition level nine. Based on significant difference, few decomposition levels have been chosen and then used for classification.

Keywords: Arrhythmia, congestive heart failure, discrete wavelet transform, electrocardiogram, myocardial ischemia, sleep apnea.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 660
9454 Adopting Artificial Intelligence and Deep Learning Techniques in Cloud Computing for Operational Efficiency

Authors: Sandesh Achar

Abstract:

Artificial intelligence (AI) is being increasingly incorporated into many applications across various sectors such as health, education, security, and agriculture. Recently, there has been rapid development in cloud computing technology, resulting in AI’s implementation into cloud computing to enhance and optimize the technology service rendered. The deployment of AI in cloud-based applications has brought about autonomous computing, whereby systems achieve stated results without human intervention. Despite the amount of research into autonomous computing, work incorporating AI/ML into cloud computing to enhance its performance and resource allocation remains a fundamental challenge. This paper highlights different manifestations, roles, trends, and challenges related to AI-based cloud computing models. This work reviews and highlights investigations and progress in the domain. Future directions are suggested for leveraging AI/ML in next-generation computing for emerging computing paradigms such as cloud environments. Adopting AI-based algorithms and techniques to increase operational efficiency, cost savings, automation, reducing energy consumption and solving complex cloud computing issues are the major findings outlined in this paper.

Keywords: Artificial intelligence, AI, cloud computing, deep learning, machine learning, ML, internet of things, IoT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 605
9453 High Capacity Data Hiding based on Predictor and Histogram Modification

Authors: Hui-Yu Huang, Shih-Hsu Chang

Abstract:

In this paper, we propose a high capacity image hiding technology based on pixel prediction and the difference of modified histogram. This approach is used the pixel prediction and the difference of modified histogram to calculate the best embedding point. This approach can improve the predictive accuracy and increase the pixel difference to advance the hiding capacity. We also use the histogram modification to prevent the overflow and underflow. Experimental results demonstrate that our proposed method within the same average hiding capacity can still keep high quality of image and low distortion

Keywords: data hiding, predictor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1878
9452 Chaos-based Secure Communication via Continuous Variable Structure Control

Authors: Cheng-Fang Huang, Meei-Ling Hung, Teh-Lu Liao, Her-Terng Yau, Jun-Juh Yan

Abstract:

The design of chaos-based secure communication via synchronized modified Chua-s systems is investigated in this paper. A continuous control law is proposed to ensure synchronization of the master and slave modified Chua-s systems by using the variable structure control technique. Particularly, the concept of extended systems is introduced such that a continuous control input is obtained to avoid chattering phenomenon. Then, it becomes possible to ensure that the message signal embedded in the transmitter can be recovered in the receiver.

Keywords: Chaos, Secure communication, Synchronization, Variable structure control (VSC)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1426
9451 The Nanobiotechnology of Obtaining of Collagen Gels from Marin Fish Skin and Yours Reological Properties for using Like New Materials in Dental Medicine

Authors: Anamaria Bechir, Rodica Sirbu, Minodora Leca, Maria Maris, Dan Artenie Maris, Emilia Mihaela Cadar, Marius Maris

Abstract:

This paper aims at presenting the biotechnology used to obtain collagen-based gels from shark (Squalus acanthias) and brill skin, marine fish growing in the Black Sea. Due to the structure of its micro-fibres, collagen can be considered a nanomaterial; in order to use collagen-based matrixes as biomaterial, rheological studies must be performed first, to state whether they are stable or not. For the triple-helix structure to remain stable within these gels at room or human body temperature, they must be stabilized by reticulation.

Keywords: Collagen, biotechnology, reticulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2966
9450 Efficient Dimensionality Reduction of Directional Overcurrent Relays Optimal Coordination Problem

Authors: Fouad Salha , X. Guillaud

Abstract:

Directional over current relays (DOCR) are commonly used in power system protection as a primary protection in distribution and sub-transmission electrical systems and as a secondary protection in transmission systems. Coordination of protective relays is necessary to obtain selective tripping. In this paper, an approach for efficiency reduction of DOCRs nonlinear optimum coordination (OC) is proposed. This was achieved by modifying the objective function and relaxing several constraints depending on the four constraints classification, non-valid, redundant, pre-obtained and valid constraints. According to this classification, the far end fault effect on the objective function and constraints, and in consequently on relay operating time, was studied. The study was carried out, firstly by taking into account the near-end and far-end faults in DOCRs coordination problem formulation; and then faults very close to the primary relays (nearend faults). The optimal coordination (OC) was achieved by simultaneously optimizing all variables (TDS and Ip) in nonlinear environment by using of Genetic algorithm nonlinear programming techniques. The results application of the above two approaches on 6-bus and 26-bus system verify that the far-end faults consideration on OC problem formulation don-t lose the optimality.

Keywords: Backup/Primary relay, Coordination time interval (CTI), directional over current relays, Genetic algorithm, time dial setting (TDS), pickup current setting (Ip), nonlinear programming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575
9449 Satellite Imagery Classification Based on Deep Convolution Network

Authors: Zhong Ma, Zhuping Wang, Congxin Liu, Xiangzeng Liu

Abstract:

Satellite imagery classification is a challenging problem with many practical applications. In this paper, we designed a deep convolution neural network (DCNN) to classify the satellite imagery. The contributions of this paper are twofold — First, to cope with the large-scale variance in the satellite image, we introduced the inception module, which has multiple filters with different size at the same level, as the building block to build our DCNN model. Second, we proposed a genetic algorithm based method to efficiently search the best hyper-parameters of the DCNN in a large search space. The proposed method is evaluated on the benchmark database. The results of the proposed hyper-parameters search method show it will guide the search towards better regions of the parameter space. Based on the found hyper-parameters, we built our DCNN models, and evaluated its performance on satellite imagery classification, the results show the classification accuracy of proposed models outperform the state of the art method.

Keywords: Satellite imagery classification, deep convolution network, genetic algorithm, hyper-parameter optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2337
9448 Developing a Customizable Serious Game and Its Applicability in the Classroom

Authors: Anita Kéri

Abstract:

Recent developments in the field of education have led to a renewed interest in teaching methodologies and practices. Gamification is fast becoming a key instrument in the education of new generations and besides other methods, serious games have become the center of attention. Ready-built serious games are available for most higher education institutions to buy and implement. However, monetary restraints and the unalterable nature of the games might deter most higher education institutions from the application of these serious games. Therefore, there is a continuously growing need for a customizable serious game that has been developed based on a concrete need analysis and experts’ opinion. There has been little evidence so far of serious games that have been created based on relevant and current need analysis from higher education institution teachers, professional practitioners and students themselves. Therefore, the aim of this current paper is to analyze the needs of higher education institution educators with special emphasis on their needs, the applicability of serious games in their classrooms, and exploring options for the development of a customizable serious game framework. The paper undertakes to analyze workshop discussions on implementing serious games in education and propose a customizable serious game framework applicable in the education of the new generation. Research results show that the most important feature of a serious game is its customizability. The fact that practitioners are able to manage different scenarios and upload their own content to a game seems to be a key to the increasingly widespread application of serious games in the classroom.

Keywords: Education, gamification, game-based learning, serious games.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 875