Search results for: parallel algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3079

Search results for: parallel algorithms

1429 Hyper Tuned RBF SVM: Approach for the Prediction of the Breast Cancer

Authors: Surita Maini, Sanjay Dhanka

Abstract:

Machine learning (ML) involves developing algorithms and statistical models that enable computers to learn and make predictions or decisions based on data without being explicitly programmed. Because of its unlimited abilities ML is gaining popularity in medical sectors; Medical Imaging, Electronic Health Records, Genomic Data Analysis, Wearable Devices, Disease Outbreak Prediction, Disease Diagnosis, etc. In the last few decades, many researchers have tried to diagnose Breast Cancer (BC) using ML, because early detection of any disease can save millions of lives. Working in this direction, the authors have proposed a hybrid ML technique RBF SVM, to predict the BC in earlier the stage. The proposed method is implemented on the Breast Cancer UCI ML dataset with 569 instances and 32 attributes. The authors recorded performance metrics of the proposed model i.e., Accuracy 98.24%, Sensitivity 98.67%, Specificity 97.43%, F1 Score 98.67%, Precision 98.67%, and run time 0.044769 seconds. The proposed method is validated by K-Fold cross-validation.

Keywords: breast cancer, support vector classifier, machine learning, hyper parameter tunning

Procedia PDF Downloads 61
1428 Tree Species Classification Using Effective Features of Polarimetric SAR and Hyperspectral Images

Authors: Milad Vahidi, Mahmod R. Sahebi, Mehrnoosh Omati, Reza Mohammadi

Abstract:

Forest management organizations need information to perform their work effectively. Remote sensing is an effective method to acquire information from the Earth. Two datasets of remote sensing images were used to classify forested regions. Firstly, all of extractable features from hyperspectral and PolSAR images were extracted. The optical features were spectral indexes related to the chemical, water contents, structural indexes, effective bands and absorption features. Also, PolSAR features were the original data, target decomposition components, and SAR discriminators features. Secondly, the particle swarm optimization (PSO) and the genetic algorithms (GA) were applied to select optimization features. Furthermore, the support vector machine (SVM) classifier was used to classify the image. The results showed that the combination of PSO and SVM had higher overall accuracy than the other cases. This combination provided overall accuracy about 90.56%. The effective features were the spectral index, the bands in shortwave infrared (SWIR) and the visible ranges and certain PolSAR features.

Keywords: hyperspectral, PolSAR, feature selection, SVM

Procedia PDF Downloads 411
1427 Methods for Distinction of Cattle Using Supervised Learning

Authors: Radoslav Židek, Veronika Šidlová, Radovan Kasarda, Birgit Fuerst-Waltl

Abstract:

Machine learning represents a set of topics dealing with the creation and evaluation of algorithms that facilitate pattern recognition, classification, and prediction, based on models derived from existing data. The data can present identification patterns which are used to classify into groups. The result of the analysis is the pattern which can be used for identification of data set without the need to obtain input data used for creation of this pattern. An important requirement in this process is careful data preparation validation of model used and its suitable interpretation. For breeders, it is important to know the origin of animals from the point of the genetic diversity. In case of missing pedigree information, other methods can be used for traceability of animal´s origin. Genetic diversity written in genetic data is holding relatively useful information to identify animals originated from individual countries. We can conclude that the application of data mining for molecular genetic data using supervised learning is an appropriate tool for hypothesis testing and identifying an individual.

Keywords: genetic data, Pinzgau cattle, supervised learning, machine learning

Procedia PDF Downloads 541
1426 Fiber Stiffness Detection of GFRP Using Combined ABAQUS and Genetic Algorithms

Authors: Gyu-Dong Kim, Wuk-Jae Yoo, Sang-Youl Lee

Abstract:

Composite structures offer numerous advantages over conventional structural systems in the form of higher specific stiffness and strength, lower life-cycle costs, and benefits such as easy installation and improved safety. Recently, there has been a considerable increase in the use of composites in engineering applications and as wraps for seismic upgrading and repairs. However, these composites deteriorate with time because of outdated materials, excessive use, repetitive loading, climatic conditions, manufacturing errors, and deficiencies in inspection methods. In particular, damaged fibers in a composite result in significant degradation of structural performance. In order to reduce the failure probability of composites in service, techniques to assess the condition of the composites to prevent continual growth of fiber damage are required. Condition assessment technology and nondestructive evaluation (NDE) techniques have provided various solutions for the safety of structures by means of detecting damage or defects from static or dynamic responses induced by external loading. A variety of techniques based on detecting the changes in static or dynamic behavior of isotropic structures has been developed in the last two decades. These methods, based on analytical approaches, are limited in their capabilities in dealing with complex systems, primarily because of their limitations in handling different loading and boundary conditions. Recently, investigators have introduced direct search methods based on metaheuristics techniques and artificial intelligence, such as genetic algorithms (GA), simulated annealing (SA) methods, and neural networks (NN), and have promisingly applied these methods to the field of structural identification. Among them, GAs attract our attention because they do not require a considerable amount of data in advance in dealing with complex problems and can make a global solution search possible as opposed to classical gradient-based optimization techniques. In this study, we propose an alternative damage-detection technique that can determine the degraded stiffness distribution of vibrating laminated composites made of Glass Fiber-reinforced Polymer (GFRP). The proposed method uses a modified form of the bivariate Gaussian distribution function to detect degraded stiffness characteristics. In addition, this study presents a method to detect the fiber property variation of laminated composite plates from the micromechanical point of view. The finite element model is used to study free vibrations of laminated composite plates for fiber stiffness degradation. In order to solve the inverse problem using the combined method, this study uses only first mode shapes in a structure for the measured frequency data. In particular, this study focuses on the effect of the interaction among various parameters, such as fiber angles, layup sequences, and damage distributions, on fiber-stiffness damage detection.

Keywords: stiffness detection, fiber damage, genetic algorithm, layup sequences

Procedia PDF Downloads 262
1425 Multi-Subpopulation Genetic Algorithm with Estimation of Distribution Algorithm for Textile Batch Dyeing Scheduling Problem

Authors: Nhat-To Huynh, Chen-Fu Chien

Abstract:

Textile batch dyeing scheduling problem is complicated which includes batch formation, batch assignment on machines, batch sequencing with sequence-dependent setup time. Most manufacturers schedule their orders manually that are time consuming and inefficient. More power methods are needed to improve the solution. Motivated by the real needs, this study aims to propose approaches in which genetic algorithm is developed with multi-subpopulation and hybridised with estimation of distribution algorithm to solve the constructed problem for minimising the makespan. A heuristic algorithm is designed and embedded into the proposed algorithms to improve the ability to get out of the local optima. In addition, an empirical study is conducted in a textile company in Taiwan to validate the proposed approaches. The results have showed that proposed approaches are more efficient than simulated annealing algorithm.

Keywords: estimation of distribution algorithm, genetic algorithm, multi-subpopulation, scheduling, textile dyeing

Procedia PDF Downloads 293
1424 Patented Free-Space Optical System for Auto Aligned Optical Beam Allowing to Compensate Mechanical Misalignments

Authors: Aurelien Boutin

Abstract:

In optical systems such as Variable Optical Delay Lines, where a collimated beam has to go back and forth, corner cubes are used in order to keep the reflected beam parallel to the incoming beam. However, the reflected beam can be laterally shifted, which will lead to losses. In this paper, we report on a patented optical design that allows keeping the reflected beam with the exact same position and direction whatever the displacement of the corner cube leading to zero losses. After explaining how the optical design works and theoretically allows to compensate for any defects in the translation of the corner cube, we will present the results of experimental comparisons between a standard layout (i.e., only corner cubes) and our optical layout. To compare both optical layouts, we used a fiber-to-fiber coupling setup. It consists of a couple of lights from one fiber to the other, thanks to two lenses. The ensemble [fiber+lense] is fixed and called a collimator so that the light is coupled from one collimator to another. Each collimator was precisely made in order to have a precise working distance. In the experiment, we measured and compared the Insertion Losses (IL) variations between both collimators with the distance between them (i.e., natural Gaussian beam coupling losses) and between both collimators in the different optical layouts tested, with the same optical length propagation. We will show that the IL variations of our setup are less than 0.05dB with respect to the IL variations of collimators alone.

Keywords: free-space optics, variable optical delay lines, optical cavity, auto-alignment

Procedia PDF Downloads 93
1423 The Optimization Process of Aortic Heart Valve Stent Geometry

Authors: Arkadiusz Mezyk, Wojciech Klein, Mariusz Pawlak, Jacek Gnilka

Abstract:

The aortic heart valve stents should fulfill many criterions. These criteria have a strong impact on the geometrical shape of the stent. Usually, the final construction of stent is a result of many year experience and knowledge. Depending on patents claims, different stent shapes are produced by different companies. This causes difficulties for biomechanics engineers narrowing the domain of feasible solutions. The paper present optimization method for stent geometry defining by a specific analytical equation based on various mathematical functions. This formula was implemented as APDL script language in ANSYS finite element environment. For the purpose of simulation tests, a few parameters were separated from developed equation. The application of the genetic algorithms allows finding the best solution due to selected objective function. Obtained solution takes into account parameters such as radial force, compression ratio and coefficient of expansion on the transverse axial.

Keywords: aortic stent, optimization process, geometry, finite element method

Procedia PDF Downloads 275
1422 Using A Blockchain-Based, End-to-End Encrypted Communication System Between Mobile Terminals to Improve Organizational Privacy

Authors: Andrei Bogdan Stanescu, Robert Stana

Abstract:

Creating private and secure communication channels between employees has become a critical aspect in order to ensure organizational integrity and avoid leaks of sensitive information. With the widespread use of modern methods of disrupting communication between users, real use-cases of advanced encryption mechanisms have emerged to avoid cyber-attackers that are willing to intercept private conversations between critical employees in an organization. This paper aims to present a custom implementation of a messaging application named “Whisper” that uses end-to-end encryption (E2EE) mechanisms and blockchain-related components to protect sensitive conversations and mitigate the risks of information breaches inside organizations. The results of this research paper aim to expand the areas of applicability of E2EE algorithms and integrations with private blockchains in chat applications as a viable method of enhancing intra-organizational communication privacy.

Keywords: end-to-end encryption, mobile communication, cryptography, communication security, data privacy

Procedia PDF Downloads 81
1421 The Role of Artificial Intelligence Algorithms in Decision-Making Policies

Authors: Marisa Almeida AraúJo

Abstract:

Artificial intelligence (AI) tools are being used (including in the criminal justice system) and becomingincreasingly popular. The many questions that these (future) super-beings pose the neuralgic center is rooted in the (old) problematic between rationality and morality. For instance, if we follow a Kantian perspective in which morality derives from AI, rationality will also surpass man in ethical and moral standards, questioning the nature of mind, the conscience of self and others, and moral. The recognition of superior intelligence in a non-human being puts us in the contingency of having to recognize a pair in a form of new coexistence and social relationship. Just think of the humanoid robot Sophia, capable of reasoning and conversation (and who has been recognized for Saudi citizenship; a fact that symbolically demonstrates our empathy with the being). Machines having a more intelligent mind, and even, eventually, with higher ethical standards to which, in the alluded categorical imperative, we would have to subject ourselves under penalty of contradiction with the universal Kantian law. Recognizing the complex ethical and legal issues and the significant impact on human rights and democratic functioning itself is the goal of our work.

Keywords: ethics, artificial intelligence, legal rules, principles, philosophy

Procedia PDF Downloads 190
1420 Influence of AAR-Induced Expansion Level on Confinement Efficiency of CFRP Wrapping Applied to Damaged Circular Concrete Columns

Authors: Thamer Kubat, Riadh Al Mahiadi, Ahmad Shayan

Abstract:

The alkali-aggregate reaction (AAR) in concrete has a negative influence on the mechanical properties and durability of concrete. Confinement by carbon fiber reinforced polymer (CFRP) is an effective method of treatment for some AAR-affected elements. Eighteen reinforced columns affected by different levels of expansion due to AAR were confined using CFRP to evaluate the effect of expansion level on confinement efficiency. Strength and strain capacities (axial and circumferential) were measured using photogrammetry under uniaxial compressive loading to evaluate the efficiency of CFRP wrapping for the rehabilitation of affected columns. In relation to uniaxial compression capacity, the results indicated that the confinement of AAR-affected columns by one layer of CFRP is sufficient to reach and exceed the load capacity of unaffected sound columns. Parallel to the experimental study, finite element (FE) modeling using ATENA software was employed to predict the behavior of CFRP-confined damaged concrete and determine the possibility of using the model in a parametric study by simulating the number of CFRP layers. A comparison of the experimental results with the results of the theoretical models showed that FE modeling could be used for the prediction of the behavior of confined AAR-damaged concrete.

Keywords: ATENA, carbon fiber reinforced polymer (CFRP), confinement efficiency, finite element (FE)

Procedia PDF Downloads 64
1419 An Electrocardiography Deep Learning Model to Detect Atrial Fibrillation on Clinical Application

Authors: Jui-Chien Hsieh

Abstract:

Background:12-lead electrocardiography(ECG) is one of frequently-used tools to detect atrial fibrillation (AF), which might degenerate into life-threaten stroke, in clinical Practice. Based on this study, the AF detection by the clinically-used 12-lead ECG device has only 0.73~0.77 positive predictive value (ppv). Objective: It is on great demand to develop a new algorithm to improve the precision of AF detection using 12-lead ECG. Due to the progress on artificial intelligence (AI), we develop an ECG deep model that has the ability to recognize AF patterns and reduce false-positive errors. Methods: In this study, (1) 570-sample 12-lead ECG reports whose computer interpretation by the ECG device was AF were collected as the training dataset. The ECG reports were interpreted by 2 senior cardiologists, and confirmed that the precision of AF detection by the ECG device is 0.73.; (2) 88 12-lead ECG reports whose computer interpretation generated by the ECG device was AF were used as test dataset. Cardiologist confirmed that 68 cases of 88 reports were AF, and others were not AF. The precision of AF detection by ECG device is about 0.77; (3) A parallel 4-layer 1 dimensional convolutional neural network (CNN) was developed to identify AF based on limb-lead ECGs and chest-lead ECGs. Results: The results indicated that this model has better performance on AF detection than traditional computer interpretation of the ECG device in 88 test samples with 0.94 ppv, 0.98 sensitivity, 0.80 specificity. Conclusions: As compared to the clinical ECG device, this AI ECG model promotes the precision of AF detection from 0.77 to 0.94, and can generate impacts on clinical applications.

Keywords: 12-lead ECG, atrial fibrillation, deep learning, convolutional neural network

Procedia PDF Downloads 112
1418 Ion Beam Induced 2D Mesophase Patterning of Nanocrystallites in Polymer

Authors: Srutirekha Giri, Manoranjan Sahoo, Anuradha Das, Pravanjan Mallick, Biswajit Mallick

Abstract:

Ion Beam (IB) technique is a very powerful experimental technique for both material synthesis and material modifications. In this work, 3MeV proton beam was generated using the 3MV Tandem machine of the Institute of Physics, Bhubaneswar and extracted into air for the irradiation-induced modification purpose[1]. The polymeric material can be modeled for a three-phase system viz. crystalline(I), amorphous(II) and mesomorphic(III). So far, our knowledge is concerned. There are only few techniques reported for the synthesis of this third-phase(III) of polymer. The IB induced technique is one of them and has been reported very recently [2-4]. It was observed that by irradiating polyethylene terephthalate (PET) fiber at very low proton fluence, 10¹⁰ - 10¹² p/s, possess 2D mesophase structure. This was confirmed using X-ray diffraction technique. A low-intensity broad peak was observed at small angle of about 2θ =6º, when the fiber axis was mounted parallel to the X-ray direction. Such peak vanished in the diffraction spectrum when the fiber axis was mounted perpendicular to the beam direction. The appearance of this extra peak in a particular orientation confirms that the phase is 2-dimensionally oriented (mesophase). It is well known that the mesophase is a 2-dimensionally ordered structure but 3-dimensionally disordered. Again, the crystallite of the mesophase peak particle was measured about 3nm. The MeV proton-induced 2D mesophase patterning of nanocrystallites (3nm) of PET due to irradiation was observed within the above low fluence range and failed in high proton fluence. This is mainly due to the breaking of crystallites, radiation-induced thermal degradation, etc.

Keywords: Ion irradiation, mesophase, nanocrystallites, polymer

Procedia PDF Downloads 191
1417 Exhaustive Study of Essential Constraint Satisfaction Problem Techniques Based on N-Queens Problem

Authors: Md. Ahsan Ayub, Kazi A. Kalpoma, Humaira Tasnim Proma, Syed Mehrab Kabir, Rakib Ibna Hamid Chowdhury

Abstract:

Constraint Satisfaction Problem (CSP) is observed in various applications, i.e., scheduling problems, timetabling problems, assignment problems, etc. Researchers adopt a CSP technique to tackle a certain problem; however, each technique follows different approaches and ways to solve a problem network. In our exhaustive study, it has been possible to visualize the processes of essential CSP algorithms from a very concrete constraint satisfaction example, NQueens Problem, in order to possess a deep understanding about how a particular constraint satisfaction problem will be dealt with by our studied and implemented techniques. Besides, benchmark results - time vs. value of N in N-Queens - have been generated from our implemented approaches, which help understand at what factor each algorithm produces solutions; especially, in N-Queens puzzle. Thus, extended decisions can be made to instantiate a real life problem within CSP’s framework.

Keywords: arc consistency (AC), backjumping algorithm (BJ), backtracking algorithm (BT), constraint satisfaction problem (CSP), forward checking (FC), least constrained values (LCV), maintaining arc consistency (MAC), minimum remaining values (MRV), N-Queens problem

Procedia PDF Downloads 357
1416 Evolution of Bioactive Components of Prickly Pear Juice (Opuntia ficus indica) and Cocktails with Orange Juice

Authors: T. Hadj Sadok, R. Hattab Bey, K. Rebiha

Abstract:

The valuation of juice from prickly pear of Opuntia ficus indica inermis as cocktails appears an attractive alternative because of their nutritional intake and functional compound has anti-radical activity (polyphenols, vitamin C, carotenoids, Betalaines, fiber and minerals). The juice from the fruit pulp is characterized by a high pH 5.85 which makes it difficult for its conservation and preservation requires a thermal treatment at high temperatures (over 100 °C) harmful for bioactive constituents compared to juice orange more acidic and processed at temperatures < 100 °C. The valuation as fig cocktails-orange is particularly interesting thanks to the contribution of polyph2nols, fiber, vitamin C, reducing sugar (sweetener) and betalaine, minerals while allowing lower temperature processing to decrease pH. The heat treatment of these juices: orange alone or in cocktails showed that the antioxidant power decreases by 12% in presence of 30% of juice treated by the heat and of 28 and 32% in the presence of 10 and 20% juice which shows the effect prickly pear juice of Opuntia. During storage for 4 weeks the loss of vitamin C is 40 and 38% in the presence of 10 and 20% juice and 33% in the presence of 30% pear juice parallel, a treatment of stabilization by heat affects relatively the polyphenols rate which decreases from 10.5% to 30% in the cocktail, and 6.11-6.71pour cocktails at 10% and 20%. Vitamin C decreases to 12 to 24 % after a heat treatment at 85°C for 30 minutes respectively for the orange juice and pear juice; this reduction is higher when the juice is in the form of cocktails composed of 10 to 30 % pear juice.

Keywords: prickly pear juice, orange cocktail, polyphenol, Opuntia ficus indica, vitamin

Procedia PDF Downloads 373
1415 Housing Price Prediction Using Machine Learning Algorithms: The Case of Melbourne City, Australia

Authors: The Danh Phan

Abstract:

House price forecasting is a main topic in the real estate market research. Effective house price prediction models could not only allow home buyers and real estate agents to make better data-driven decisions but may also be beneficial for the property policymaking process. This study investigates the housing market by using machine learning techniques to analyze real historical house sale transactions in Australia. It seeks useful models which could be deployed as an application for house buyers and sellers. Data analytics show a high discrepancy between the house price in the most expensive suburbs and the most affordable suburbs in the city of Melbourne. In addition, experiments demonstrate that the combination of Stepwise and Support Vector Machine (SVM), based on the Mean Squared Error (MSE) measurement, consistently outperforms other models in terms of prediction accuracy.

Keywords: house price prediction, regression trees, neural network, support vector machine, stepwise

Procedia PDF Downloads 217
1414 Blended Intensive Programmes: A Way Forward to Promote Internationalization in Higher Education

Authors: Sonja Gögele, Petra Kletzenbauer

Abstract:

International strategies are ranked as one of the core activities in the development plans of Austrian universities. This has led to numerous promising activities in terms of internationalization (i.e. development of international degree programmes, increased staff and student mobility, and blended international projects). The latest innovative approach in terms of Erasmus+ are so called Blended Intensive Programmes (BIP) which combine jointly delivered teaching and learning elements of at least three participating ERASMUS universities in a virtual and short-term mobility setup. Students who participate in BIP can maintain their study plans at their home institution and include BIP as a parallel activity. This paper presents the experiences of this programme on the topic of sustainable computing hosted by the University of Applied Sciences FH JOANNEUM. By means of an online survey and face-to-face interviews with all stakeholders (20 students, 8 professors), the empirical study addresses the challenges of hosting an international blended learning programme (i.e. virtual phase and on-site intensive phase) and discusses the impact of such activities in terms of internationalization and Englishization. In this context, key roles are assigned to the development of future transnational and transdisciplinary curricula by considering innovative aspects for learning and teaching (i.e. virtual collaboration, research-based learning).

Keywords: internationalization, englishization, short-term mobility, international teaching and learning

Procedia PDF Downloads 117
1413 BEATRICE: A Low-Cost Manipulator Arm for an Educational Planetary Rover

Authors: T. Pakulski, L. Kryza, A. Linossier

Abstract:

The BEar Articulated TeleRobotic Inspection and Clasping Extremity is a lightweight, 5 DoF robotic manipulator for the Berlin Educational Assistant Rover (BEAR). BEAR is one of the educational planetary rovers developed under the Space Rover projects at the Chair of Space Technology of the Technische Universität Berlin. The projects serve to conduct research and train engineers by developing rovers for competitions like the European Rover Challenge and the DLR SpaceBot Cup. BEATRICE is the result of a cost-driven design process to deliver a simple but capable platform for a variety of competition tasks: object grasping and manipulation, inspection, instrument wielding and more. The manipulator’s simple mechatronic design, based on a combination of servomotors and stepper motors with planetary gearboxes, also makes it a practical tool for developing embedded control systems. The platform’s initial implementation relies on tele-operated control but is fully instrumented for future autonomous functionality. This paper describes BEATRICE’s development from its preliminary link model to its structural and mechatronic design, embedded control and AI and T. In parallel, it examines the influence of budget constraints and high personnel turnover commonly associated with student teams on the manipulator’s design. Finally, it comments on the utility of robot design projects for educating future engineers.

Keywords: education, low-cost, manipulator, robotics, rover

Procedia PDF Downloads 249
1412 Estimation of the Road Traffic Emissions and Dispersion in the Developing Countries Conditions

Authors: Hicham Gourgue, Ahmed Aharoune, Ahmed Ihlal

Abstract:

We present in this work our model of road traffic emissions (line sources) and dispersion of these emissions, named DISPOLSPEM (Dispersion of Poly Sources and Pollutants Emission Model). In its emission part, this model was designed to keep the consistent bottom-up and top-down approaches. It also allows to generate emission inventories from reduced input parameters being adapted to existing conditions in Morocco and in the other developing countries. While several simplifications are made, all the performance of the model results are kept. A further important advantage of the model is that it allows the uncertainty calculation and emission rate uncertainty according to each of the input parameters. In the dispersion part of the model, an improved line source model has been developed, implemented and tested against a reference solution. It provides improvement in accuracy over previous formulas of line source Gaussian plume model, without being too demanding in terms of computational resources. In the case study presented here, the biggest errors were associated with the ends of line source sections; these errors will be canceled by adjacent sections of line sources during the simulation of a road network. In cases where the wind is parallel to the source line, the use of the combination discretized source and analytical line source formulas minimizes remarkably the error. Because this combination is applied only for a small number of wind directions, it should not excessively increase the calculation time.

Keywords: air pollution, dispersion, emissions, line sources, road traffic, urban transport

Procedia PDF Downloads 435
1411 Numerical Investigation of Nanofluid Based Thermosyphon System

Authors: Kiran Kumar K., Ramesh Babu Bejjam, Atul Najan

Abstract:

A thermosyphon system is a heat transfer loop which operates on the basis of gravity and buoyancy forces. It guarantees a good reliability and low maintenance cost as it does not involve any mechanical pump. Therefore it can be used in many industrial applications such as refrigeration and air conditioning, electronic cooling, nuclear reactors, geothermal heat extraction, etc. But flow instabilities and loop configuration are the major problems in this system. Several previous researchers studied that stabilities can be suppressed by using nanofluids as loop fluid. In the present study a rectangular thermosyphon loop with end heat exchangers are considered for the study. This configuration is more appropriate for many practical applications such as solar water heater, geothermal heat extraction, etc. In the present work, steady-state analysis is carried out on thermosyphon loop with parallel flow coaxial heat exchangers at heat source and heat sink. In this loop nano fluid is considered as the loop fluid and water is considered as the external fluid in both hot and cold heat exchangers. For this analysis one-dimensional homogeneous model is developed. In this model, conservation equations like conservation of mass, momentum, energy are discretized using finite difference method. A computer code is written in MATLAB to simulate the flow in thermosyphon loop. A comparison in terms of heat transfer is made between water and nano fluid as working fluids in the loop.

Keywords: heat exchanger, heat transfer, nanofluid, thermosyphon loop

Procedia PDF Downloads 471
1410 HPPDFIM-HD: Transaction Distortion and Connected Perturbation Approach for Hierarchical Privacy Preserving Distributed Frequent Itemset Mining over Horizontally-Partitioned Dataset

Authors: Fuad Ali Mohammed Al-Yarimi

Abstract:

Many algorithms have been proposed to provide privacy preserving in data mining. These protocols are based on two main approaches named as: the perturbation approach and the Cryptographic approach. The first one is based on perturbation of the valuable information while the second one uses cryptographic techniques. The perturbation approach is much more efficient with reduced accuracy while the cryptographic approach can provide solutions with perfect accuracy. However, the cryptographic approach is a much slower method and requires considerable computation and communication overhead. In this paper, a new scalable protocol is proposed which combines the advantages of the perturbation and distortion along with cryptographic approach to perform privacy preserving in distributed frequent itemset mining on horizontally distributed data. Both the privacy and performance characteristics of the proposed protocol are studied empirically.

Keywords: anonymity data, data mining, distributed frequent itemset mining, gaussian perturbation, perturbation approach, privacy preserving data mining

Procedia PDF Downloads 499
1409 Approximation of Convex Set by Compactly Semidefinite Representable Set

Authors: Anusuya Ghosh, Vishnu Narayanan

Abstract:

The approximation of convex set by semidefinite representable set plays an important role in semidefinite programming, especially in modern convex optimization. To optimize a linear function over a convex set is a hard problem. But optimizing the linear function over the semidefinite representable set which approximates the convex set is easy to solve as there exists numerous efficient algorithms to solve semidefinite programming problems. So, our approximation technique is significant in optimization. We develop a technique to approximate any closed convex set, say K by compactly semidefinite representable set. Further we prove that there exists a sequence of compactly semidefinite representable sets which give tighter approximation of the closed convex set, K gradually. We discuss about the convergence of the sequence of compactly semidefinite representable sets to closed convex set K. The recession cone of K and the recession cone of the compactly semidefinite representable set are equal. So, we say that the sequence of compactly semidefinite representable sets converge strongly to the closed convex set. Thus, this approximation technique is very useful development in semidefinite programming.

Keywords: semidefinite programming, semidefinite representable set, compactly semidefinite representable set, approximation

Procedia PDF Downloads 375
1408 Autism Disease Detection Using Transfer Learning Techniques: Performance Comparison between Central Processing Unit vs. Graphics Processing Unit Functions for Neural Networks

Authors: Mst Shapna Akter, Hossain Shahriar

Abstract:

Neural network approaches are machine learning methods used in many domains, such as healthcare and cyber security. Neural networks are mostly known for dealing with image datasets. While training with the images, several fundamental mathematical operations are carried out in the Neural Network. The operation includes a number of algebraic and mathematical functions, including derivative, convolution, and matrix inversion and transposition. Such operations require higher processing power than is typically needed for computer usage. Central Processing Unit (CPU) is not appropriate for a large image size of the dataset as it is built with serial processing. While Graphics Processing Unit (GPU) has parallel processing capabilities and, therefore, has higher speed. This paper uses advanced Neural Network techniques such as VGG16, Resnet50, Densenet, Inceptionv3, Xception, Mobilenet, XGBOOST-VGG16, and our proposed models to compare CPU and GPU resources. A system for classifying autism disease using face images of an autistic and non-autistic child was used to compare performance during testing. We used evaluation matrices such as Accuracy, F1 score, Precision, Recall, and Execution time. It has been observed that GPU runs faster than the CPU in all tests performed. Moreover, the performance of the Neural Network models in terms of accuracy increases on GPU compared to CPU.

Keywords: autism disease, neural network, CPU, GPU, transfer learning

Procedia PDF Downloads 107
1407 Novel Correlations for P-Substituted Phenols in NMR Spectroscopy

Authors: Khodzhaberdi Allaberdiev

Abstract:

Substituted phenols are widely used for the synthesis of advanced polycondensation polymers. In terms of the structure regularity and practical value of obtained polymers are of special interest the p-substituted phenols. The lanthanide induced shifts (LIS) of the aromatic ring and the OH protons by addition Eu(fod)3 to various p-substituted phenols in CDCL3 solvent were measured Nuclear Magnetic Resonance spectroscopy. A linear relationship has been observed between the LIS of protons (∆=δcomplex –δsubstrate) and Eu(fod)3/substrate molar ratios. The LIS protons of the investigated phenols decreases in the following order: ОН > ortho > meta. The LIS of these protons also depends on both steric and electronic effects of p-substituents. The effect on the LIS of protons steric hindrance of substituents by way of example p-substituted alkyl phenols was studied. Alkyl phenols exhibit pronounced europium- induced shifts, their sensitivity increasing in the order: CH3 > C2H5 > sym-C5H11 > tert-C5H11 > tert-C4H9, i.e. in parallel with decreasing steric hindrance. The influence steric hindrance p-substituents of phenols on the LIS of protons in sequence following decreases: OH> meta >ortho. Contrary to the expectations, it is found that the LIS of the ortho protons an excellent linear correlation with meta-substituent constants, σm for 14 p-substituted phenols: ∆H2, 6=8.165-9.896 σm (r2=0,999). Moreover, a linear correlation between the LIS of the ortho protons and ionization constants, РКa of p-substituted phenols has been revealed. Similarly, the linear relationships for the LIS of the meta and the OH protons were obtained. Use the LIS of the phenolic hydroxyl groups for linear relationships is necessary with care, because of the signal broadening of the OH protons. New constants may be determinate with unusual case by this approach.

Keywords: novel correlations, NMR spectroscopy, phenols, shift reagent

Procedia PDF Downloads 299
1406 Deep Learning in Chest Computed Tomography to Differentiate COVID-19 from Influenza

Authors: Hongmei Wang, Ziyun Xiang, Ying liu, Li Yu, Dongsheng Yue

Abstract:

Intro: The COVID-19 (Corona Virus Disease 2019) has greatly changed the global economic, political and financial ecology. The mutation of the coronavirus in the UK in December 2020 has brought new panic to the world. Deep learning was performed on Chest Computed tomography (CT) of COVID-19 and Influenza and describes their characteristics. The predominant features of COVID-19 pneumonia was ground-glass opacification, followed by consolidation. Lesion density: most lesions appear as ground-glass shadows, and some lesions coexist with solid lesions. Lesion distribution: the focus is mainly on the dorsal side of the periphery of the lung, with the lower lobe of the lungs as the focus, and it is often close to the pleura. Other features it has are grid-like shadows in ground glass lesions, thickening signs of diseased vessels, air bronchi signs and halo signs. The severe disease involves whole bilateral lungs, showing white lung signs, air bronchograms can be seen, and there can be a small amount of pleural effusion in the bilateral chest cavity. At the same time, this year's flu season could be near its peak after surging throughout the United States for months. Chest CT for Influenza infection is characterized by focal ground glass shadows in the lungs, with or without patchy consolidation, and bronchiole air bronchograms are visible in the concentration. There are patchy ground-glass shadows, consolidation, air bronchus signs, mosaic lung perfusion, etc. The lesions are mostly fused, which is prominent near the hilar and two lungs. Grid-like shadows and small patchy ground-glass shadows are visible. Deep neural networks have great potential in image analysis and diagnosis that traditional machine learning algorithms do not. Method: Aiming at the two major infectious diseases COVID-19 and influenza, which are currently circulating in the world, the chest CT of patients with two infectious diseases is classified and diagnosed using deep learning algorithms. The residual network is proposed to solve the problem of network degradation when there are too many hidden layers in a deep neural network (DNN). The proposed deep residual system (ResNet) is a milestone in the history of the Convolutional neural network (CNN) images, which solves the problem of difficult training of deep CNN models. Many visual tasks can get excellent results through fine-tuning ResNet. The pre-trained convolutional neural network ResNet is introduced as a feature extractor, eliminating the need to design complex models and time-consuming training. Fastai is based on Pytorch, packaging best practices for in-depth learning strategies, and finding the best way to handle diagnoses issues. Based on the one-cycle approach of the Fastai algorithm, the classification diagnosis of lung CT for two infectious diseases is realized, and a higher recognition rate is obtained. Results: A deep learning model was developed to efficiently identify the differences between COVID-19 and influenza using chest CT.

Keywords: COVID-19, Fastai, influenza, transfer network

Procedia PDF Downloads 137
1405 Neotectonic Features of the Fethiye-Burdur Fault Zone between Kozluca and Burdur, SW Anatolia, Turkey

Authors: Berkant Coşkuner, Rahmi Aksoy

Abstract:

The aim of this study is to present some preliminary stratigraphic and structural evidence for the Fethiye-Burdur fault zone between Kozluca and Burdur. The Fethiye-Burdur fault zone, the easternmost extension of the west Anatolian extensional province, extends from the Gulf of Fethiye northeastward through Burdur, a distance of about 300 km. The research area is located in the Burdur segment of the fault zone. Here, the fault zone includes several parallel to subparallel fault branching and en-echelon faults that lie within a linear belt, as much as 20 km in width. The direction of movement in the fault zone has been oblique-slip in the left lateral sense. The basement of the study area consists of the Triassic-Eocene Lycian Nappes, the Eocene-Oligocene molasse sediments and the lower Miocene marine rocks. The Burdur basin contains two basin infills. The ancient and deformed basin fill is composed of lacustrine sediments of the upper Miocene-lower Pliocene age. The younger and undeformed basin fill comprises Plio-Quaternary alluvial fan and recent basin-floor deposits and unconformably overlies the ancient basin infill. The Burdur basin is bounded by the NE-SW trending, left lateral oblique-slip normal faults, the Karakent fault on the northwest and the Burdur fault on the southeast. These faults played a key role in the development of the Burdur basin as a pull-apart basin.

Keywords: Burdur basin, Fethiye-Burdur fault zone, left lateral oblique-slip fault, Western Anatolia

Procedia PDF Downloads 405
1404 Multi-Objective Variable Neighborhood Search Algorithm to Solving Scheduling Problem with Transportation Times

Authors: Majid Khalili

Abstract:

This paper deals with a bi-objective hybrid no-wait flowshop scheduling problem minimizing the makespan and total weighted tardiness, in which we consider transportation times between stages. Obtaining an optimal solution for this type of complex, large-sized problem in reasonable computational time by using traditional approaches and optimization tools is extremely difficult. This paper presents a new multi-objective variable neighborhood algorithm (MOVNS). A set of experimental instances are carried out to evaluate the algorithm by advanced multi-objective performance measures. The algorithm is carefully evaluated for its performance against available algorithm by means of multi-objective performance measures and statistical tools. The related results show that a variant of our proposed MOVNS provides sound performance comparing with other algorithms.

Keywords: no-wait hybrid flowshop scheduling; multi-objective variable neighborhood algorithm; makespan; total weighted tardiness

Procedia PDF Downloads 410
1403 Neural Networks and Genetic Algorithms Approach for Word Correction and Prediction

Authors: Rodrigo S. Fonseca, Antônio C. P. Veiga

Abstract:

Aiming at helping people with some movement limitation that makes typing and communication difficult, there is a need to customize an assistive tool with a learning environment that helps the user in order to optimize text input, identifying the error and providing the correction and possibilities of choice in the Portuguese language. The work presents an Orthographic and Grammatical System that can be incorporated into writing environments, improving and facilitating the use of an alphanumeric keyboard, using a prototype built using a genetic algorithm in addition to carrying out the prediction, which can occur based on the quantity and position of the inserted letters and even placement in the sentence, ensuring the sequence of ideas using a Long Short Term Memory (LSTM) neural network. The prototype optimizes data entry, being a component of assistive technology for the textual formulation, detecting errors, seeking solutions and informing the user of accurate predictions quickly and effectively through machine learning.

Keywords: genetic algorithm, neural networks, word prediction, machine learning

Procedia PDF Downloads 185
1402 Relay Mining: Verifiable Multi-Tenant Distributed Rate Limiting

Authors: Daniel Olshansky, Ramiro Rodrıguez Colmeiro

Abstract:

Relay Mining presents a scalable solution employing probabilistic mechanisms and crypto-economic incentives to estimate RPC volume usage, facilitating decentralized multitenant rate limiting. Network traffic from individual applications can be concurrently serviced by multiple RPC service providers, with costs, rewards, and rate limiting governed by a native cryptocurrency on a distributed ledger. Building upon established research in token bucket algorithms and distributed rate-limiting penalty models, our approach harnesses a feedback loop control mechanism to adjust the difficulty of mining relay rewards, dynamically scaling with network usage growth. By leveraging crypto-economic incentives, we reduce coordination overhead costs and introduce a mechanism for providing RPC services that are both geopolitically and geographically distributed.

Keywords: remote procedure call, crypto-economic, commit-reveal, decentralization, scalability, blockchain, rate limiting, token bucket

Procedia PDF Downloads 50
1401 Nations in Labour: Incorporating National Narratives in Sociological Models of Cultural Labour

Authors: Anna Lytvynova

Abstract:

This essay presents labour as a performatively national phenomenon from a cultural perspective. Considering Engels’ proposition of labour as the epicentre of development of social structures and communities, it theorizes the formation and sustainment of group identities through labour identities. Taking labour in the cultural sector as the starting point case study, the essay further enunciates such labour and labour identity as a form of engaged citizenship. In doing so, this piece hopes to arrive at a potential contemporary understanding of labour as having a central and dynamic role in cultural organization and citizenship. A parallel goal is to de-link sociological models of cultural labor from narratives of art and culture as something that stands separate from the 'real world' and the economy and exists in precarity. Combining discourse from cultural sociology, performance studies, and economics and grounding it in historical archive, the essay makes a primarily discursive theoretical contribution. Taking North American theatre organizations as the exemplifying starting point, this project positions cultural workers not solely as workers in a professional industry but as active citizen-subjects who are deeply involved in their society’s democratic processes. The resulting discourse can be used to shape more effective labour policies, as well as help art and cultural organizations find more effective organizational structures to engage the arts in the economic, political, and social spheres.

Keywords: arts labour, cultural sociology, national identity, performativity

Procedia PDF Downloads 122
1400 Design of an Automated Deep Learning Recurrent Neural Networks System Integrated with IoT for Anomaly Detection in Residential Electric Vehicle Charging in Smart Cities

Authors: Wanchalerm Patanacharoenwong, Panaya Sudta, Prachya Bumrungkun

Abstract:

The paper focuses on the development of a system that combines Internet of Things (IoT) technologies and deep learning algorithms for anomaly detection in residential Electric Vehicle (EV) charging in smart cities. With the increasing number of EVs, ensuring efficient and reliable charging systems has become crucial. The aim of this research is to develop an integrated IoT and deep learning system for detecting anomalies in residential EV charging and enhancing EV load profiling and event detection in smart cities. This approach utilizes IoT devices equipped with infrared cameras to collect thermal images and household EV charging profiles from the database of Thailand utility, subsequently transmitting this data to a cloud database for comprehensive analysis. The methodology includes the use of advanced deep learning techniques such as Recurrent Neural Networks (RNN) and Long Short-Term Memory (LSTM) algorithms. IoT devices equipped with infrared cameras are used to collect thermal images and EV charging profiles. The data is transmitted to a cloud database for comprehensive analysis. The researchers also utilize feature-based Gaussian mixture models for EV load profiling and event detection. Moreover, the research findings demonstrate the effectiveness of the developed system in detecting anomalies and critical profiles in EV charging behavior. The system provides timely alarms to users regarding potential issues and categorizes the severity of detected problems based on a health index for each charging device. The system also outperforms existing models in event detection accuracy. This research contributes to the field by showcasing the potential of integrating IoT and deep learning techniques in managing residential EV charging in smart cities. The system ensures operational safety and efficiency while also promoting sustainable energy management. The data is collected using IoT devices equipped with infrared cameras and is stored in a cloud database for analysis. The collected data is then analyzed using RNN, LSTM, and feature-based Gaussian mixture models. The approach includes both EV load profiling and event detection, utilizing a feature-based Gaussian mixture model. This comprehensive method aids in identifying unique power consumption patterns among EV owners and outperforms existing models in event detection accuracy. In summary, the research concludes that integrating IoT and deep learning techniques can effectively detect anomalies in residential EV charging and enhance EV load profiling and event detection accuracy. The developed system ensures operational safety and efficiency, contributing to sustainable energy management in smart cities.

Keywords: cloud computing framework, recurrent neural networks, long short-term memory, Iot, EV charging, smart grids

Procedia PDF Downloads 57