Search results for: bare machine computing
1476 Computing Transition Intensity Using Time-Homogeneous Markov Jump Process: Case of South African HIV/AIDS Disposition
Authors: A. Bayaga
Abstract:
This research provides a technical account of estimating Transition Probability using Time-homogeneous Markov Jump Process applying by South African HIV/AIDS data from the Statistics South Africa. It employs Maximum Likelihood Estimator (MLE) model to explore the possible influence of Transition Probability of mortality cases in which case the data was based on actual Statistics South Africa. This was conducted via an integrated demographic and epidemiological model of South African HIV/AIDS epidemic. The model was fitted to age-specific HIV prevalence data and recorded death data using MLE model. Though the previous model results suggest HIV in South Africa has declined and AIDS mortality rates have declined since 2002 – 2013, in contrast, our results differ evidently with the generally accepted HIV models (Spectrum/EPP and ASSA2008) in South Africa. However, there is the need for supplementary research to be conducted to enhance the demographic parameters in the model and as well apply it to each of the nine (9) provinces of South Africa.Keywords: AIDS mortality rates, epidemiological model, time-homogeneous markov jump process, transition probability, statistics South Africa
Procedia PDF Downloads 4971475 A Nonlinear Feature Selection Method for Hyperspectral Image Classification
Authors: Pei-Jyun Hsieh, Cheng-Hsuan Li, Bor-Chen Kuo
Abstract:
For hyperspectral image classification, feature reduction is an important pre-processing for avoiding the Hughes phenomena due to the difficulty for collecting training samples. Hence, lots of researches developed feature selection methods such as F-score, HSIC (Hilbert-Schmidt Independence Criterion), and etc., to improve hyperspectral image classification. However, most of them only consider the class separability in the original space, i.e., a linear class separability. In this study, we proposed a nonlinear class separability measure based on kernel trick for selecting an appropriate feature subset. The proposed nonlinear class separability was formed by a generalized RBF kernel with different bandwidths with respect to different features. Moreover, it considered the within-class separability and the between-class separability. A genetic algorithm was applied to tune these bandwidths such that the smallest with-class separability and the largest between-class separability simultaneously. This indicates the corresponding feature space is more suitable for classification. In addition, the corresponding nonlinear classification boundary can separate classes very well. These optimal bandwidths also show the importance of bands for hyperspectral image classification. The reciprocals of these bandwidths can be viewed as weights of bands. The smaller bandwidth, the larger weight of the band, and the more importance for classification. Hence, the descending order of the reciprocals of the bands gives an order for selecting the appropriate feature subsets. In the experiments, three hyperspectral image data sets, the Indian Pine Site data set, the PAVIA data set, and the Salinas A data set, were used to demonstrate the selected feature subsets by the proposed nonlinear feature selection method are more appropriate for hyperspectral image classification. Only ten percent of samples were randomly selected to form the training dataset. All non-background samples were used to form the testing dataset. The support vector machine was applied to classify these testing samples based on selected feature subsets. According to the experiments on the Indian Pine Site data set with 220 bands, the highest accuracies by applying the proposed method, F-score, and HSIC are 0.8795, 0.8795, and 0.87404, respectively. However, the proposed method selects 158 features. F-score and HSIC select 168 features and 217 features, respectively. Moreover, the classification accuracies increase dramatically only using first few features. The classification accuracies with respect to feature subsets of 10 features, 20 features, 50 features, and 110 features are 0.69587, 0.7348, 0.79217, and 0.84164, respectively. Furthermore, only using half selected features (110 features) of the proposed method, the corresponding classification accuracy (0.84168) is approximate to the highest classification accuracy, 0.8795. For other two hyperspectral image data sets, the PAVIA data set and Salinas A data set, we can obtain the similar results. These results illustrate our proposed method can efficiently find feature subsets to improve hyperspectral image classification. One can apply the proposed method to determine the suitable feature subset first according to specific purposes. Then researchers can only use the corresponding sensors to obtain the hyperspectral image and classify the samples. This can not only improve the classification performance but also reduce the cost for obtaining hyperspectral images.Keywords: hyperspectral image classification, nonlinear feature selection, kernel trick, support vector machine
Procedia PDF Downloads 2651474 A Heart Arrhythmia Prediction Using Machine Learning’s Classification Approach and the Concept of Data Mining
Authors: Roshani S. Golhar, Neerajkumar S. Sathawane, Snehal Dongre
Abstract:
Background and objectives: As the, cardiovascular illnesses increasing and becoming cause of mortality worldwide, killing around lot of people each year. Arrhythmia is a type of cardiac illness characterized by a change in the linearity of the heartbeat. The goal of this study is to develop novel deep learning algorithms for successfully interpreting arrhythmia using a single second segment. Because the ECG signal indicates unique electrical heart activity across time, considerable changes between time intervals are detected. Such variances, as well as the limited number of learning data available for each arrhythmia, make standard learning methods difficult, and so impede its exaggeration. Conclusions: The proposed method was able to outperform several state-of-the-art methods. Also proposed technique is an effective and convenient approach to deep learning for heartbeat interpretation, that could be probably used in real-time healthcare monitoring systemsKeywords: electrocardiogram, ECG classification, neural networks, convolutional neural networks, portable document format
Procedia PDF Downloads 721473 Diagnosis of the Lubrification System of a Gas Turbine Using the Adaptive Neuro-Fuzzy Inference System
Authors: H. Mahdjoub, B. Hamaidi, B. Zerouali, S. Rouabhia
Abstract:
The issue of fault detection and diagnosis (FDD) has gained widespread industrial interest in process condition monitoring applications. Accordingly, the use of neuro-fuzzy technic seems very promising. This paper treats a diagnosis modeling a strategic equipment of an industrial installation. We propose a diagnostic tool based on adaptive neuro-fuzzy inference system (ANFIS). The neuro-fuzzy network provides an abductive diagnosis. Moreover, it takes into account the uncertainties on the maintenance knowledge by giving a fuzzy characterization of each cause. This work was carried out with real data of a lubrication circuit from the gas turbine. The machine of interest is a gas turbine placed in a gas compressor station at South Industrial Centre (SIC Hassi Messaoud Ouargla, Algeria). We have defined the zones of good and bad functioning, and the results are presented to demonstrate the advantages of the proposed method.Keywords: fault detection and diagnosis, lubrication system, turbine, ANFIS, training, pattern recognition
Procedia PDF Downloads 4911472 A Comparison of Single of Decision Tree, Decision Tree Forest and Group Method of Data Handling to Evaluate the Surface Roughness in Machining Process
Authors: S. Ghorbani, N. I. Polushin
Abstract:
The machinability of workpieces (AISI 1045 Steel, AA2024 aluminum alloy, A48-class30 gray cast iron) in turning operation has been carried out using different types of cutting tool (conventional, cutting tool with holes in toolholder and cutting tool filled up with composite material) under dry conditions on a turning machine at different stages of spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev), depth of cut (0.05-0.15 mm) and tool overhang (41-65 mm). Experimentation was performed as per Taguchi’s orthogonal array. To evaluate the relative importance of factors affecting surface roughness the single decision tree (SDT), Decision tree forest (DTF) and Group method of data handling (GMDH) were applied.Keywords: decision tree forest, GMDH, surface roughness, Taguchi method, turning process
Procedia PDF Downloads 4451471 A Survey of Response Generation of Dialogue Systems
Authors: Yifan Fan, Xudong Luo, Pingping Lin
Abstract:
An essential task in the field of artificial intelligence is to allow computers to interact with people through natural language. Therefore, researches such as virtual assistants and dialogue systems have received widespread attention from industry and academia. The response generation plays a crucial role in dialogue systems, so to push forward the research on this topic, this paper surveys various methods for response generation. We sort out these methods into three categories. First one includes finite state machine methods, framework methods, and instance methods. The second contains full-text indexing methods, ontology methods, vast knowledge base method, and some other methods. The third covers retrieval methods and generative methods. We also discuss some hybrid methods based knowledge and deep learning. We compare their disadvantages and advantages and point out in which ways these studies can be improved further. Our discussion covers some studies published in leading conferences such as IJCAI and AAAI in recent years.Keywords: deep learning, generative, knowledge, response generation, retrieval
Procedia PDF Downloads 1341470 Optimization of Process Parameters for Rotary Electro Discharge Machining Using EN31 Tool Steel: Present and Future Scope
Authors: Goutam Dubey, Varun Dutta
Abstract:
In the present study, rotary-electro discharge machining of EN31 tool steel has been carried out using a pure copper electrode. Various response variables such as Material Removal Rate (MRR), Tool Wear Rate (TWR), and Machining Rate (MR) have been studied against the selected process variables. The selected process variables were peak current (I), voltage (V), duty cycle, and electrode rotation (N). EN31 Tool Steel is hardened, high carbon steel which increases its hardness and reduces its machinability. Reduced machinability means it not economical to use conventional methods to machine EN31 Tool Steel. So, non-conventional methods play an important role in machining of such materials.Keywords: electric discharge machining, EDM, tool steel, tool wear rate, optimization techniques
Procedia PDF Downloads 2041469 The Design, Control and Dynamic Performance of an Interior Permanent Magnet Synchronous Generator for Wind Power System
Authors: Olusegun Solomon
Abstract:
This paper describes the concept for the design and maximum power point tracking control for an interior permanent magnet synchronous generator wind turbine system. Two design concepts are compared to outline the effect of magnet design on the performance of the interior permanent magnet synchronous generator. An approximate model that includes the effect of core losses has been developed for the machine to simulate the dynamic performance of the wind energy system. An algorithm for Maximum Power Point Tracking control is included to describe the process for maximum power extraction.Keywords: permanent magnet synchronous generator, wind power system, wind turbine
Procedia PDF Downloads 2241468 Improving Overall Equipment Effectiveness of CNC-VMC by Implementing Kobetsu Kaizen
Authors: Nakul Agrawal, Y. M. Puri
Abstract:
TPM methodology is a proven approach to increase Overall Equipment Effectiveness (OEE) of machine. OEE is an established method to monitor and improve the effectiveness of manufacturing process. OEE is a product of equipment availability, performance efficiency and quality performance of manufacturing operations. The paper presents a project work for improving OEE of CNC-VMC in a manufacturing industry with the help of TPM tools Kaizen and Autonomous Maintenance. The aim of paper is to enhance OEE by minimizing the breakdown and re-work, increase availability, performance and quality. The calculated OEE of bottle necking machines for 4 months is lower of 53.3%. Root Cause Analysis RCA tools like fishbone diagram, Pareto chart are used for determining the reasons behind low OEE. While Tool like Why-Why analysis is use for determining the basis reasons for low OEE. Tools like Kaizen and Autonomous Maintenance are effectively implemented on CNC-VMC which eliminate the causes of breakdown and prevent from reoccurring. The result obtains from approach shows that OEE of CNC-VMC improved from 53.3% to 73.7% which saves an average sum of Rs.3, 19,000.Keywords: OEE, TPM, Kaizen, CNC-VMC, why-why analysis, RCA
Procedia PDF Downloads 3951467 The Urban Stray Animal Identification Management System Based on YOLOv5
Authors: Chen Xi, LIU Xuebin, Kuan Sinman, LI Haofeng, Huang Hongming, Zeng Chengyu, Lao Xuerui
Abstract:
Stray animals are on the rise in mainland China's cities. There are legal reasons for this, namely the lack of protection for domestic pets in mainland China, where only wildlife protection laws exist. At a social level, the ease with which families adopt pets and the lack of a social view of animal nature have led to the frequent abandonment and loss of stray animals. If left unmanaged, conflicts between humans and stray animals can also increase. This project provides an inexpensive and widely applicable management tool for urban management by collecting videos and pictures of stray animals captured by surveillance or transmitted by humans and using artificial intelligence technology (mainly using Yolov5 recognition technology) and recording and managing them in a database.Keywords: urban planning, urban governance, artificial intelligence, convolutional neural network, machine vision
Procedia PDF Downloads 991466 HcDD: The Hybrid Combination of Disk Drives in Active Storage Systems
Authors: Shu Yin, Zhiyang Ding, Jianzhong Huang, Xiaojun Ruan, Xiaomin Zhu, Xiao Qin
Abstract:
Since large-scale and data-intensive applications have been widely deployed, there is a growing demand for high-performance storage systems to support data-intensive applications. Compared with traditional storage systems, next-generation systems will embrace dedicated processor to reduce computational load of host machines and will have hybrid combinations of different storage devices. The advent of flash- memory-based solid state disk has become a critical role in revolutionizing the storage world. However, instead of simply replacing the traditional magnetic hard disk with the solid state disk, it is believed that finding a complementary approach to corporate both of them is more challenging and attractive. This paper explores an idea of active storage, an emerging new storage configuration, in terms of the architecture and design, the parallel processing capability, the cooperation of other machines in cluster computing environment, and a disk configuration, the hybrid combination of different types of disk drives. Experimental results indicate that the proposed HcDD achieves better I/O performance and longer storage system lifespan.Keywords: arallel storage system, hybrid storage system, data inten- sive, solid state disks, reliability
Procedia PDF Downloads 4501465 Modeling and Simulation of Flow Shop Scheduling Problem through Petri Net Tools
Authors: Joselito Medina Marin, Norberto Hernández Romero, Juan Carlos Seck Tuoh Mora, Erick S. Martinez Gomez
Abstract:
The Flow Shop Scheduling Problem (FSSP) is a typical problem that is faced by production planning managers in Flexible Manufacturing Systems (FMS). This problem consists in finding the optimal scheduling to carry out a set of jobs, which are processed in a set of machines or shared resources. Moreover, all the jobs are processed in the same machine sequence. As in all the scheduling problems, the makespan can be obtained by drawing the Gantt chart according to the operations order, among other alternatives. On this way, an FMS presenting the FSSP can be modeled by Petri nets (PNs), which are a powerful tool that has been used to model and analyze discrete event systems. Then, the makespan can be obtained by simulating the PN through the token game animation and incidence matrix. In this work, we present an adaptive PN to obtain the makespan of FSSP by applying PN analytical tools.Keywords: flow-shop scheduling problem, makespan, Petri nets, state equation
Procedia PDF Downloads 2991464 Forecasting Thermal Energy Demand in District Heating and Cooling Systems Using Long Short-Term Memory Neural Networks
Authors: Kostas Kouvaris, Anastasia Eleftheriou, Georgios A. Sarantitis, Apostolos Chondronasios
Abstract:
To achieve the objective of almost zero carbon energy solutions by 2050, the EU needs to accelerate the development of integrated, highly efficient and environmentally friendly solutions. In this direction, district heating and cooling (DHC) emerges as a viable and more efficient alternative to conventional, decentralized heating and cooling systems, enabling a combination of more efficient renewable and competitive energy supplies. In this paper, we develop a forecasting tool for near real-time local weather and thermal energy demand predictions for an entire DHC network. In this fashion, we are able to extend the functionality and to improve the energy efficiency of the DHC network by predicting and adjusting the heat load that is distributed from the heat generation plant to the connected buildings by the heat pipe network. Two case-studies are considered; one for Vransko, Slovenia and one for Montpellier, France. The data consists of i) local weather data, such as humidity, temperature, and precipitation, ii) weather forecast data, such as the outdoor temperature and iii) DHC operational parameters, such as the mass flow rate, supply and return temperature. The external temperature is found to be the most important energy-related variable for space conditioning, and thus it is used as an external parameter for the energy demand models. For the development of the forecasting tool, we use state-of-the-art deep neural networks and more specifically, recurrent networks with long-short-term memory cells, which are able to capture complex non-linear relations among temporal variables. Firstly, we develop models to forecast outdoor temperatures for the next 24 hours using local weather data for each case-study. Subsequently, we develop models to forecast thermal demand for the same period, taking under consideration past energy demand values as well as the predicted temperature values from the weather forecasting models. The contributions to the scientific and industrial community are three-fold, and the empirical results are highly encouraging. First, we are able to predict future thermal demand levels for the two locations under consideration with minimal errors. Second, we examine the impact of the outdoor temperature on the predictive ability of the models and how the accuracy of the energy demand forecasts decreases with the forecast horizon. Third, we extend the relevant literature with a new dataset of thermal demand and examine the performance and applicability of machine learning techniques to solve real-world problems. Overall, the solution proposed in this paper is in accordance with EU targets, providing an automated smart energy management system, decreasing human errors and reducing excessive energy production.Keywords: machine learning, LSTMs, district heating and cooling system, thermal demand
Procedia PDF Downloads 1431463 Preliminary Proposal to Use Adaptive Computer Games in the Virtual Rehabilitation Therapy
Authors: Mamoun S. Ideis, Zein Salah
Abstract:
Virtual Rehabilitation (VR) refers to using Virtual Reality’s hardware and simulations as means of exercising tools to rehabilitate patients in need. These patients will undergo their treatment exercises while playing different computer games, which helps achieve greater motivation for patients undergoing their therapeutic exercises. Virtual Rehabilitation systems adopt computer games as part of the treatment therapy. In this paper, we present a preliminary proposal to using adaptive computer games in Virtual Rehabilitation therapy. We also present some tips in designing those adaptive computer games by using different machine learning algorithms in order to create a personalized experience for each patient, which in turn, increases the potential benefits of the treatment that each patient receives. Furthermore, we propose a method of comparing the results of treatment using the adaptive computer games with the results of using static and classical computer games.Keywords: virtual rehabilitation, physiotherapy, adaptive computer games, post-stroke, game design
Procedia PDF Downloads 2991462 Magnetic Solid-Phase Separation of Uranium from Aqueous Solution Using High Capacity Diethylenetriamine Tethered Magnetic Adsorbents
Authors: Amesh P, Suneesh A S, Venkatesan K A
Abstract:
The magnetic solid-phase extraction is a relatively new method among the other solid-phase extraction techniques for the separating of metal ions from aqueous solutions, such as mine water and groundwater, contaminated wastes, etc. However, the bare magnetic particles (Fe3O4) exhibit poor selectivity due to the absence of target-specific functional groups for sequestering the metal ions. The selectivity of these magnetic particles can be remarkably improved by covalently tethering the task-specific ligands on magnetic surfaces. The magnetic particles offer a number of advantages such as quick phase separation aided by the external magnetic field. As a result, the solid adsorbent can be prepared with the particle size ranging from a few micrometers to the nanometer, which again offers the advantages such as enhanced kinetics of extraction, higher extraction capacity, etc. Conventionally, the magnetite (Fe3O4) particles were prepared by the hydrolysis and co-precipitation of ferrous and ferric salts in aqueous ammonia solution. Since the covalent linking of task-specific functionalities on Fe3O4 was difficult, and it is also susceptible to redox reaction in the presence of acid or alkali, it is necessary to modify the surface of Fe3O4 by silica coating. This silica coating is usually carried out by hydrolysis and condensation of tetraethyl orthosilicate over the surface of magnetite to yield a thin layer of silica-coated magnetite particles. Since the silica-coated magnetite particles amenable for further surface modification, it can be reacted with task-specific functional groups to obtain the functionalized magnetic particles. The surface area exhibited by such magnetic particles usually falls in the range of 50 to 150 m2.g-1, which offer advantage such as quick phase separation, as compared to the other solid-phase extraction systems. In addition, the magnetic (Fe3O4) particles covalently linked on mesoporous silica matrix (MCM-41) and task-specific ligands offer further advantages in terms of extraction kinetics, high stability, longer reusable cycles, and metal extraction capacity, due to the large surface area, ample porosity and enhanced number of functional groups per unit area on these adsorbents. In view of this, the present paper deals with the synthesis of uranium specific diethylenetriamine ligand (DETA) ligand anchored on silica-coated magnetite (Fe-DETA) as well as on magnetic mesoporous silica (MCM-Fe-DETA) and studies on the extraction of uranium from aqueous solution spiked with uranium to mimic the mine water or groundwater contaminated with uranium. The synthesized solid-phase adsorbents were characterized by FT-IR, Raman, TG-DTA, XRD, and SEM. The extraction behavior of uranium on the solid-phase was studied under several conditions like the effect of pH, initial concentration of uranium, rate of extraction and its variation with pH and initial concentration of uranium, effect of interference ions like CO32-, Na+, Fe+2, Ni+2, and Cr+3, etc. The maximum extraction capacity of 233 mg.g-1 was obtained for Fe-DETA, and a huge capacity of 1047 mg.g-1 was obtained for MCM-Fe-DETA. The mechanism of extraction, speciation of uranium, extraction studies, reusability, and the other results obtained in the present study suggests Fe-DETA and MCM-Fe-DETA are the potential candidates for the extraction of uranium from mine water, and groundwater.Keywords: diethylenetriamine, magnetic mesoporous silica, magnetic solid-phase extraction, uranium extraction, wastewater treatment
Procedia PDF Downloads 1711461 Enabling Citizen Participation in Urban Planning through Geospatial Gamification
Authors: Joanne F. Hayek
Abstract:
This study explores the use of gamification to promote citizen e-participation in urban planning. The research departs from a case study: the ‘Shape Your City’ web app designed and programmed by the author and presented as part of the 2021 Dubai Design Week to engage citizens in the co-creation of the future of their city through a gamified experience. The paper documents the design and development methodology of the web app and concludes with the findings of its pilot release. The case study explores the use of mobile interactive mapping, real-time data visualization, augmented reality, and machine learning as tools to enable co-planning. The paper also details the user interface design strategies employed to integrate complex cross-sector e-planning systems and make them accessible to citizens.Keywords: gamification, co-planning, citizen e-participation, mobile interactive mapping, real-time data visualization
Procedia PDF Downloads 1421460 Design of Control System Based On PLC and Kingview for Granulation Product Line
Authors: Mei-Feng, Yude-Fan, Min-Zhu
Abstract:
Based on PLC and kingview, this paper proposed a method that designed a set of the automatic control system according to the craft flow and demands for granulation product line. There were the main station and subordinate stations in PLC which were communicated by PROFIBUS network. PLC and computer were communicated by Ethernet network. The conversation function between human and machine was realized by kingview software, including actual time craft flows, historic report curves and product report forms. The construction of the control system, hardware collocation and software design were introduced. Besides these, PROFIBUS network frequency conversion control, the difficult points and configuration software design were elaborated. The running results showed that there were several advantages in the control system. They were high automatic degree, perfect function, perfect steady and convenient operation.Keywords: PLC, PROFIBUS, configuration, frequency
Procedia PDF Downloads 4031459 Effect of Clay Content on the Drained Shear Strength
Authors: Navid Khayat
Abstract:
Drained shear strength of saturated soils is fully understood. Shear strength of unsaturated soils is usually expressed in terms of soil suction. Evaluation of shear strength of compacted mixtures of sand–clay at optimum water content is main purpose of this research. To prepare the required samples, first clay and sand are mixed in 10, 30, 50, and 70 percent by dry weight and then compacted at the proper optimum water content according to the standard proctor test. The samples were sheared in direct shear machine. Stress –strain relationship of samples indicated a ductile behavior. Most of the samples showed a dilatancy behavior during the shear and the tendency for dilatancy increased with the increase in sand proportion. The results show that with the increase in percentage of sand a decrease in cohesion intercept c' for mixtures and an increase in the angle of internal friction Φ’is observed.Keywords: clay, sand, drained shear strength, cohesion intercept
Procedia PDF Downloads 4401458 Traffic Sign Recognition System Using Convolutional Neural NetworkDevineni
Authors: Devineni Vijay Bhaskar, Yendluri Raja
Abstract:
We recommend a model for traffic sign detection stranded on Convolutional Neural Networks (CNN). We first renovate the unique image into the gray scale image through with support vector machines, then use convolutional neural networks with fixed and learnable layers for revealing and understanding. The permanent layer can reduction the amount of attention areas to notice and crop the limits very close to the boundaries of traffic signs. The learnable coverings can rise the accuracy of detection significantly. Besides, we use bootstrap procedures to progress the accuracy and avoid overfitting problem. In the German Traffic Sign Detection Benchmark, we obtained modest results, with an area under the precision-recall curve (AUC) of 99.49% in the group “Risk”, and an AUC of 96.62% in the group “Obligatory”.Keywords: convolutional neural network, support vector machine, detection, traffic signs, bootstrap procedures, precision-recall curve
Procedia PDF Downloads 1231457 Testing of the Decreasing Bond Strength of Polyvinyl Acetate Adhesive by Low Temperatures
Authors: Pavel Boška, Jan Bomba, Tomáš Beránek, Jiří Procházka
Abstract:
When using wood products bonded by polyvinyl acetate, glues such as windows are the most limiting element of degradation of the glued joint due to weather changes. In addition to moisture and high temperatures, the joint may damage the low temperature below freezing point, where dimensional changes in the material and distortion of the adhesive film occur. During the experiments, the joints were exposed to several degrees of sub-zero temperatures from 0 °C to -40 °C and then to compare how the decreasing temperature affects the strength of the joint. The experiment was performed on wood beech samples (Fagus sylvatica), bonded with PVAc with D3 resistance and the shear strength of bond was measured. The glued and treated samples were tested on a laboratory testing machine, recording the strength of the joint. The statistical results have given us information that the strength of the joint gradually decreases with decreasing temperature, but a noticeable and statistically significant change is achieved only at very low temperatures.Keywords: adhesives, bond strength, low temperatures, polyvinyl acetate
Procedia PDF Downloads 3501456 Optimization of Roster Construction In Sports
Authors: Elijah Cavan
Abstract:
In Major League Sports (MLB, NBA, NHL, NFL), it is the Front Office Staff (FOS) who make decisions about who plays for their respective team. The FOS bear the brunt of the responsibility for acquiring players through drafting, trading and signing players in free agency while typically contesting with maximum roster salary constraints. The players themselves are volatile assets of these teams- their value fluctuates with age and performance. A simple comparison can be made when viewing players as assets. The problem here is similar to that of optimizing your investment portfolio. The The goal is ultimately to maximize your periodic returns while tolerating a fixed risk (degree of uncertainty/ potential loss). Each franchise may value assets differently, and some may only tolerate lower risk levels- these are examples of factors that introduce additional constraints into the model. In this talk, we will detail the mathematical formulation of this problem as a constrained optimization problem- which can be solved with classical machine learning methods but is also well posed as a problem to be solved on quantum computersKeywords: optimization, financial mathematics, sports analytics, simulated annealing
Procedia PDF Downloads 1221455 Critical Review of Web Content Mining Extraction Mechanisms
Authors: Rabia Bashir, Sajjad Akbar
Abstract:
There is an inevitable demand of web mining due to rapid increase of huge information on the Internet, but the striking variety of web structures has made required content retrieval a difficult task. To counter this issue, Web Content Mining (WCM) emerges as a potential candidate which extracts and integrates suitable resources of data to users. In past few years, research has been done on several extraction techniques for WCM i.e. agent-based, template-based, assumption-based, statistic-based, wrapper-based and machine learning. However, it is still unclear that either these approaches are efficiently tackling the significant challenges of WCM or not. To answer this question, this paper identifies these challenges such as language independency, structure flexibility, performance, automation, dynamicity, redundancy handling, intelligence, relevant content retrieval, and privacy. Further, mapping of these challenges is done with existing extraction mechanisms which helps to adopt the most suitable WCM approach, given some conditions and characteristics at hand.Keywords: content mining challenges, web content mining, web content extraction approaches, web information retrieval
Procedia PDF Downloads 5491454 A Parallel Cellular Automaton Model of Tumor Growth for Multicore and GPU Programming
Authors: Manuel I. Capel, Antonio Tomeu, Alberto Salguero
Abstract:
Tumor growth from a transformed cancer-cell up to a clinically apparent mass spans through a range of spatial and temporal magnitudes. Through computer simulations, Cellular Automata (CA) can accurately describe the complexity of the development of tumors. Tumor development prognosis can now be made -without making patients undergo through annoying medical examinations or painful invasive procedures- if we develop appropriate CA-based software tools. In silico testing mainly refers to Computational Biology research studies of application to clinical actions in Medicine. To establish sound computer-based models of cellular behavior, certainly reduces costs and saves precious time with respect to carrying out experiments in vitro at labs or in vivo with living cells and organisms. These aim to produce scientifically relevant results compared to traditional in vitro testing, which is slow, expensive, and does not generally have acceptable reproducibility under the same conditions. For speeding up computer simulations of cellular models, specific literature shows recent proposals based on the CA approach that include advanced techniques, such the clever use of supporting efficient data structures when modeling with deterministic stochastic cellular automata. Multiparadigm and multiscale simulation of tumor dynamics is just beginning to be developed by the concerned research community. The use of stochastic cellular automata (SCA), whose parallel programming implementations are open to yield a high computational performance, are of much interest to be explored up to their computational limits. There have been some approaches based on optimizations to advance in multiparadigm models of tumor growth, which mainly pursuit to improve performance of these models through efficient memory accesses guarantee, or considering the dynamic evolution of the memory space (grids, trees,…) that holds crucial data in simulations. In our opinion, the different optimizations mentioned above are not decisive enough to achieve the high performance computing power that cell-behavior simulation programs actually need. The possibility of using multicore and GPU parallelism as a promising multiplatform and framework to develop new programming techniques to speed-up the computation time of simulations is just starting to be explored in the few last years. This paper presents a model that incorporates parallel processing, identifying the synchronization necessary for speeding up tumor growth simulations implemented in Java and C++ programming environments. The speed up improvement that specific parallel syntactic constructs, such as executors (thread pools) in Java, are studied. The new tumor growth parallel model is proved using implementations with Java and C++ languages on two different platforms: chipset Intel core i-X and a HPC cluster of processors at our university. The parallelization of Polesczuk and Enderling model (normally used by researchers in mathematical oncology) proposed here is analyzed with respect to performance gain. We intend to apply the model and overall parallelization technique presented here to solid tumors of specific affiliation such as prostate, breast, or colon. Our final objective is to set up a multiparadigm model capable of modelling angiogenesis, or the growth inhibition induced by chemotaxis, as well as the effect of therapies based on the presence of cytotoxic/cytostatic drugs.Keywords: cellular automaton, tumor growth model, simulation, multicore and manycore programming, parallel programming, high performance computing, speed up
Procedia PDF Downloads 2441453 The Beam Expansion Method, A Simplified and Efficient Approach of Field Propagation and Resonators Modes Study
Authors: Zaia Derrar Kaddour
Abstract:
The study of a beam throughout an optical path is generally achieved by means of diffraction integral. Unfortunately, in some problems, this tool turns out to be not very friendly and hard to implement. Instead, the beam expansion method for computing field profiles appears to be an interesting alternative. The beam expansion method consists of expanding the field pattern as a series expansion in a set of orthogonal functions. Propagating each individual component through a circuit and adding up the derived elements leads easily to the result. The problem is then reduced to finding how the expansion coefficients change in a circuit. The beam expansion method requires a systematic study of each type of optical element that can be met in the considered optical path. In this work, we analyze the following fundamental elements: first order optical systems, hard apertures and waveguides. We show that the former element type is completely defined thanks to the Gouy phase shift expression we provide and the latters require a suitable mode conversion. For endorsing the usefulness and relevance of the beam expansion approach, we show here some of its applications such as the treatment of the thermal lens effect and the study of unstable resonators.Keywords: gouy phase shift, modes, optical resonators, unstable resonators
Procedia PDF Downloads 631452 Graded Orientation of the Linear Polymers
Authors: Levan Nadareishvili, Roland Bakuradze, Barbara Kilosanidze, Nona Topuridze, Liana Sharashidze, Ineza Pavlenishvili
Abstract:
Some regularities of formation of a new structural state of the thermoplastic polymers-gradually oriented (stretched) state (GOS) are discussed. Transition into GOS is realized by the graded oriented stretching-by action of inhomogeneous mechanical field on the isotropic linear polymers or by zonal stretching that is implemented on a standard tensile-testing machine with using a specially designed zone stretching device (ZSD). Both technical approaches (especially zonal stretching method) allows to manage the such quantitative parameters of gradually oriented polymers as a range of change in relative elongation/orientation degree, length of this change and profile (linear, hyperbolic, parabolic, logarithmic, etc.). Uniaxial graded stretching method should be considered as an effective technological solution to create polymer materials with a predetermined gradient of physical properties.Keywords: controlled graded stretching, gradually oriented state, linear polymers, zone stretching device
Procedia PDF Downloads 4381451 Modal Analysis for Optimal Location of Doubly Fed Induction-Generator-Based Wind Farms for Reduction of Small Signal Oscillation
Authors: Meet Patel, Darshan Patel, Nilay Shah
Abstract:
Excess growth of wind-based renewable energy sources is required to identify the optimal location and damping capacity of doubly fed induction-generator-based (DFIG) wind farms while it penetrates into the transmission network. In this analysis, various ratings of DFIG wind farms are penetrated into the Single Machine Infinite Bus (SMIB ) at a different distance of the transmission line. On the basis of detailed examinations, a prime position is evaluated to maximize the stability of overall systems. A damping controller is designed at an optimum location to mitigate the small oscillations. The proposed model was validated using eigenvalue analysis, calculation of the participation factor, and time-domain simulation.Keywords: DFIG, small signal stability, eigenvalues, time domain simulation
Procedia PDF Downloads 1131450 A Preliminary Study for Building an Arabic Corpus of Pair Questions-Texts from the Web: Aqa-Webcorp
Authors: Wided Bakari, Patrce Bellot, Mahmoud Neji
Abstract:
With the development of electronic media and the heterogeneity of Arabic data on the Web, the idea of building a clean corpus for certain applications of natural language processing, including machine translation, information retrieval, question answer, become more and more pressing. In this manuscript, we seek to create and develop our own corpus of pair’s questions-texts. This constitution then will provide a better base for our experimentation step. Thus, we try to model this constitution by a method for Arabic insofar as it recovers texts from the web that could prove to be answers to our factual questions. To do this, we had to develop a java script that can extract from a given query a list of html pages. Then clean these pages to the extent of having a database of texts and a corpus of pair’s question-texts. In addition, we give preliminary results of our proposal method. Some investigations for the construction of Arabic corpus are also presented in this document.Keywords: Arabic, web, corpus, search engine, URL, question, corpus building, script, Google, html, txt
Procedia PDF Downloads 3241449 Orientational Pair Correlation Functions Modelling of the LiCl6H2O by the Hybrid Reverse Monte Carlo: Using an Environment Dependence Interaction Potential
Authors: Mohammed Habchi, Sidi Mohammed Mesli, Rafik Benallal, Mohammed Kotbi
Abstract:
On the basis of four partial correlation functions and some geometric constraints obtained from neutron scattering experiments, a Reverse Monte Carlo (RMC) simulation has been performed in the study of the aqueous electrolyte LiCl6H2O at the glassy state. The obtained 3-dimensional model allows computing pair radial and orientational distribution functions in order to explore the structural features of the system. Unrealistic features appeared in some coordination peaks. To remedy to this, we use the Hybrid Reverse Monte Carlo (HRMC), incorporating an additional energy constraint in addition to the usual constraints derived from experiments. The energy of the system is calculated using an Environment Dependence Interaction Potential (EDIP). Ions effects is studied by comparing correlations between water molecules in the solution and in pure water at room temperature Our results show a good agreement between experimental and computed partial distribution functions (PDFs) as well as a significant improvement in orientational distribution curves.Keywords: LiCl6H2O, glassy state, RMC, HRMC
Procedia PDF Downloads 4721448 Delamination of Scale in a Fe Carbon Steel Surface by Effect of Interface Roughness and Oxide Scale Thickness
Authors: J. M. Lee, W. R. Noh, C. Y. Kim, M. G. Lee
Abstract:
Delamination of oxide scale has been often discovered at the interface between Fe carbon steel and oxide scale. Among several mechanisms of this delamination behavior, the normal tensile stress to the substrate-scale interface has been described as one of the main factors. The stress distribution at the interface is also known to be affected by thermal expansion mismatch between substrate and oxide scale, creep behavior during cooling and the geometry of the interface. In this study, stress states near the interface in a Fe carbon steel with oxide scale have been investigated using FE simulations. The thermal and mechanical properties of oxide scales are indicated in literature and Fe carbon steel is measured using tensile testing machine. In particular, the normal and shear stress components developed at the interface during bending are investigated. Preliminary numerical sensitivity analyses are provided to explain the effects of the interface geometry and oxide thickness on the delamination behavior.Keywords: oxide scale, delamination, Fe analysis, roughness, thickness, stress state
Procedia PDF Downloads 3451447 Forthcoming Big Data on Smart Buildings and Cities: An Experimental Study on Correlations among Urban Data
Authors: Yu-Mi Song, Sung-Ah Kim, Dongyoun Shin
Abstract:
Cities are complex systems of diverse and inter-tangled activities. These activities and their complex interrelationships create diverse urban phenomena. And such urban phenomena have considerable influences on the lives of citizens. This research aimed to develop a method to reveal the causes and effects among diverse urban elements in order to enable better understanding of urban activities and, therefrom, to make better urban planning strategies. Specifically, this study was conducted to solve a data-recommendation problem found on a Korean public data homepage. First, a correlation analysis was conducted to find the correlations among random urban data. Then, based on the results of that correlation analysis, the weighted data network of each urban data was provided to people. It is expected that the weights of urban data thereby obtained will provide us with insights into cities and show us how diverse urban activities influence each other and induce feedback.Keywords: big data, machine learning, ontology model, urban data model
Procedia PDF Downloads 419