Search results for: equivalent transformation algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4349

Search results for: equivalent transformation algorithms

3899 Implicit Force Control of a Position Controlled Robot - A Comparison with Explicit Algorithms

Authors: Alexander Winkler, Jozef Suchý

Abstract:

This paper investigates simple implicit force control algorithms realizable with industrial robots. A lot of approaches already published are difficult to implement in commercial robot controllers, because the access to the robot joint torques is necessary or the complete dynamic model of the manipulator is used. In the past we already deal with explicit force control of a position controlled robot. Well known schemes of implicit force control are stiffness control, damping control and impedance control. Using such algorithms the contact force cannot be set directly. It is further the result of controller impedance, environment impedance and the commanded robot motion/position. The relationships of these properties are worked out in this paper in detail for the chosen implicit approaches. They have been adapted to be implementable on a position controlled robot. The behaviors of stiffness control and damping control are verified by practical experiments. For this purpose a suitable test bed was configured. Using the full mechanical impedance within the controller structure will not be practical in the case when the robot is in physical contact with the environment. This fact will be verified by simulation.

Keywords: robot force control, stiffness control, damping control, impedance control, stability

Procedia PDF Downloads 497
3898 Performance of Non-Deterministic Structural Optimization Algorithms Applied to a Steel Truss Structure

Authors: Ersilio Tushaj

Abstract:

The efficient solution that satisfies the optimal condition is an important issue in the structural engineering design problem. The new codes of structural design consist in design methodology that looks after the exploitation of the total resources of the construction material. In recent years some non-deterministic or meta-heuristic structural optimization algorithms have been developed widely in the research community. These methods search the optimum condition starting from the simulation of a natural phenomenon, such as survival of the fittest, the immune system, swarm intelligence or the cooling process of molten metal through annealing. Among these techniques the most known are: the genetic algorithms, simulated annealing, evolution strategies, particle swarm optimization, tabu search, ant colony optimization, harmony search and big bang crunch optimization. In this study, five of these algorithms are applied for the optimum weight design of a steel truss structure with variable geometry but fixed topology. The design process selects optimum distances and size sections from a set of commercial steel profiles. In the formulation of the design problem are considered deflection limitations, buckling and allowable stress constraints. The approach is repeated starting from different initial populations. The design problem topology is taken from an existing steel structure. The optimization process helps the engineer to achieve good final solutions, avoiding the repetitive evaluation of alternative designs in a time consuming process. The algorithms used for the application, the results of the optimal solutions, the number of iterations and the minimal weight designs, will be reported in the paper. Based on these results, it would be estimated, the amount of the steel that could be saved by applying structural analysis combined with non-deterministic optimization methods.

Keywords: structural optimization, non-deterministic methods, truss structures, steel truss

Procedia PDF Downloads 201
3897 Naturalization of Aliens in Consideration of Turkish Constitutional Law: Recent Governmental Practices

Authors: Zeynep Ozkan, Cigdem Serra Uzunpinar

Abstract:

Citizenship is a legal bond that binds a person to a certain state. How constitutions define ‘the citizen’ and how they regulate the elements of citizenship have great importance in terms of individuals’ duties before the state as well as the rights they own. Especially in multi-segmented societies that contain foreign elements, it becomes necessary to examinate the institution of naturalization in terms of individuals’ duty of constitutional citizenship. The meaning of citizenship in Turkey has transformed due to the changes in practices of naturalization, in parallel to receiving huge amount of immagrants with the recent Syrian Crisis, the change in the governmental system and facing economic crisis. This transformation took place in the way of a diversion from the states’ initial motive of building the bond of citizenship with the aim of founding/sustaining political unity. Hence, rising of the economic and political motives in naturalization practices are in question, instead of objective and subjective criterias, that are traditionally used on defining the notion of nation. In this study, firstly the regime of citizenship and the legal regime of aliens in Turkish legislation will be given place. Then, the transformation, that the notion of constitutional citizenship underwent, will be studied, especially on the basis of governmental practices of naturalization. The assessment will be made in the context of legal institutions brought with the new governmental system as a result of recent constitutional amendment.

Keywords: constitutional citizenship, naturalization, naturalization practices in Turkish legal system, transformation of the notion of constitutional citizenship

Procedia PDF Downloads 100
3896 An Equivalence between a Harmonic Form and a Closed Co-Closed Differential Form in L^Q and Non-L^Q Spaces

Authors: Lina Wu, Ye Li

Abstract:

An equivalent relation between a harmonic form and a closed co-closed form is established on a complete non-compact manifold. This equivalence has been generalized for a differential k-form ω from Lq spaces to non-Lq spaces when q=2 in the context of p-balanced growth where p=2. Especially for a simple differential k-form on a complete non-compact manifold, the equivalent relation has been verified with the extended scope of q for from finite q-energy in Lq spaces to infinite q-energy in non-Lq spaces when with 2-balanced growth. Generalized Hadamard Theorem, Cauchy-Schwarz Inequality, and Calculus skills including Integration by Parts as well as Convergent Series have been applied as estimation techniques to evaluate growth rates for a differential form. In particular, energy growth rates as indicated by an appropriate power range in a selected test function lead to a balance between a harmonic differential form and a closed co-closed differential form. Research ideas and computational methods in this paper could provide an innovative way in the study of broadening Lq spaces to non-Lq spaces with a wide variety of infinite energy growth for a differential form.

Keywords: closed forms, co-closed forms, harmonic forms, L^q spaces, p-balanced growth, simple differential k-forms

Procedia PDF Downloads 425
3895 Effect of Nano/Micro Alumina Matrix on Alumina-Cubic Boron Nitride Composites Consolidated by Spark Plasma Sintering

Authors: A. S. Hakeem, B. Ahmed, M. Ehsan, A. Ibrahim, H. M. Irshad, T. Laoui

Abstract:

Alumina (Al2O3) - cubic boron nitride (cBN) ceramic composites were sintered by spark plasma sintering (SPS) using α-Al2O3 particle sizes; 150 µm, 150 nm and cBN particle size of 42 µm. Alumina-cBN composites containing 10, 20 and 30wt% cBN with and without Ni coated were sintering at an elevated temperature of 1400°C at a constant uniaxial pressure of 50 MPa. The effect of matrix particle size, cBN and Ni content on mechanical properties and thermal properties, i.e., thermal conductivity, diffusivity, expansion, densification, phase transformation, microstructure, hardness and toughness of the Al2O3-cBN/(Ni) composites under specific sintering conditions were investigated. The highest relative densification of 150 nm-Al2O3 containing 30wt% cBN (Ni coated) composite was 99% at TSPS = 1400°C. In case of 150 µm- Al2O3 compositions, the phase transformation of cBN to hBN were observed, and the relative densification decreased. Thermal conductivity depicts maximum value in case of 150 nm- Al2O3-30wt% cBN-Ni composition. The Vickers hardness of this composition at TSPS = 1400°C also showed the highest value of 29 GPa.

Keywords: alumina composite, cubic boron nitride, mechanical properties, phase transformation, Spark plasma sintering

Procedia PDF Downloads 320
3894 Incorporating Multiple Supervised Learning Algorithms for Effective Intrusion Detection

Authors: Umar Albalawi, Sang C. Suh, Jinoh Kim

Abstract:

As internet continues to expand its usage with an enormous number of applications, cyber-threats have significantly increased accordingly. Thus, accurate detection of malicious traffic in a timely manner is a critical concern in today’s Internet for security. One approach for intrusion detection is to use Machine Learning (ML) techniques. Several methods based on ML algorithms have been introduced over the past years, but they are largely limited in terms of detection accuracy and/or time and space complexity to run. In this work, we present a novel method for intrusion detection that incorporates a set of supervised learning algorithms. The proposed technique provides high accuracy and outperforms existing techniques that simply utilizes a single learning method. In addition, our technique relies on partial flow information (rather than full information) for detection, and thus, it is light-weight and desirable for online operations with the property of early identification. With the mid-Atlantic CCDC intrusion dataset publicly available, we show that our proposed technique yields a high degree of detection rate over 99% with a very low false alarm rate (0.4%).

Keywords: intrusion detection, supervised learning, traffic classification, computer networks

Procedia PDF Downloads 325
3893 Short Text Classification Using Part of Speech Feature to Analyze Students' Feedback of Assessment Components

Authors: Zainab Mutlaq Ibrahim, Mohamed Bader-El-Den, Mihaela Cocea

Abstract:

Students' textual feedback can hold unique patterns and useful information about learning process, it can hold information about advantages and disadvantages of teaching methods, assessment components, facilities, and other aspects of teaching. The results of analysing such a feedback can form a key point for institutions’ decision makers to advance and update their systems accordingly. This paper proposes a data mining framework for analysing end of unit general textual feedback using part of speech feature (PoS) with four machine learning algorithms: support vector machines, decision tree, random forest, and naive bays. The proposed framework has two tasks: first, to use the above algorithms to build an optimal model that automatically classifies the whole data set into two subsets, one subset is tailored to assessment practices (assessment related), and the other one is the non-assessment related data. Second task to use the same algorithms to build an optimal model for whole data set, and the new data subsets to automatically detect their sentiment. The significance of this paper is to compare the performance of the above four algorithms using part of speech feature to the performance of the same algorithms using n-grams feature. The paper follows Knowledge Discovery and Data Mining (KDDM) framework to construct the classification and sentiment analysis models, which is understanding the assessment domain, cleaning and pre-processing the data set, selecting and running the data mining algorithm, interpreting mined patterns, and consolidating the discovered knowledge. The results of this paper experiments show that both models which used both features performed very well regarding first task. But regarding the second task, models that used part of speech feature has underperformed in comparison with models that used unigrams and bigrams.

Keywords: assessment, part of speech, sentiment analysis, student feedback

Procedia PDF Downloads 116
3892 Comparative Analysis of Reinforcement Learning Algorithms for Autonomous Driving

Authors: Migena Mana, Ahmed Khalid Syed, Abdul Malik, Nikhil Cherian

Abstract:

In recent years, advancements in deep learning enabled researchers to tackle the problem of self-driving cars. Car companies use huge datasets to train their deep learning models to make autonomous cars a reality. However, this approach has certain drawbacks in that the state space of possible actions for a car is so huge that there cannot be a dataset for every possible road scenario. To overcome this problem, the concept of reinforcement learning (RL) is being investigated in this research. Since the problem of autonomous driving can be modeled in a simulation, it lends itself naturally to the domain of reinforcement learning. The advantage of this approach is that we can model different and complex road scenarios in a simulation without having to deploy in the real world. The autonomous agent can learn to drive by finding the optimal policy. This learned model can then be easily deployed in a real-world setting. In this project, we focus on three RL algorithms: Q-learning, Deep Deterministic Policy Gradient (DDPG), and Proximal Policy Optimization (PPO). To model the environment, we have used TORCS (The Open Racing Car Simulator), which provides us with a strong foundation to test our model. The inputs to the algorithms are the sensor data provided by the simulator such as velocity, distance from side pavement, etc. The outcome of this research project is a comparative analysis of these algorithms. Based on the comparison, the PPO algorithm gives the best results. When using PPO algorithm, the reward is greater, and the acceleration, steering angle and braking are more stable compared to the other algorithms, which means that the agent learns to drive in a better and more efficient way in this case. Additionally, we have come up with a dataset taken from the training of the agent with DDPG and PPO algorithms. It contains all the steps of the agent during one full training in the form: (all input values, acceleration, steering angle, break, loss, reward). This study can serve as a base for further complex road scenarios. Furthermore, it can be enlarged in the field of computer vision, using the images to find the best policy.

Keywords: autonomous driving, DDPG (deep deterministic policy gradient), PPO (proximal policy optimization), reinforcement learning

Procedia PDF Downloads 121
3891 Non-Linear Behavior of Granular Materials in Pavement Design

Authors: Mounir Tichamakdj, Khaled Sandjak, Boualem Tiliouine

Abstract:

The design of flexible pavements is currently carried out using a multilayer elastic theory. However, for thin-surface pavements subject to light or medium traffic volumes, the importance of the non-linear stress-strain behavior of unbound granular materials requires the use of more sophisticated numerical models for the structural design of these pavements. The simplified analysis of the nonlinear behavior of granular materials in pavement design will be developed in this study. To achieve this objective, an equivalent linear model derived from a volumetric shear stress model is used to simulate the nonlinear elastic behavior of two unlinked local granular materials often used in pavements. This model is included here to adequately incorporate material non-linearity due to stress dependence and stiffness of the granular layers in the flexible pavement analysis. The sensitivity of the pavement design criteria to the likely variations in asphalt layer thickness and the mineralogical nature of unbound granular materials commonly used in pavement structures are also evaluated.

Keywords: granular materials, linear equivalent model, non-linear behavior, pavement design, shear volumetric strain model

Procedia PDF Downloads 154
3890 Markowitz and Implementation of a Multi-Objective Evolutionary Technique Applied to the Colombia Stock Exchange (2009-2015)

Authors: Feijoo E. Colomine Duran, Carlos E. Peñaloza Corredor

Abstract:

There modeling component selection financial investment (Portfolio) a variety of problems that can be addressed with optimization techniques under evolutionary schemes. For his feature, the problem of selection of investment components of a dichotomous relationship between two elements that are opposed: The Portfolio Performance and Risk presented by choosing it. This relationship was modeled by Markowitz through a media problem (Performance) - variance (risk), ie must Maximize Performance and Minimize Risk. This research included the study and implementation of multi-objective evolutionary techniques to solve these problems, taking as experimental framework financial market equities Colombia Stock Exchange between 2009-2015. Comparisons three multiobjective evolutionary algorithms, namely the Nondominated Sorting Genetic Algorithm II (NSGA-II), the Strength Pareto Evolutionary Algorithm 2 (SPEA2) and Indicator-Based Selection in Multiobjective Search (IBEA) were performed using two measures well known performance: The Hypervolume indicator and R_2 indicator, also it became a nonparametric statistical analysis and the Wilcoxon rank-sum test. The comparative analysis also includes an evaluation of the financial efficiency of the investment portfolio chosen by the implementation of various algorithms through the Sharpe ratio. It is shown that the portfolio provided by the implementation of the algorithms mentioned above is very well located between the different stock indices provided by the Colombia Stock Exchange.

Keywords: finance, optimization, portfolio, Markowitz, evolutionary algorithms

Procedia PDF Downloads 273
3889 A Review on the Perception of Beşiktaş Public Square

Authors: Neslinur Hizli, Berrak Kirbaş Akyürek

Abstract:

Beşiktaş, one of the historical coastal district of İstanbul, is on the very edge of the radical transformation because of an approaching ‘Beşiktaş Public Square Project’. At this juncture, due its location, presence on the coast, population density and distance to the other centers of the city, the decisions to be taken are critical to whole Istanbul that will be majorly affected from this transformation. As the new project aims to pedestrianize the area by placing the vehicular traffic under the ground, Beşiktaş and its square will change from top to bottom. Among those considerations, through the advantages and disadvantages the perception of the existing conditions of the Beşiktaş play significant role. The motive of this paper is the lack of determination and clarity on the cognition of the Square. After brief analysis on the historical transformation of the area, prominent studies on the criteria of public square are revised. Through cognitive mapping methodology, characteristics of the Square and the public space in general find a place to discuss from individual views. This study aims to discuss and review Beşiktaş Public Square from perspective, mind and behavior of the users. Cognitive map study with thirty subjects (30) is evaluated and categorized upon the five elements that Kevin Lynch defined as the images of the city. The results obtained digitized and represented with tables and graphs. Findings of the research underline the crucial issues on the approaching change in Beşiktaş. Thus, this study may help to develop comprehensive ideas and new suggestions on the Square.

Keywords: Beşiktaş public square, cognitive map, perception, public space

Procedia PDF Downloads 245
3888 The Evaluation of Transformational Leadership Characteristics and Behaviors in Air Forces

Authors: Cuma Şimşek

Abstract:

Nowadays our globalized world is in a very rapid and sophisticated change. In the information age, notion of ‘information’ has begun to spread faster than ever also in this age, changes and transformation has gained tremendous momentum with technology boom. This continuous change and transformation, increased the competition between existing organizations and corporations. Besides, the organizations which show resistance to change has been put out of action in this competitive environment. It is not possible to sustain the existence of organizations without adapting to change and transformation by isolating itself from developments. As a consequence of improved communication and dialog possibilities by means of increasing knowledge level, there has been made a change of scene in administrative mentality, style and activation, especially in 21th century. Leaders emerge as the most important factor in this process of perception and success. At the same time it is not enough to adapt the alteration with conventional leadership abilities and behaviors. In parallel with alteration, new types of leadership are coming up. The optimal leadership type for our era and a trending topic "Transformational Leadership" is in great demand now. In this research, current situation of the Air Forces which use high-technology weapons efficiently, operates in an environment full of threats and is analyzed. It is evaluated that in order to be ready for war continuously and adjusting itself to changing terms of warfare atmosphere , Air Forces need ‘transformational leaders’ who are innovative, foreseeing and having a vision so that they can develop new methods and strategies for complex problems. Because it is the Air Force which is responsible for being the deterrent force of its country.

Keywords: transformational, change, air force, leadership

Procedia PDF Downloads 422
3887 EnumTree: An Enumerative Biclustering Algorithm for DNA Microarray Data

Authors: Haifa Ben Saber, Mourad Elloumi

Abstract:

In a number of domains, like in DNA microarray data analysis, we need to cluster simultaneously rows (genes) and columns (conditions) of a data matrix to identify groups of constant rows with a group of columns. This kind of clustering is called biclustering. Biclustering algorithms are extensively used in DNA microarray data analysis. More effective biclustering algorithms are highly desirable and needed. We introduce a new algorithm called, Enumerative tree (EnumTree) for biclustering of binary microarray data. is an algorithm adopting the approach of enumerating biclusters. This algorithm extracts all biclusters consistent good quality. The main idea of ​​EnumLat is the construction of a new tree structure to represent adequately different biclusters discovered during the process of enumeration. This algorithm adopts the strategy of all biclusters at a time. The performance of the proposed algorithm is assessed using both synthetic and real DNA micryarray data, our algorithm outperforms other biclustering algorithms for binary microarray data. Biclusters with different numbers of rows. Moreover, we test the biological significance using a gene annotation web tool to show that our proposed method is able to produce biologically relevent biclusters.

Keywords: DNA microarray, biclustering, gene expression data, tree, datamining.

Procedia PDF Downloads 353
3886 A Hybrid Watermarking Scheme Using Discrete and Discrete Stationary Wavelet Transformation For Color Images

Authors: Bülent Kantar, Numan Ünaldı

Abstract:

This paper presents a new method which includes robust and invisible digital watermarking on images that is colored. Colored images are used as watermark. Frequency region is used for digital watermarking. Discrete wavelet transform and discrete stationary wavelet transform are used for frequency region transformation. Low, medium and high frequency coefficients are obtained by applying the two-level discrete wavelet transform to the original image. Low frequency coefficients are obtained by applying one level discrete stationary wavelet transform separately to all frequency coefficient of the two-level discrete wavelet transformation of the original image. For every low frequency coefficient obtained from one level discrete stationary wavelet transformation, watermarks are added. Watermarks are added to all frequency coefficients of two-level discrete wavelet transform. Totally, four watermarks are added to original image. In order to get back the watermark, the original and watermarked images are applied with two-level discrete wavelet transform and one level discrete stationary wavelet transform. The watermark is obtained from difference of the discrete stationary wavelet transform of the low frequency coefficients. A total of four watermarks are obtained from all frequency of two-level discrete wavelet transform. Obtained watermark results are compared with real watermark results, and a similarity result is obtained. A watermark is obtained from the highest similarity values. Proposed methods of watermarking are tested against attacks of the geometric and image processing. The results show that proposed watermarking method is robust and invisible. All features of frequencies of two level discrete wavelet transform watermarking are combined to get back the watermark from the watermarked image. Watermarks have been added to the image by converting the binary image. These operations provide us with better results in getting back the watermark from watermarked image by attacking of the geometric and image processing.

Keywords: watermarking, DWT, DSWT, copy right protection, RGB

Procedia PDF Downloads 511
3885 Image Encryption Using Eureqa to Generate an Automated Mathematical Key

Authors: Halima Adel Halim Shnishah, David Mulvaney

Abstract:

Applying traditional symmetric cryptography algorithms while computing encryption and decryption provides immunity to secret keys against different attacks. One of the popular techniques generating automated secret keys is evolutionary computing by using Eureqa API tool, which got attention in 2013. In this paper, we are generating automated secret keys for image encryption and decryption using Eureqa API (tool which is used in evolutionary computing technique). Eureqa API models pseudo-random input data obtained from a suitable source to generate secret keys. The validation of generated secret keys is investigated by performing various statistical tests (histogram, chi-square, correlation of two adjacent pixels, correlation between original and encrypted images, entropy and key sensitivity). Experimental results obtained from methods including histogram analysis, correlation coefficient, entropy and key sensitivity, show that the proposed image encryption algorithms are secure and reliable, with the potential to be adapted for secure image communication applications.

Keywords: image encryption algorithms, Eureqa, statistical measurements, automated key generation

Procedia PDF Downloads 461
3884 A Novel Guided Search Based Multi-Objective Evolutionary Algorithm

Authors: A. Baviskar, C. Sandeep, K. Shankar

Abstract:

Solving Multi-objective Optimization Problems requires faster convergence and better spread. Though existing Evolutionary Algorithms (EA's) are able to achieve this, the computation effort can further be reduced by hybridizing them with innovative strategies. This study is focuses on converging to the pareto front faster while adapting the advantages of Strength Pareto Evolutionary Algorithm-II (SPEA-II) for a better spread. Two different approaches based on optimizing the objective functions independently are implemented. In the first method, the decision variables corresponding to the optima of individual objective functions are strategically used to guide the search towards the pareto front. In the second method, boundary points of the pareto front are calculated and their decision variables are seeded to the initial population. Both the methods are applied to different constrained and unconstrained multi-objective test functions. It is observed that proposed guided search based algorithm gives better convergence and diversity than several well-known existing algorithms (such as NSGA-II and SPEA-II) in considerably less number of iterations.

Keywords: boundary points, evolutionary algorithms (EA's), guided search, strength pareto evolutionary algorithm-II (SPEA-II)

Procedia PDF Downloads 248
3883 Rhetoric and Renarrative Structure of Digital Images in Trans-Media

Authors: Yang Geng, Anqi Zhao

Abstract:

The misreading theory of Harold Bloom provides a new diachronic perspective as an approach to the consistency between rhetoric of digital technology, dynamic movement of digital images and uncertain meaning of text. Reinterpreting the diachroneity of 'intertextuality' in the context of misreading theory extended the range of the 'intermediality' of transmedia to the intense tension between digital images and symbolic images throughout history of images. With the analogy between six categories of revisionary ratios and six steps of digital transformation, digital rhetoric might be illustrated as a linear process reflecting dynamic, intensive relations between digital moving images and original static images. Finally, it was concluded that two-way framework of the rhetoric of transformation of digital images and reversed served as a renarrative structure to revive static images by reconnecting them with digital moving images.

Keywords: rhetoric, digital art, intermediality, misreading theory

Procedia PDF Downloads 226
3882 Institutional Transformation in a Finnish University of Applied Sciences

Authors: Perttu Heino

Abstract:

Universities of applied sciences (UASs) were introduced as part of the Finnish higher education system in the early 1990’s. Research, development and innovation (RDI) were mentioned in the legislation on UASs from the very beginning, but very little attention was paid to it in the early years of UASs due to reasons that are easy to understand. Required changes in the organization of the UAS and its educational offering were a big challenge, and it took several years to get things in order. There were RDI projects already then, but their number was low and there was no systematical coordination or management of those activities. In this paper, the institutional transformation of UASs is discussed based on how Tampere University of Applied Sciences has transformed during the years from a traditional tertiary level school to a modern higher education institution with a strong RDI activity, characterized by lively university-industry interaction and tight integration to education.

Keywords: research, development, management, practices

Procedia PDF Downloads 24
3881 Radioactivity Assessment of Sediments in Negombo Lagoon Sri Lanka

Authors: H. M. N. L. Handagiripathira

Abstract:

The distributions of naturally occurring and anthropogenic radioactive materials were determined in surface sediments taken at 27 different locations along the bank of Negombo Lagoon in Sri Lanka. Hydrographic parameters of lagoon water and the grain size analyses of the sediment samples were also carried out for this study. The conductivity of the adjacent water was varied from 13.6 mS/cm to 55.4 mS/cm near to the southern end and the northern end of the lagoon, respectively, and equally salinity levels varied from 7.2 psu to 32.1 psu. The average pH in the water was 7.6 and average water temperature was 28.7 °C. The grain size analysis emphasized the mass fractions of the samples as sand (60.9%), fine sand (30.6%) and fine silt+clay (1.3%) in the sampling locations. The surface sediment samples of wet weight, 1 kg each from upper 5-10 cm layer, were oven dried at 105 °C for 24 hours to get a constant weight, homogenized and sieved through a 2 mm sieve (IAEA technical series no. 295). The radioactivity concentrations were determined using gamma spectrometry technique. Ultra Low Background Broad Energy High Purity Ge Detector, BEGe (Model BE5030, Canberra) was used for radioactivity measurement with Canberra Industries' Laboratory Source-less Calibration Software (LabSOCS) mathematical efficiency calibration approach and Geometry composer software. The mean activity concentration was found to be 24 ± 4, 67 ± 9, 181 ± 10, 59 ± 8, 3.5 ± 0.4 and 0.47 ± 0.08 Bq/kg for 238U, 232Th, 40K, 210Pb, 235U and 137Cs respectively. The mean absorbed dose rate in air, radium equivalent activity, external hazard index, annual gonadal dose equivalent and annual effective dose equivalent were 60.8 nGy/h, 137.3 Bq/kg, 0.4, 425.3 mSv/year and 74.6 mSv/year, respectively. The results of this study will provide baseline information on the natural and artificial radioactive isotopes and environmental pollution associated with information on radiological risk.

Keywords: gamma spectrometry, lagoon, radioactivity, sediments

Procedia PDF Downloads 118
3880 Dissolved Black Carbon Accelerates the Photo-Degradation of Polystyrene Microplastics

Authors: Qin Ou, Yanghui Xu, Xintu Wang, Kim Maren Lompe, Gang Liu, Jan Peter Van Der Hoek

Abstract:

Microplastics (MPs) can undergo the photooxidation process under ultraviolet (UV) exposure, which determines their transformation and fate in environments. The presence of dissolved organic matter (DOM) can interact with MPs and take participate in the photo-degradation of MPs. As an important DOM component, dissolved black carbon (DBC), widely distributed in aquatic environments, can accelerate or inhibit the sunlight-driven photo-transformation of environmental pollutants. However, the role and underlying mechanism of DBC in the photooxidation of MPs are not clear. Herein, the DBC (< 0.45 µm) was extracted from wood biochar and fractionated by molecular weight (i.e., <3 KDa, 3 KDa−30 KDa, 30 KDa−0.45 µm). The effects of DBC chemical composition (i.e., molecular weight and chemical structure) in DBC-mediated photo-transformation of polystyrene (PS) MPs were investigated. The results showed that DBC initially inhibited the photo-degradation of MPs due to light shielding. Under UV exposure for 6−24 h, the presence of 5 mg/L DBC decreased the carbonyl index of MPs compared to the control. This inhibitory effect of DBC was found to decrease with increasing irradiation time. Notably, DBC initially decreased but then increased the hydroxyl index with aging time, suggesting that the role of DBC may shift from inhibition to acceleration. In terms of the different DBC fractions, the results showed that the smallest fraction of DBC (<3 KDa) significantly accelerated the photooxidation of PS MPs since it acted as reactive oxygen species (ROS) generators, especially in promoting the production of ¹O₂ and ³DBC* and •OH. With the increase in molecular weight, the acceleration effect of DBC on the degradation of MPs was decreased due to the increase of light shielding and possible decrease of photosensitization ability. This study thoroughly investigated the critical role of DBC chemical composition in the photooxidation process, which helps to assess the duration of aging and transformation of MPs during long-term weathering in natural waters.

Keywords: microplastics, photo-degradation, dissolved black carbon, molecular weight, photosensitization

Procedia PDF Downloads 54
3879 Feature Weighting Comparison Based on Clustering Centers in the Detection of Diabetic Retinopathy

Authors: Kemal Polat

Abstract:

In this paper, three feature weighting methods have been used to improve the classification performance of diabetic retinopathy (DR). To classify the diabetic retinopathy, features extracted from the output of several retinal image processing algorithms, such as image-level, lesion-specific and anatomical components, have been used and fed them into the classifier algorithms. The dataset used in this study has been taken from University of California, Irvine (UCI) machine learning repository. Feature weighting methods including the fuzzy c-means clustering based feature weighting, subtractive clustering based feature weighting, and Gaussian mixture clustering based feature weighting, have been used and compered with each other in the classification of DR. After feature weighting, five different classifier algorithms comprising multi-layer perceptron (MLP), k- nearest neighbor (k-NN), decision tree, support vector machine (SVM), and Naïve Bayes have been used. The hybrid method based on combination of subtractive clustering based feature weighting and decision tree classifier has been obtained the classification accuracy of 100% in the screening of DR. These results have demonstrated that the proposed hybrid scheme is very promising in the medical data set classification.

Keywords: machine learning, data weighting, classification, data mining

Procedia PDF Downloads 307
3878 A Method for Compression of Short Unicode Strings

Authors: Masoud Abedi, Abbas Malekpour, Peter Luksch, Mohammad Reza Mojtabaei

Abstract:

The use of short texts in communication has been greatly increasing in recent years. Applying different languages in short texts has led to compulsory use of Unicode strings. These strings need twice the space of common strings, hence, applying algorithms of compression for the purpose of accelerating transmission and reducing cost is worthwhile. Nevertheless, other compression methods like gzip, bzip2 or PAQ due to high overhead data size are not appropriate. The Huffman algorithm is one of the rare algorithms effective in reducing the size of short Unicode strings. In this paper, an algorithm is proposed for compression of very short Unicode strings. At first, every new character to be sent to a destination is inserted in the proposed mapping table. At the beginning, every character is new. In case the character is repeated for the same destination, it is not considered as a new character. Next, the new characters together with the mapping value of repeated characters are arranged through a specific technique and specially formatted to be transmitted. The results obtained from an assessment made on a set of short Persian and Arabic strings indicate that this proposed algorithm outperforms the Huffman algorithm in size reduction.

Keywords: Algorithms, Data Compression, Decoding, Encoding, Huffman Codes, Text Communication

Procedia PDF Downloads 329
3877 The Clustering of Multiple Sclerosis Subgroups through L2 Norm Multifractal Denoising Technique

Authors: Yeliz Karaca, Rana Karabudak

Abstract:

Multifractal Denoising techniques are used in the identification of significant attributes by removing the noise of the dataset. Magnetic resonance (MR) image technique is the most sensitive method so as to identify chronic disorders of the nervous system such as Multiple Sclerosis. MRI and Expanded Disability Status Scale (EDSS) data belonging to 120 individuals who have one of the subgroups of MS (Relapsing Remitting MS (RRMS), Secondary Progressive MS (SPMS), Primary Progressive MS (PPMS)) as well as 19 healthy individuals in the control group have been used in this study. The study is comprised of the following stages: (i) L2 Norm Multifractal Denoising technique, one of the multifractal technique, has been used with the application on the MS data (MRI and EDSS). In this way, the new dataset has been obtained. (ii) The new MS dataset obtained from the MS dataset and L2 Multifractal Denoising technique has been applied to the K-Means and Fuzzy C Means clustering algorithms which are among the unsupervised methods. Thus, the clustering performances have been compared. (iii) In the identification of significant attributes in the MS dataset through the Multifractal denoising (L2 Norm) technique using K-Means and FCM algorithms on the MS subgroups and control group of healthy individuals, excellent performance outcome has been yielded. According to the clustering results based on the MS subgroups obtained in the study, successful clustering results have been obtained in the K-Means and FCM algorithms by applying the L2 norm of multifractal denoising technique for the MS dataset. Clustering performance has been more successful with the MS Dataset (L2_Norm MS Data Set) K-Means and FCM in which significant attributes are obtained by applying L2 Norm Denoising technique.

Keywords: clinical decision support, clustering algorithms, multiple sclerosis, multifractal techniques

Procedia PDF Downloads 144
3876 On the Application of Heuristics of the Traveling Salesman Problem for the Task of Restoring the DNA Matrix

Authors: Boris Melnikov, Dmitrii Chaikovskii, Elena Melnikova

Abstract:

The traveling salesman problem (TSP) is a well-known optimization problem that seeks to find the shortest possible route that visits a set of points and returns to the starting point. In this paper, we apply some heuristics of the TSP for the task of restoring the DNA matrix. This restoration problem is often considered in biocybernetics. For it, we must recover the matrix of distances between DNA sequences if not all the elements of the matrix under consideration are known at the input. We consider the possibility of using this method in the testing of distance calculation algorithms between a pair of DNAs to restore the partially filled matrix.

Keywords: optimization problems, DNA matrix, partially filled matrix, traveling salesman problem, heuristic algorithms

Procedia PDF Downloads 128
3875 Modelling of Phase Transformation Kinetics in Post Heat-Treated Resistance Spot Weld of AISI 1010 Mild Steel

Authors: B. V. Feujofack Kemda, N. Barka, M. Jahazi, D. Osmani

Abstract:

Automobile manufacturers are constantly seeking means to reduce the weight of car bodies. The usage of several steel grades in auto body assembling has been found to be a good technique to enlighten vehicles weight. This few years, the usage of dual phase (DP) steels, transformation induced plasticity (TRIP) steels and boron steels in some parts of the auto body have become a necessity because of their lightweight. However, these steels are martensitic, when they undergo a fast heat treatment, the resultant microstructure is essential, made of martensite. Resistance spot welding (RSW), one of the most used techniques in assembling auto bodies, becomes problematic in the case of these steels. RSW being indeed a process were steel is heated and cooled in a very short period of time, the resulting weld nugget is mostly fully martensitic, especially in the case of DP, TRIP and boron steels but that also holds for plain carbon steels as AISI 1010 grade which is extensively used in auto body inner parts. Martensite in its turn must be avoided as most as possible when welding steel because it is the principal source of brittleness and it weakens weld nugget. Thus, this work aims to find a mean to reduce martensite fraction in weld nugget when using RSW for assembling. The prediction of phase transformation kinetics during RSW has been done. That phase transformation kinetics prediction has been made possible through the modelling of the whole welding process, and a technique called post weld heat treatment (PWHT) have been applied in order to reduce martensite fraction in the weld nugget. Simulation has been performed for AISI 1010 grade, and results show that the application of PWHT leads to the formation of not only martensite but also ferrite, bainite and pearlite during the cooling of weld nugget. Welding experiments have been done in parallel and micrographic analyses show the presence of several phases in the weld nugget. Experimental weld geometry and phase proportions are in good agreement with simulation results, showing here the validity of the model.

Keywords: resistance spot welding, AISI 1010, modeling, post weld heat treatment, phase transformation, kinetics

Procedia PDF Downloads 95
3874 Data Mining in Medicine Domain Using Decision Trees and Vector Support Machine

Authors: Djamila Benhaddouche, Abdelkader Benyettou

Abstract:

In this paper, we used data mining to extract biomedical knowledge. In general, complex biomedical data collected in studies of populations are treated by statistical methods, although they are robust, they are not sufficient in themselves to harness the potential wealth of data. For that you used in step two learning algorithms: the Decision Trees and Support Vector Machine (SVM). These supervised classification methods are used to make the diagnosis of thyroid disease. In this context, we propose to promote the study and use of symbolic data mining techniques.

Keywords: biomedical data, learning, classifier, algorithms decision tree, knowledge extraction

Procedia PDF Downloads 526
3873 Predication Model for Leukemia Diseases Based on Data Mining Classification Algorithms with Best Accuracy

Authors: Fahd Sabry Esmail, M. Badr Senousy, Mohamed Ragaie

Abstract:

In recent years, there has been an explosion in the rate of using technology that help discovering the diseases. For example, DNA microarrays allow us for the first time to obtain a "global" view of the cell. It has great potential to provide accurate medical diagnosis, to help in finding the right treatment and cure for many diseases. Various classification algorithms can be applied on such micro-array datasets to devise methods that can predict the occurrence of Leukemia disease. In this study, we compared the classification accuracy and response time among eleven decision tree methods and six rule classifier methods using five performance criteria. The experiment results show that the performance of Random Tree is producing better result. Also it takes lowest time to build model in tree classifier. The classification rules algorithms such as nearest- neighbor-like algorithm (NNge) is the best algorithm due to the high accuracy and it takes lowest time to build model in classification.

Keywords: data mining, classification techniques, decision tree, classification rule, leukemia diseases, microarray data

Procedia PDF Downloads 296
3872 Numerical Investigations on Dynamic Stall of a Pitching-Plunging Helicopter Blade Airfoil

Authors: Xie Kai, Laith K. Abbas, Chen Dongyang, Yang Fufeng, Rui Xiaoting

Abstract:

Effect of plunging motion on the pitch oscillating NACA0012 airfoil is investigated using computational fluid dynamics (CFD). A simulation model based on overset grid technology and k - ω shear stress transport (SST) turbulence model is established, and the numerical simulation results are compared with available experimental data and other simulations. Two cases of phase angle φ = 0, μ which represents the phase difference between the pitching and plunging motions of an airfoil are performed. Airfoil vortex generation, moving, and shedding are discussed in detail. Good agreements have been achieved with the available literature. The upward plunging motion made the equivalent angle of attack less than the actual one during pitching analysis. It is observed that the formation of the stall vortex is suppressed, resulting in a decrease in the lift coefficient and a delay of the stall angle. However, the downward plunging motion made the equivalent angle of attack higher the actual one.

Keywords: dynamic stall, pitching-plunging, computational fluid dynamics, helicopter blade rotor, airfoil

Procedia PDF Downloads 204
3871 MapReduce Logistic Regression Algorithms with RHadoop

Authors: Byung Ho Jung, Dong Hoon Lim

Abstract:

Logistic regression is a statistical method for analyzing a dataset in which there are one or more independent variables that determine an outcome. Logistic regression is used extensively in numerous disciplines, including the medical and social science fields. In this paper, we address the problem of estimating parameters in the logistic regression based on MapReduce framework with RHadoop that integrates R and Hadoop environment applicable to large scale data. There exist three learning algorithms for logistic regression, namely Gradient descent method, Cost minimization method and Newton-Rhapson's method. The Newton-Rhapson's method does not require a learning rate, while gradient descent and cost minimization methods need to manually pick a learning rate. The experimental results demonstrated that our learning algorithms using RHadoop can scale well and efficiently process large data sets on commodity hardware. We also compared the performance of our Newton-Rhapson's method with gradient descent and cost minimization methods. The results showed that our newton's method appeared to be the most robust to all data tested.

Keywords: big data, logistic regression, MapReduce, RHadoop

Procedia PDF Downloads 253
3870 Transformation of the Business Model in an Occupational Health Care Company Embedded in an Emerging Personal Data Ecosystem: A Case Study in Finland

Authors: Tero Huhtala, Minna Pikkarainen, Saila Saraniemi

Abstract:

Information technology has long been used as an enabler of exchange for goods and services. Services are evolving from generic to personalized, and the reverse use of customer data has been discussed in both academia and industry for the past few years. This article presents the results of an empirical case study in the area of preventive health care services. The primary data were gathered in workshops, in which future personal data-based services were conceptualized by analyzing future scenarios from a business perspective. The aim of this study is to understand business model transformation in emerging personal data ecosystems. The work was done as a case study in the context of occupational healthcare. The results have implications to theory and practice, indicating that adopting personal data management principles requires transformation of the business model, which, if successfully managed, may provide access to more resources, potential to offer better value, and additional customer channels. These advantages correlate with the broadening of the business ecosystem. Expanding the scope of this study to include more actors would improve the validity of the research. The results draw from existing literature and are based on findings from a case study and the economic properties of the healthcare industry in Finland.

Keywords: ecosystem, business model, personal data, preventive healthcare

Procedia PDF Downloads 228