Search results for: sample entropy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6125

Search results for: sample entropy

6005 Phase Stability and Grain Growth Kinetics of Oxide Dispersed CoCrFeMnNi

Authors: Prangya P. Sahoo, B. S. Murty

Abstract:

The present study deals with phase evolution of oxide dispersed CoCrFeMnNi high entropy alloy as a function of amount of added Y2O3 during mechanical alloying and analysis of grain growth kinetics of CoCrFeMnNi high entropy alloy without and with oxide dispersion. Mechanical alloying of CoCrFeMnNi resulted in a single FCC phase. However, evolution of chromium carbide was observed after heat treatment between 1073 and 1473 K. Comparison of grain growth time exponents and activation energy barrier is also reported. Micro structural investigations, using electron microscopy and EBSD techniques, were carried out to confirm the enhanced grain growth resistance which is attributed to the presence oxide dispersoids.

Keywords: grain growth kinetics, mechanical alloying, oxide dispersion, phase evolution

Procedia PDF Downloads 400
6004 Highly Accurate Target Motion Compensation Using Entropy Function Minimization

Authors: Amin Aghatabar Roodbary, Mohammad Hassan Bastani

Abstract:

One of the defects of stepped frequency radar systems is their sensitivity to target motion. In such systems, target motion causes range cell shift, false peaks, Signal to Noise Ratio (SNR) reduction and range profile spreading because of power spectrum interference of each range cell in adjacent range cells which induces distortion in High Resolution Range Profile (HRRP) and disrupt target recognition process. Thus Target Motion Parameters (TMPs) effects compensation should be employed. In this paper, such a method for estimating TMPs (velocity and acceleration) and consequently eliminating or suppressing the unwanted effects on HRRP based on entropy minimization has been proposed. This method is carried out in two major steps: in the first step, a discrete search method has been utilized over the whole acceleration-velocity lattice network, in a specific interval seeking to find a less-accurate minimum point of the entropy function. Then in the second step, a 1-D search over velocity is done in locus of the minimum for several constant acceleration lines, in order to enhance the accuracy of the minimum point found in the first step. The provided simulation results demonstrate the effectiveness of the proposed method.

Keywords: automatic target recognition (ATR), high resolution range profile (HRRP), motion compensation, stepped frequency waveform technique (SFW), target motion parameters (TMPs)

Procedia PDF Downloads 126
6003 Entropy-Based Multichannel Stationary Measure for Characterization of Non-Stationary Patterns

Authors: J. D. Martínez-Vargas, C. Castro-Hoyos, G. Castellanos-Dominguez

Abstract:

In this work, we propose a novel approach for measuring the stationarity level of a multichannel time-series. This measure is based on a stationarity definition over time-varying spectrum, and it is aimed to quantify the relation between local stationarity (single-channel) and global dynamic behavior (multichannel dynamics). To assess the proposed approach validity, we use a well known EEG-BCI database, that was constructed for separate between motor/imagery tasks. Thus, based on the statement that imagination of movements implies an increase on the EEG dynamics, we use as discriminant features the proposed measure computed over an estimation of the non-stationary components of input time-series. As measure of separability we use a t-student test, and the obtained results evidence that such measure is able to accurately detect the brain areas projected on the scalp where motor tasks are realized.

Keywords: stationary measure, entropy, sub-space projection, multichannel dynamics

Procedia PDF Downloads 368
6002 Explaining Irregularity in Music by Entropy and Information Content

Authors: Lorena Mihelac, Janez Povh

Abstract:

In 2017, we conducted a research study using data consisting of 160 musical excerpts from different musical styles, to analyze the impact of entropy of the harmony on the acceptability of music. In measuring the entropy of harmony, we were interested in unigrams (individual chords in the harmonic progression) and bigrams (the connection of two adjacent chords). In this study, it has been found that 53 musical excerpts out from 160 were evaluated by participants as very complex, although the entropy of the harmonic progression (unigrams and bigrams) was calculated as low. We have explained this by particularities of chord progression, which impact the listener's feeling of complexity and acceptability. We have evaluated the same data twice with new participants in 2018 and with the same participants for the third time in 2019. These three evaluations have shown that the same 53 musical excerpts, found to be difficult and complex in the study conducted in 2017, are exhibiting a high feeling of complexity again. It was proposed that the content of these musical excerpts, defined as “irregular,” is not meeting the listener's expectancy and the basic perceptual principles, creating a higher feeling of difficulty and complexity. As the “irregularities” in these 53 musical excerpts seem to be perceived by the participants without being aware of it, affecting the pleasantness and the feeling of complexity, they have been defined as “subliminal irregularities” and the 53 musical excerpts as “irregular.” In our recent study (2019) of the same data (used in previous research works), we have proposed a new measure of the complexity of harmony, “regularity,” based on the irregularities in the harmonic progression and other plausible particularities in the musical structure found in previous studies. We have in this study also proposed a list of 10 different particularities for which we were assuming that they are impacting the participant’s perception of complexity in harmony. These ten particularities have been tested in this paper, by extending the analysis in our 53 irregular musical excerpts from harmony to melody. In the examining of melody, we have used the computational model “Information Dynamics of Music” (IDyOM) and two information-theoretic measures: entropy - the uncertainty of the prediction before the next event is heard, and information content - the unexpectedness of an event in a sequence. In order to describe the features of melody in these musical examples, we have used four different viewpoints: pitch, interval, duration, scale degree. The results have shown that the texture of melody (e.g., multiple voices, homorhythmic structure) and structure of melody (e.g., huge interval leaps, syncopated rhythm, implied harmony in compound melodies) in these musical excerpts are impacting the participant’s perception of complexity. High information content values were found in compound melodies in which implied harmonies seem to have suggested additional harmonies, affecting the participant’s perception of the chord progression in harmony by creating a sense of an ambiguous musical structure.

Keywords: entropy and information content, harmony, subliminal (ir)regularity, IDyOM

Procedia PDF Downloads 106
6001 Digital Watermarking Based on Visual Cryptography and Histogram

Authors: R. Rama Kishore, Sunesh

Abstract:

Nowadays, robust and secure watermarking algorithm and its optimization have been need of the hour. A watermarking algorithm is presented to achieve the copy right protection of the owner based on visual cryptography, histogram shape property and entropy. In this, both host image and watermark are preprocessed. Host image is preprocessed by using Butterworth filter, and watermark is with visual cryptography. Applying visual cryptography on water mark generates two shares. One share is used for embedding the watermark, and the other one is used for solving any dispute with the aid of trusted authority. Usage of histogram shape makes the process more robust against geometric and signal processing attacks. The combination of visual cryptography, Butterworth filter, histogram, and entropy can make the algorithm more robust, imperceptible, and copy right protection of the owner.

Keywords: digital watermarking, visual cryptography, histogram, butter worth filter

Procedia PDF Downloads 324
6000 An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN

Authors: Yang Zhou, Kangfeng Zheng, Wei Ni, Ren Ping Liu

Abstract:

Software-defined networking (SDN) provides a solution for scalable network framework with decoupled control and data plane. However, this architecture also induces a particular distributed denial-of-service (DDoS) attack that can affect or even overwhelm the SDN network. DDoS attack detection problem has to date been mostly researched as entropy comparison problem. However, this problem lacks the utilization of SDN, and the results are not accurate. In this paper, we propose a DDoS attack detection method, which interprets DDoS detection as a signature matching problem and is formulated as Earth Mover’s Distance (EMD) model. Considering the feasibility and accuracy, we further propose to define the cost function of EMD to be a generalized Kullback-Leibler divergence. Simulation results show that our proposed method can detect DDoS attacks by comparing EMD values with the ones computed in the case without attacks. Moreover, our method can significantly increase the true positive rate of detection.

Keywords: DDoS detection, EMD, relative entropy, SDN

Procedia PDF Downloads 303
5999 2D and 3D Unsteady Simulation of the Heat Transfer in the Sample during Heat Treatment by Moving Heat Source

Authors: Zdeněk Veselý, Milan Honner, Jiří Mach

Abstract:

The aim of the performed work is to establish the 2D and 3D model of direct unsteady task of sample heat treatment by moving source employing computer model on the basis of finite element method. The complex boundary condition on heat loaded sample surface is the essential feature of the task. Computer model describes heat treatment of the sample during heat source movement over the sample surface. It is started from the 2D task of sample cross section as a basic model. Possibilities of extension from 2D to 3D task are discussed. The effect of the addition of third model dimension on the temperature distribution in the sample is showed. Comparison of various model parameters on the sample temperatures is observed. Influence of heat source motion on the depth of material heat treatment is shown for several velocities of the movement. Presented computer model is prepared for the utilization in laser treatment of machine parts.

Keywords: computer simulation, unsteady model, heat treatment, complex boundary condition, moving heat source

Procedia PDF Downloads 361
5998 Real-Time Episodic Memory Construction for Optimal Action Selection in Cognitive Robotics

Authors: Deon de Jager, Yahya Zweiri, Dimitrios Makris

Abstract:

The three most important components in the cognitive architecture for cognitive robotics is memory representation, memory recall, and action-selection performed by the executive. In this paper, action selection, performed by the executive, is defined as a memory quantification and optimization process. The methodology describes the real-time construction of episodic memory through semantic memory optimization. The optimization is performed by set-based particle swarm optimization, using an adaptive entropy memory quantification approach for fitness evaluation. The performance of the approach is experimentally evaluated by simulation, where a UAV is tasked with the collection and delivery of a medical package. The experiments show that the UAV dynamically uses the episodic memory to autonomously control its velocity, while successfully completing its mission.

Keywords: cognitive robotics, semantic memory, episodic memory, maximum entropy principle, particle swarm optimization

Procedia PDF Downloads 119
5997 Prediction of the Thermodynamic Properties of Hydrocarbons Using Gaussian Process Regression

Authors: N. Alhazmi

Abstract:

Knowing the thermodynamics properties of hydrocarbons is vital when it comes to analyzing the related chemical reaction outcomes and understanding the reaction process, especially in terms of petrochemical industrial applications, combustions, and catalytic reactions. However, measuring the thermodynamics properties experimentally is time-consuming and costly. In this paper, Gaussian process regression (GPR) has been used to directly predict the main thermodynamic properties - standard enthalpy of formation, standard entropy, and heat capacity -for more than 360 cyclic and non-cyclic alkanes, alkenes, and alkynes. A simple workflow has been proposed that can be applied to directly predict the main properties of any hydrocarbon by knowing its descriptors and chemical structure and can be generalized to predict the main properties of any material. The model was evaluated by calculating the statistical error R², which was more than 0.9794 for all the predicted properties.

Keywords: thermodynamic, Gaussian process regression, hydrocarbons, regression, supervised learning, entropy, enthalpy, heat capacity

Procedia PDF Downloads 188
5996 The Bayesian Premium Under Entropy Loss

Authors: Farouk Metiri, Halim Zeghdoudi, Mohamed Riad Remita

Abstract:

Credibility theory is an experience rating technique in actuarial science which can be seen as one of quantitative tools that allows the insurers to perform experience rating, that is, to adjust future premiums based on past experiences. It is used usually in automobile insurance, worker's compensation premium, and IBNR (incurred but not reported claims to the insurer) where credibility theory can be used to estimate the claim size amount. In this study, we focused on a popular tool in credibility theory which is the Bayesian premium estimator, considering Lindley distribution as a claim distribution. We derive this estimator under entropy loss which is asymmetric and squared error loss which is a symmetric loss function with informative and non-informative priors. In a purely Bayesian setting, the prior distribution represents the insurer’s prior belief about the insured’s risk level after collection of the insured’s data at the end of the period. However, the explicit form of the Bayesian premium in the case when the prior is not a member of the exponential family could be quite difficult to obtain as it involves a number of integrations which are not analytically solvable. The paper finds a solution to this problem by deriving this estimator using numerical approximation (Lindley approximation) which is one of the suitable approximation methods for solving such problems, it approaches the ratio of the integrals as a whole and produces a single numerical result. Simulation study using Monte Carlo method is then performed to evaluate this estimator and mean squared error technique is made to compare the Bayesian premium estimator under the above loss functions.

Keywords: bayesian estimator, credibility theory, entropy loss, monte carlo simulation

Procedia PDF Downloads 298
5995 A Relative Entropy Regularization Approach for Fuzzy C-Means Clustering Problem

Authors: Ouafa Amira, Jiangshe Zhang

Abstract:

Clustering is an unsupervised machine learning technique; its aim is to extract the data structures, in which similar data objects are grouped in the same cluster, whereas dissimilar objects are grouped in different clusters. Clustering methods are widely utilized in different fields, such as: image processing, computer vision , and pattern recognition, etc. Fuzzy c-means clustering (fcm) is one of the most well known fuzzy clustering methods. It is based on solving an optimization problem, in which a minimization of a given cost function has been studied. This minimization aims to decrease the dissimilarity inside clusters, where the dissimilarity here is measured by the distances between data objects and cluster centers. The degree of belonging of a data point in a cluster is measured by a membership function which is included in the interval [0, 1]. In fcm clustering, the membership degree is constrained with the condition that the sum of a data object’s memberships in all clusters must be equal to one. This constraint can cause several problems, specially when our data objects are included in a noisy space. Regularization approach took a part in fuzzy c-means clustering technique. This process introduces an additional information in order to solve an ill-posed optimization problem. In this study, we focus on regularization by relative entropy approach, where in our optimization problem we aim to minimize the dissimilarity inside clusters. Finding an appropriate membership degree to each data object is our objective, because an appropriate membership degree leads to an accurate clustering result. Our clustering results in synthetic data sets, gaussian based data sets, and real world data sets show that our proposed model achieves a good accuracy.

Keywords: clustering, fuzzy c-means, regularization, relative entropy

Procedia PDF Downloads 240
5994 Influence of Sintering Temperature on Microhardness and Tribological Properties of Equi-Atomic Ti-Al-Mo-Si-W Multicomponent Alloy

Authors: Rudolf L. Kanyane, Nicolaus Malatji, Patritia A. Popoola

Abstract:

Tribological failure of materials during application can lead to catastrophic events which also carry economic penalties. High entropy alloys (HEAs) have shown outstanding tribological properties in applications such as mechanical parts were moving parts under high friction are required. This work aims to investigate the effect of sintering temperature on microhardness properties and tribological properties of novel equiatomic TiAlMoSiW HEAs fabricated via spark plasma sintering. The effect of Spark plasma sintering temperature on morphological evolution and phase formation was also investigated. The microstructure and the phases formed for the developed HEAs were examined using scanning electron microscopy (SEM) and X-ray diffractometry (XRD) respectively. The microhardness and tribological properties were studied using a diamond base microhardness tester Rtec tribometer. The developed HEAs showed improved mechanical properties as the sintering temperature increases.

Keywords: sintering, high entropy alloy, microhardness, tribology

Procedia PDF Downloads 106
5993 Study on Hydrogen Isotope Permeability of High Entropy Alloy Coating

Authors: Long Wang, Yongjin Feng, Xiaofang Luo

Abstract:

Tritium permeation through structural materials is a significant issue for fusion demonstration (DEMO) reactor blankets in terms of fuel cycle efficiency and radiological safety. Reduced activation ferritic (RAFM) steel CLF-1 is a prime candidate for the China’s CFETR blanket structural material, facing high permeability of hydrogen isotopes at reactor operational temperature. To confine tritium as much as possible in the reactor, surface modification of the steels including fabrication of tritium permeation barrier (TPB) attracts much attention. As a new alloy system, high entropy alloy (HEA) contains at least five principal elements, each of which ranges from 5 at% to 35 at%. This high mixing effect entitles HEA extraordinary comprehensive performance. So it is attractive to lead HEA into surface alloying for protective use. At present, studies on the hydrogen isotope permeability of HEA coatings is still insufficient and corresponding mechanism isn’t clear. In our study, we prepared three kinds of HEA coatings, including AlCrTaTiZr, (AlCrTaTiZr)N and (AlCrTaTiZr)O. After comprehensive characterization of SEM, XPS, AFM, XRD and TEM, the structure and composition of the HEA coatings were obtained. Deuterium permeation tests were conducted to evaluate the hydrogen isotope permeability of AlCrTaTiZr, (AlCrTaTiZr)N and (AlCrTaTiZr)O HEA coatings. Results proved that the (AlCrTaTiZr)N and (AlCrTaTiZr)O HEA coatings had better hydrogen isotope permeation resistance. Through analyzing and characterizing the hydrogen isotope permeation results of the corroded samples, an internal link between hydrogen isotope permeation behavior and structure of HEA coatings was established. The results provide valuable reference in engineering design of structural and TPB materials for future fusion device.

Keywords: high entropy alloy, hydrogen isotope permeability, tritium permeation barrier, fusion demonstration reactor

Procedia PDF Downloads 142
5992 Understanding the Processwise Entropy Framework in a Heat-powered Cooling Cycle

Authors: P. R. Chauhan, S. K. Tyagi

Abstract:

Adsorption refrigeration technology offers a sustainable and energy-efficient cooling alternative over traditional refrigeration technologies for meeting the fast-growing cooling demands. With its ability to utilize natural refrigerants, low-grade heat sources, and modular configurations, it has the potential to revolutionize the cooling industry. Despite these benefits, the commercial viability of this technology is hampered by several fundamental limiting constraints, including its large size, low uptake capacity, and poor performance as a result of deficient heat and mass transfer characteristics. The primary cause of adequate heat and mass transfer characteristics and magnitude of exergy loss in various real processes of adsorption cooling system can be assessed by the entropy generation rate analysis, i. e. Second law of Thermodynamics. Therefore, this article presents the second law of thermodynamic-based investigation in terms of entropy generation rate (EGR) to identify the energy losses in various processes of the HPCC-based adsorption system using MATLAB R2021b software. The adsorption technology-based cooling system consists of two beds made up of silica gel and arranged in a single stage, while the water is employed as a refrigerant, coolant, and hot fluid. The variation in process-wise EGR is examined corresponding to cycle time, and a comparative analysis is also presented. Moreover, the EGR is also evaluated in the external units, such as the heat source and heat sink unit used for regeneration and heat dump, respectively. The research findings revealed that the combination of adsorber and desorber, which operates across heat reservoirs with a higher temperature gradient, shares more than half of the total amount of EGR. Moreover, the EGR caused by the heat transfer process is determined to be the highest, followed by a heat sink, heat source, and mass transfer, respectively. in case of heat transfer process, the operation of the valve is determined to be responsible for more than half (54.9%) of the overall EGR during the heat transfer. However, the combined contribution of the external units, such as the source (18.03%) and sink (21.55%), to the total EGR, is 35.59%. The analysis and findings of the present research are expected to pinpoint the source of the energy waste in HPCC based adsorption cooling systems.

Keywords: adsorption cooling cycle, heat transfer, mass transfer, entropy generation, silica gel-water

Procedia PDF Downloads 76
5991 A Theorem Related to Sample Moments and Two Types of Moment-Based Density Estimates

Authors: Serge B. Provost

Abstract:

Numerous statistical inference and modeling methodologies are based on sample moments rather than the actual observations. A result justifying the validity of this approach is introduced. More specifically, it will be established that given the first n moments of a sample of size n, one can recover the original n sample points. This implies that a sample of size n and its first associated n moments contain precisely the same amount of information. However, it is efficient to make use of a limited number of initial moments as most of the relevant distributional information is included in them. Two types of density estimation techniques that rely on such moments will be discussed. The first one expresses a density estimate as the product of a suitable base density and a polynomial adjustment whose coefficients are determined by equating the moments of the density estimate to the sample moments. The second one assumes that the derivative of the logarithm of a density function can be represented as a rational function. This gives rise to a system of linear equations involving sample moments, the density estimate is then obtained by solving a differential equation. Unlike kernel density estimation, these methodologies are ideally suited to model ‘big data’ as they only require a limited number of moments, irrespective of the sample size. What is more, they produce simple closed form expressions that are amenable to algebraic manipulations. They also turn out to be more accurate as will be shown in several illustrative examples.

Keywords: density estimation, log-density, polynomial adjustments, sample moments

Procedia PDF Downloads 129
5990 Utilizing Waste Heat from Thermal Power Plants to Generate Power by Modelling an Atmospheric Vortex Engine

Authors: Mohammed Nabeel Khan, C. Perisamy

Abstract:

Convective vortices are normal highlights of air that ingest lower-entropy-energy at higher temperatures than they dismiss higher-entropy-energy to space. By means of the thermodynamic proficiency, it has been anticipated that the force of convective vortices relies upon the profundity of the convective layer. The atmospheric vortex engine is proposed as a gadget for delivering mechanical energy by methods for artificially produced vortex. The task of the engine is in view of the certainties that the environment is warmed from the base and cooled from the top. By generation of the artificial vortex, it is planned to take out the physical solar updraft tower and decrease the capital of the solar chimney power plants. The study shows the essentials of the atmospheric vortex engine, furthermore, audits the cutting edge in subject. Moreover, the study talks about a thought on using the solar energy as heat source to work the framework. All in all, the framework is attainable and promising for electrical power production.

Keywords: AVE, atmospheric vortex engine, atmosphere, updraft, vortex

Procedia PDF Downloads 128
5989 The Cardiac Diagnostic Prediction Applied to a Designed Holter

Authors: Leonardo Juan Ramírez López, Javier Oswaldo Rodriguez Velasquez

Abstract:

We have designed a Holter that measures the heart´s activity for over 24 hours, implemented a prediction methodology, and generate alarms as well as indicators to patients and treating physicians. Various diagnostic advances have been developed in clinical cardiology thanks to Holter implementation; however, their interpretation has largely been conditioned to clinical analysis and measurements adjusted to diverse population characteristics, thus turning it into a subjective examination. This, however, requires vast population studies to be validated that, in turn, have not achieved the ultimate goal: mortality prediction. Given this context, our Insight Research Group developed a mathematical methodology that assesses cardiac dynamics through entropy and probability, creating a numerical and geometrical attractor which allows quantifying the normalcy of chronic and acute disease as well as the evolution between such states, and our Tigum Research Group developed a holter device with 12 channels and advanced computer software. This has been shown in different contexts with 100% sensitivity and specificity results.

Keywords: attractor , cardiac, entropy, holter, mathematical , prediction

Procedia PDF Downloads 136
5988 Spatio-temporal Distribution of the Groundwater Quality in the El Milia Plain, Kebir Rhumel Basin, Algeria

Authors: Lazhar Belkhiri, Ammar Tiri, Lotfi Mouni

Abstract:

In this research, we analyzed the groundwater quality index in the El Milia plain, Kebir Rhumel Basin, Algeria. Thirty-three groundwater samples were collected from wells in the El Milia plain during April 2015. In this study, pH and electrical conductivity (EC) were conducted at each sampling well. Eight hydrochemical parameters such as calcium (Ca), magnesium (Mg), sodium (Na), potassium (K), chlorid (Cl), sulfate (SO4), bicarbonate (HCO3), and Nnitrate (NO3) were analysed. The entropy water quality index (EWQI) method was employed to evaluate the groundwater quality in the study area. Moran’s I and the ordinary kriging (OK) interpolation technique were used to examine the spatial distribution pattern of the hydrochemical parameters in the groundwater. It was found that the hydrochemical parameters Ca, Cl, and HCO3 showed strong spatial autocorrelation in the El Milia plain, indicating a spatial dependence and clustering of these parameters in the groundwater. The groundwater quality was evaluated using the entropy water quality index (EWQI). The results showed that approximately 86% of the total groundwater samples in the study area fall within the moderate groundwater quality category. The spatial map of the EWQI values indicated an increasing trend from the south-west to the northeast, following the direction of groundwater flow. The highest EWQI values were observed near El Milia city in the center of the plain. This spatial pattern suggests variations in groundwater quality across the study area, with potentially higher risks near the city center. Therefore, the results obtained in this research provide very useful information to decision-makers.

Keywords: entropy water quality index (EWQI), moran’s i, ordinary kriging interpolation, el milia plain

Procedia PDF Downloads 22
5987 The Effect of Non-Normality on CB-SEM and PLS-SEM Path Estimates

Authors: Z. Jannoo, B. W. Yap, N. Auchoybur, M. A. Lazim

Abstract:

The two common approaches to Structural Equation Modeling (SEM) are the Covariance-Based SEM (CB-SEM) and Partial Least Squares SEM (PLS-SEM). There is much debate on the performance of CB-SEM and PLS-SEM for small sample size and when distributions are non-normal. This study evaluates the performance of CB-SEM and PLS-SEM under normality and non-normality conditions via a simulation. Monte Carlo Simulation in R programming language was employed to generate data based on the theoretical model with one endogenous and four exogenous variables. Each latent variable has three indicators. For normal distributions, CB-SEM estimates were found to be inaccurate for small sample size while PLS-SEM could produce the path estimates. Meanwhile, for a larger sample size, CB-SEM estimates have lower variability compared to PLS-SEM. Under non-normality, CB-SEM path estimates were inaccurate for small sample size. However, CB-SEM estimates are more accurate than those of PLS-SEM for sample size of 50 and above. The PLS-SEM estimates are not accurate unless sample size is very large.

Keywords: CB-SEM, Monte Carlo simulation, normality conditions, non-normality, PLS-SEM

Procedia PDF Downloads 371
5986 The Microstructure and Corrosion Behavior of High Entropy Metallic Layers Electrodeposited by Low and High-Temperature Methods

Authors: Zbigniew Szklarz, Aldona Garbacz-Klempka, Magdalena Bisztyga-Szklarz

Abstract:

Typical metallic alloys bases on one major alloying component, where the addition of other elements is intended to improve or modify certain properties, most of all the mechanical properties. However, in 1995 a new concept of metallic alloys was described and defined. High Entropy Alloys (HEA) contains at least five alloying elements in an amount from 5 to 20 at.%. A common feature this type of alloys is an absence of intermetallic phases, high homogeneity of the microstructure and unique chemical composition, what leads to obtaining materials with very high strength indicators, stable structures (also at high temperatures) and excellent corrosion resistance. Hence, HEA can be successfully used as a substitutes for typical metallic alloys in various applications where a sufficiently high properties are desirable. For fabricating HEA, a few ways are applied: 1/ from liquid phase i.e. casting (usually arc melting); 2/ from solid phase i.e. powder metallurgy (sintering methods preceded by mechanical synthesis) and 3/ from gas phase e.g. sputtering or 4/ other deposition methods like electrodeposition from liquids. Application of different production methods creates different microstructures of HEA, which can entail differences in their properties. The last two methods also allows to obtain coatings with HEA structures, hereinafter referred to as High Entropy Films (HEF). With reference to above, the crucial aim of this work was the optimization of the manufacturing process of the multi-component metallic layers (HEF) by the low- and high temperature electrochemical deposition ( ED). The low-temperature deposition process was crried out at ambient or elevated temperature (up to 100 ᵒC) in organic electrolyte. The high-temperature electrodeposition (several hundred Celcius degrees), in turn, allowed to form the HEF layer by electrochemical reduction of metals from molten salts. The basic chemical composition of the coatings was CoCrFeMnNi (known as Cantor’s alloy). However, it was modified by other, selected elements like Al or Cu. The optimization of the parameters that allow to obtain as far as it possible homogeneous and equimolar composition of HEF is the main result of presented studies. In order to analyse and compare the microstructure, SEM/EBSD, TEM and XRD techniques were employed. Morover, the determination of corrosion resistance of the CoCrFeMnNi(Cu or Al) layers in selected electrolytes (i.e. organic and non-organic liquids) was no less important than the above mentioned objectives.

Keywords: high entropy alloys, electrodeposition, corrosion behavior, microstructure

Procedia PDF Downloads 53
5985 Entropy Generation Analyze Due to the Steady Natural Convection of Newtonian Fluid in a Square Enclosure

Authors: T. T. Naas, Y. Lasbet, C. Kezrane

Abstract:

The thermal control in many systems is widely accomplished applying mixed convection process due to its low cost, reliability and easy maintenance. Typical applications include the aircraft electronic equipment, rotating-disc heat exchangers, turbo machinery, and nuclear reactors, etc. Natural convection in an inclined square enclosure heated via wall heater has been studied numerically. Finite volume method is used for solving momentum and energy equations in the form of stream function–vorticity. The right and left walls are kept at a constant temperature, while the other parts are adiabatic. The range of the inclination angle covers a whole revolution. The method is validated for a vertical cavity. A general power law dependence of the Nusselt number with respect to the Rayleigh number with the coefficient and exponent as functions of the inclination angle is presented. For a fixed Rayleigh number, the inclination angle increases or decreases is found.

Keywords: natural convection in enclosure, inclined enclosure, Nusselt number, entropy generation analyze

Procedia PDF Downloads 221
5984 Hub Traveler Guidance Signage Evaluation via Panoramic Visualization Using Entropy Weight Method and TOPSIS

Authors: Si-yang Zhang, Chi Zhao

Abstract:

Comprehensive transportation hubs are important nodes of the transportation network, and their internal signage the functions as guidance and distribution assistance, which directly affects the operational efficiency of traffic in and around the hubs. Reasonably installed signage effectively attracts the visual focus of travelers and improves wayfinding efficiency. Among the elements of signage, the visual guidance effect is the key factor affecting the information conveyance, whom should be evaluated during design and optimization process. However, existing evaluation methods mostly focus on the layout, and are not able to fully understand if signage caters travelers’ need. This study conducted field investigations and developed panoramic videos for multiple transportation hubs in China, and designed survey accordingly. Human subjects are recruited to watch panoramic videos via virtual reality (VR) and respond to the surveys. In this paper, Pudong Airport and Xi'an North Railway Station were studied and compared as examples due to their high traveler volume and relatively well-developed traveler service systems. Visual attention was captured by eye tracker and subjective satisfaction ratings were collected through surveys. Entropy Weight Method (EWM) was utilized to evaluate the effectiveness of signage elements and Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) was used to further rank the importance of the elements. The results show that the degree of visual attention of travelers significantly affects the evaluation results of guidance signage. Key factors affecting visual attention include accurate legibility, obstruction and defacement rates, informativeness, and whether signage is set up in a hierarchical manner.

Keywords: traveler guidance signage, panoramic video, visual attention, entropy weight method, TOPSIS

Procedia PDF Downloads 32
5983 Application of Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Multipoint Optimal Minimum Entropy Deconvolution in Railway Bearings Fault Diagnosis

Authors: Yao Cheng, Weihua Zhang

Abstract:

Although the measured vibration signal contains rich information on machine health conditions, the white noise interferences and the discrete harmonic coming from blade, shaft and mash make the fault diagnosis of rolling element bearings difficult. In order to overcome the interferences of useless signals, a new fault diagnosis method combining Complete Ensemble Empirical Mode Decomposition with adaptive noise (CEEMDAN) and Multipoint Optimal Minimum Entropy Deconvolution (MOMED) is proposed for the fault diagnosis of high-speed train bearings. Firstly, the CEEMDAN technique is applied to adaptively decompose the raw vibration signal into a series of finite intrinsic mode functions (IMFs) and a residue. Compared with Ensemble Empirical Mode Decomposition (EEMD), the CEEMDAN can provide an exact reconstruction of the original signal and a better spectral separation of the modes, which improves the accuracy of fault diagnosis. An effective sensitivity index based on the Pearson's correlation coefficients between IMFs and raw signal is adopted to select sensitive IMFs that contain bearing fault information. The composite signal of the sensitive IMFs is applied to further analysis of fault identification. Next, for propose of identifying the fault information precisely, the MOMED is utilized to enhance the periodic impulses in composite signal. As a non-iterative method, the MOMED has better deconvolution performance than the classical deconvolution methods such Minimum Entropy Deconvolution (MED) and Maximum Correlated Kurtosis Deconvolution (MCKD). Third, the envelope spectrum analysis is applied to detect the existence of bearing fault. The simulated bearing fault signals with white noise and discrete harmonic interferences are used to validate the effectiveness of the proposed method. Finally, the superiorities of the proposed method are further demonstrated by high-speed train bearing fault datasets measured from test rig. The analysis results indicate that the proposed method has strong practicability.

Keywords: bearing, complete ensemble empirical mode decomposition with adaptive noise, fault diagnosis, multipoint optimal minimum entropy deconvolution

Procedia PDF Downloads 339
5982 Frank Norris’ McTeague: An Entropic Melodrama

Authors: Mohsen Masoomi, Fazel Asadi Amjad, Monireh Arvin

Abstract:

According to Naturalistic principles, human destiny in the form of blind chance and determinism, entraps the individual, so man is a defenceless creature unable to escape from the ruthless paws of a stoical universe. In Naturalism; nonetheless, melodrama mirrors a conscious alternative with a peculiar function. A typical American Naturalistic character thus cannot be a subject for social criticism of American society since they are not victims of the ongoing virtual slavery, capitalist system, nor of a ruined milieu, but of their own volition, and more importantly, their character frailty. Through a Postmodern viewpoint, each Naturalistic work can encompass some entropic trends and changes culminating in an entire failure and devastation. Frank Norris in McTeague displays the futile struggles of ordinary men and how they end up brutes. McTeague encompasses intoxication, abuse, violation, and ruthless homicides. Norris’ depictions of the falling individual as a demon represent the entropic dimension of Naturalistic novels. McTeague’s defeat is somewhat his own fault, the result of his own blunders and resolution, not the result of sheer accident. Throughout the novel, each character is a kind of insane quester indicating McTeague’s decadence and, by inference, the decadence of Western civilisation. McTeague seems to designate Norris’ solicitude for a community fabricated by the elements of human negative demeanours and conducts hauling acute symptoms of infectious dehumanisation. The aim of this article is to illustrate how one specific negative human disposition gradually, like a running fire, can spread everywhere and burn everything in itself. The author applies the concept of entropy metaphorically to describe the individual devolutions that necessarily comprise community entropy in McTeague, a dying universe.

Keywords: animal imagery, entropy, Gypsy, melodrama

Procedia PDF Downloads 256
5981 The Relationship between Rhythmic Complexity and Listening Engagement as a Proxy for Perceptual Interest

Authors: Noah R. Fram

Abstract:

Although it has been confirmed by multiple studies, the inverted-U relationship between stimulus complexity and preference (liking) remains contentious. Research aimed at substantiating the model are largely reliant upon anecdotal self-assessments of subjects and basic measures of complexity, leaving potential confounds unresolved. This study attempts to address the topic by assessing listening time as a behavioral correlate of liking (with the assumption that engagement prolongs listening time) and by looking for latent factors underlying several measures of rhythmic complexity. Participants listened to groups of rhythms, stopping each one when they started to lose interest and were asked to rate each rhythm in each group in terms of interest, complexity, and preference. Subjects were not informed that the time spent listening to each rhythm was the primary measure of interest. The hypothesis that listening time does demonstrate the same inverted-U relationship with complexity as verbal reports of liking was confirmed using a variety of metrics for rhythmic complexity, including meter-dependent measures of syncopation and meter-independent measures of entropy.

Keywords: complexity, entropy, rhythm, syncopation

Procedia PDF Downloads 143
5980 Determination Power and Sample Size Zero-Inflated Negative Binomial Dependent Death Rate of Age Model (ZINBD): Regression Analysis Mortality Acquired Immune Deficiency De ciency Syndrome (AIDS)

Authors: Mohd Asrul Affendi Bin Abdullah

Abstract:

Sample size calculation is especially important for zero inflated models because a large sample size is required to detect a significant effect with this model. This paper verify how to present percentage of power approximation for categorical and then extended to zero inflated models. Wald test was chosen to determine power sample size of AIDS death rate because it is frequently used due to its approachability and its natural for several major recent contribution in sample size calculation for this test. Power calculation can be conducted when covariates are used in the modeling ‘excessing zero’ data and assist categorical covariate. Analysis of AIDS death rate study is used for this paper. Aims of this study to determine the power of sample size (N = 945) categorical death rate based on parameter estimate in the simulation of the study.

Keywords: power sample size, Wald test, standardize rate, ZINBDR

Procedia PDF Downloads 409
5979 The Analysis of Changes in Urban Hierarchy of Isfahan Province in the Fifty-Year Period (1956-2006)

Authors: Hamidreza Joudaki, Yousefali Ziari

Abstract:

The appearance of city and urbanism is one of the important processes which have affected social communities. Being industrialized urbanism developed along with each other in the history. In addition, they have had simple relationship for more than six thousand years, that is, from the appearance of the first cities. In 18th century by coming out of industrial capitalism, progressive development took place in urbanism in the world. In Iran, the city of each region made its decision by itself and the capital of region (downtown) was the only central part and also the regional city without any hierarchy, controlled its realm. However, this method of ruling during these three decays, because of changing in political, social and economic issues that have caused changes in rural and urban relationship. Moreover, it has changed the variety of performance of cities and systematic urban network in Iran. Today, urban system has very vast imbalanced apace and performance. In Isfahan, the trend of urbanism is like the other part of Iran and systematic urban hierarchy is not suitable and normal. This article is a quantitative and analytical. The statistical communities are Isfahan Province cities and the changes in urban network and its hierarchy during the period of fifty years (1956 -2006) has been surveyed. In addition, those data have been analyzed by model of Rank and size and Entropy index. In this article Iran cities and also the factor of entropy of primate city and urban hierarchy of Isfahan Province have been introduced. Urban residents of this Province have been reached from 55 percent to 83% (2006). As we see the analytical data reflects that there is mismatching and imbalance between cities. Because the entropy index was.91 in 1956.And it decreased to.63 in 2006. Isfahan city is the primate city in the whole of these periods. Moreover, the second and the third cities have population gap with regard to the other cities and finally, they do not follow the system of rank-size.

Keywords: urban network, urban hierarchy, primate city, Isfahan province, urbanism, first cities

Procedia PDF Downloads 222
5978 Image Encryption Using Eureqa to Generate an Automated Mathematical Key

Authors: Halima Adel Halim Shnishah, David Mulvaney

Abstract:

Applying traditional symmetric cryptography algorithms while computing encryption and decryption provides immunity to secret keys against different attacks. One of the popular techniques generating automated secret keys is evolutionary computing by using Eureqa API tool, which got attention in 2013. In this paper, we are generating automated secret keys for image encryption and decryption using Eureqa API (tool which is used in evolutionary computing technique). Eureqa API models pseudo-random input data obtained from a suitable source to generate secret keys. The validation of generated secret keys is investigated by performing various statistical tests (histogram, chi-square, correlation of two adjacent pixels, correlation between original and encrypted images, entropy and key sensitivity). Experimental results obtained from methods including histogram analysis, correlation coefficient, entropy and key sensitivity, show that the proposed image encryption algorithms are secure and reliable, with the potential to be adapted for secure image communication applications.

Keywords: image encryption algorithms, Eureqa, statistical measurements, automated key generation

Procedia PDF Downloads 455
5977 Gray Level Image Encryption

Authors: Roza Afarin, Saeed Mozaffari

Abstract:

The aim of this paper is image encryption using Genetic Algorithm (GA). The proposed encryption method consists of two phases. In modification phase, pixels locations are altered to reduce correlation among adjacent pixels. Then, pixels values are changed in the diffusion phase to encrypt the input image. Both phases are performed by GA with binary chromosomes. For modification phase, these binary patterns are generated by Local Binary Pattern (LBP) operator while for diffusion phase binary chromosomes are obtained by Bit Plane Slicing (BPS). Initial population in GA includes rows and columns of the input image. Instead of subjective selection of parents from this initial population, a random generator with predefined key is utilized. It is necessary to decrypt the coded image and reconstruct the initial input image. Fitness function is defined as average of transition from 0 to 1 in LBP image and histogram uniformity in modification and diffusion phases, respectively. Randomness of the encrypted image is measured by entropy, correlation coefficients and histogram analysis. Experimental results show that the proposed method is fast enough and can be used effectively for image encryption.

Keywords: correlation coefficients, genetic algorithm, image encryption, image entropy

Procedia PDF Downloads 302
5976 Study of the Impact of Synthesis Method and Chemical Composition on Photocatalytic Properties of Cobalt Ferrite Catalysts

Authors: Katerina Zaharieva, Vicente Rives, Martin Tsvetkov, Raquel Trujillano, Boris Kunev, Ivan Mitov, Maria Milanova, Zara Cherkezova-Zheleva

Abstract:

The nanostructured cobalt ferrite-type materials Sample A - Co0.25Fe2.75O4, Sample B - Co0.5Fe2.5O4, and Sample C - CoFe2O4 were prepared by co-precipitation in our previous investigations. The co-precipitated Sample B and Sample C were mechanochemically activated in order to produce Sample D - Co0.5Fe2.5O4 and Sample E- CoFe2O4. The PXRD, Moessbauer and FTIR spectroscopies, specific surface area determination by the BET method, thermal analysis, element chemical analysis and temperature-programmed reduction were used to investigate the prepared nano-sized samples. The changes of the Malachite green dye concentration during reaction of the photocatalytic decolorization using nanostructured cobalt ferrite-type catalysts with different chemical composition are included. The photocatalytic results show that the increase in the degree of incorporation of cobalt ions in the magnetite host structure for co-precipitated cobalt ferrite-type samples results in an increase of the photocatalytic activity: Sample A (4 х10-3 min-1) < Sample B (5 х10-3 min-1) < Sample C (7 х10-3 min-1). Mechanochemically activated photocatalysts showed a higher activity than the co-precipitated ferrite materials: Sample D (16 х10-3 min-1) > Sample E (14 х10-3 min-1) > Sample C (7 х10-3 min-1) > Sample B (5 х10-3 min-1) > Sample A (4 х10-3 min-1). On decreasing the degree of substitution of iron ions by cobalt ones a higher sorption ability of the dye after the dark period for the co-precipitated cobalt ferrite materials was observed: Sample C (72 %) < Sample B (78 %) < Sample A (80 %). Mechanochemically treated ferrite catalysts and co-precipitated Sample B possess similar sorption capacities, Sample D (78 %) ~ Sample E (78 %) ~ Sample B (78 %). The prepared nano-sized cobalt ferrite-type materials demonstrate good photocatalytic and sorption properties. Mechanochemically activated Sample D - Co0.5Fe2.5O4 (16х10-3 min-1) and Sample E-CoFe2O4 (14х10-3 min-1) possess higher photocatalytic activity than that of the most common used UV-light catalyst Degussa P25 (12х10-3 min-1). The dependence of the photo-catalytic activity and sorption properties on the preparation method and different degree of substitution of iron ions by cobalt ions in synthesized cobalt ferrite samples is established. The mechanochemical activation leads to formation of nano-structured cobalt ferrite-type catalysts (Sample D and Sample E) with higher rate constants than those of the ferrite materials (Sample A, Sample B, and Sample C) prepared by the co-precipitation procedure. The increase in the degree of substitution of iron ions by cobalt ones leads to improved photocatalytic properties and lower sorption capacities of the co-precipitated ferrite samples. The good sorption properties between 72 and 80% of the prepared ferrite-type materials show that they could be used as potential cheap absorbents for purification of polluted waters.

Keywords: nanodimensional cobalt ferrites, photocatalyst, synthesis, mechanochemical activation

Procedia PDF Downloads 242