Search results for: bayesian method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19136

Search results for: bayesian method

16076 Using Self Organizing Feature Maps for Classification in RGB Images

Authors: Hassan Masoumi, Ahad Salimi, Nazanin Barhemmat, Babak Gholami

Abstract:

Artificial neural networks have gained a lot of interest as empirical models for their powerful representational capacity, multi input and output mapping characteristics. In fact, most feed-forward networks with nonlinear nodal functions have been proved to be universal approximates. In this paper, we propose a new supervised method for color image classification based on self organizing feature maps (SOFM). This algorithm is based on competitive learning. The method partitions the input space using self-organizing feature maps to introduce the concept of local neighborhoods. Our image classification system entered into RGB image. Experiments with simulated data showed that separability of classes increased when increasing training time. In additional, the result shows proposed algorithms are effective for color image classification.

Keywords: classification, SOFM algorithm, neural network, neighborhood, RGB image

Procedia PDF Downloads 478
16075 DC Bus Voltage Ripple Control of Photo Voltaic Inverter in Low Voltage Ride-Trough Operation

Authors: Afshin Kadri

Abstract:

Using Renewable Energy Resources (RES) as a type of DG unit is developing in distribution systems. The connection of these generation units to existing AC distribution systems changes the structure and some of the operational aspects of these grids. Most of the RES requires to power electronic-based interfaces for connection to AC systems. These interfaces consist of at least one DC/AC conversion unit. Nowadays, grid-connected inverters must have the required feature to support the grid under sag voltage conditions. There are two curves in these conditions that show the magnitude of the reactive component of current as a function of voltage drop value and the required minimum time value, which must be connected to the grid. This feature is named low voltage ride-through (LVRT). Implementing this feature causes problems in the operation of the inverter that increases the amplitude of high-frequency components of the injected current and working out of maximum power point in the photovoltaic panel connected inverters are some of them. The important phenomenon in these conditions is ripples in the DC bus voltage that affects the operation of the inverter directly and indirectly. The losses of DC bus capacitors which are electrolytic capacitors, cause increasing their temperature and decreasing its lifespan. In addition, if the inverter is connected to the photovoltaic panels directly and has the duty of maximum power point tracking, these ripples cause oscillations around the operating point and decrease the generating energy. Using a bidirectional converter in the DC bus, which works as a buck and boost converter and transfers the ripples to its DC bus, is the traditional method to eliminate these ripples. In spite of eliminating the ripples in the DC bus, this method cannot solve the problem of reliability because it uses an electrolytic capacitor in its DC bus. In this work, a control method is proposed which uses the bidirectional converter as the fourth leg of the inverter and eliminates the DC bus ripples using an injection of unbalanced currents into the grid. Moreover, the proposed method works based on constant power control. In this way, in addition, to supporting the amplitude of grid voltage, it stabilizes its frequency by injecting active power. Also, the proposed method can eliminate the DC bus ripples in deep voltage drops, which cause increasing the amplitude of the reference current more than the nominal current of the inverter. The amplitude of the injected current for the faulty phases in these conditions is kept at the nominal value and its phase, together with the phase and amplitude of the other phases, are adjusted, which at the end, the ripples in the DC bus are eliminated, however, the generated power decreases.

Keywords: renewable energy resources, voltage drop value, DC bus ripples, bidirectional converter

Procedia PDF Downloads 76
16074 Evaluation of Seismic Behavior of Steel Shear Wall with Opening with Hardener and Beam with Reduced Cross Section under Cycle Loading with Finite Element Analysis Method

Authors: Masoud Mahdavi

Abstract:

During an earthquake, the structure is subjected to seismic loads that cause tension in the members of the building. The use of energy dissipation elements in the structure reduces the percentage of seismic forces on the main members of the building (especially the columns). Steel plate shear wall, as one of the most widely used types of energy dissipation element, has evolved today, and regular drilling of its inner plate is one of the common cases. In the present study, using a finite element method, the shear wall of the steel plate is designed as a floor (with dimensions of 447 × 6/246 cm) with Abacus software and in three different modes on which a cyclic load has been applied. The steel shear wall has a horizontal element (beam) with a reduced beam section (RBS). The hole in the interior plate of the models is created in such a way that it has the process of increasing the area, which makes the effect of increasing the surface area of the hole on the seismic performance of the steel shear wall completely clear. In the end, it was found that with increasing the opening level in the steel shear wall (with reduced cross-section beam), total displacement and plastic strain indicators increased, structural capacity and total energy indicators decreased and the Mises Monson stress index did not change much.

Keywords: steel plate shear wall with opening, cyclic loading, reduced cross-section beam, finite element method, Abaqus software

Procedia PDF Downloads 123
16073 Effect of Thermal Energy on Inorganic Coagulation for the Treatment of Industrial Wastewater

Authors: Abhishek Singh, Rajlakshmi Barman, Tanmay Shah

Abstract:

Coagulation is considered to be one of the predominant water treatment processes which improve the cost effectiveness of wastewater. The sole purpose of this experiment on thermal coagulation is to increase the efficiency and the rate of reaction. The process uses renewable sources of energy which comprises of improved and minimized time method in order to eradicate the water scarcity of the regions which are on the brink of depletion. This paper includes the various effects of temperature on the standard coagulation treatment of wastewater and their effect on water quality. In addition, the coagulation is done with the mix of bottom/fly-ash that will act as an adsorbent and removes most of the minor and macro particles by means of adsorption which not only helps to reduce the environmental burden of fly ash but also enhance economic benefit. Also, the method of sand filtration is amalgamated in the process. The sand filter is an environmentally-friendly wastewater treatment method, which is relatively simple and inexpensive. The existing parameters were satisfied with the experimental results obtained in this study and were found satisfactory. The initial turbidity of the wastewater is 162 NTU. The initial temperature of the wastewater is 27 C. The temperature variation of the entire process is 50 C-80 C. The concentration of alum in wastewater is 60mg/L-320mg/L. The turbidity range is 8.31-28.1 NTU after treatment. pH variation is 7.73-8.29. The effective time taken is 10 minutes for thermal mixing and sedimentation. The results indicate that the presence of thermal energy affects the coagulation treatment process. The influence of thermal energy on turbidity is assessed along with renewable energy sources and increase of the rate of reaction of the treatment process.

Keywords: adsorbent, sand filter, temperature, thermal coagulation

Procedia PDF Downloads 321
16072 The Criteria of the Aesthetic Quality of Art: Contemporary Photography

Authors: Artem Surkov

Abstract:

This work is devoted to a problem of aesthetic quality determinism in the context of contemporary art. The object of study is photography regarding as a kind of art which demands specific system of quality marking. Objective: To define aesthetic criteria in photography art. For current searching different kind of texts by such powerful authors like Clement Greenberg and Rosalind Krauss, Theodor Adorno and Herbert Marcuse, Charlott Cotton and Boris Groys, Viktor Miziano and Ekaterina Degot' were analyzed. Before all, there are two different kinds of photography: the classic art photography (by Ansel Adams) and the photography as kind of art (by Andreas Gursky). In this text we are talking about the photography as kind of art. The main principle of current searching is synthesis of two different approaches: modernism and postmodernism. This method helps us to define uniform criteria of aesthetic quality in photography as kind of art. The criteria mentioned in conclusion paragraph are: aesthetic rationality, aesthetic economy, awareness (using photographic technics or references), and intention to go beyond form, practice and method.

Keywords: aesthetic, art, criteria of quality, photography, visually

Procedia PDF Downloads 418
16071 Robust Heart Rate Estimation from Multiple Cardiovascular and Non-Cardiovascular Physiological Signals Using Signal Quality Indices and Kalman Filter

Authors: Shalini Rankawat, Mansi Rankawat, Rahul Dubey, Mazad Zaveri

Abstract:

Physiological signals such as electrocardiogram (ECG) and arterial blood pressure (ABP) in the intensive care unit (ICU) are often seriously corrupted by noise, artifacts, and missing data, which lead to errors in the estimation of heart rate (HR) and incidences of false alarm from ICU monitors. Clinical support in ICU requires most reliable heart rate estimation. Cardiac activity, because of its relatively high electrical energy, may introduce artifacts in Electroencephalogram (EEG), Electrooculogram (EOG), and Electromyogram (EMG) recordings. This paper presents a robust heart rate estimation method by detection of R-peaks of ECG artifacts in EEG, EMG & EOG signals, using energy-based function and a novel Signal Quality Index (SQI) assessment technique. SQIs of physiological signals (EEG, EMG, & EOG) were obtained by correlation of nonlinear energy operator (teager energy) of these signals with either ECG or ABP signal. HR is estimated from ECG, ABP, EEG, EMG, and EOG signals from separate Kalman filter based upon individual SQIs. Data fusion of each HR estimate was then performed by weighing each estimate by the Kalman filters’ SQI modified innovations. The fused signal HR estimate is more accurate and robust than any of the individual HR estimate. This method was evaluated on MIMIC II data base of PhysioNet from bedside monitors of ICU patients. The method provides an accurate HR estimate even in the presence of noise and artifacts.

Keywords: ECG, ABP, EEG, EMG, EOG, ECG artifacts, Teager-Kaiser energy, heart rate, signal quality index, Kalman filter, data fusion

Procedia PDF Downloads 696
16070 Ultrasonic Pulse Velocity Investigation of Polypropylene and Steel Fiber Reinforced Concrete

Authors: Erjola Reufi, Jozefita Marku, Thomas Bier

Abstract:

Ultrasonic pulse velocity (UPV) method has been shown for some time to provide a reliable means of estimating properties and offers a unique opportunity for direct, quick and safe control of building damaged by earthquake, fatigue, conflagration and catastrophic scenarios. On this investigation hybrid reinforced concrete has been investigated by UPV method. Hooked end steel fiber of length 50 and 30 mm was added to concrete in different proportion 0, 0.25, 0.5, and 1 % by the volume of concrete. On the other hand, polypropylene fiber of length 12, 6, 3 mm was added to concrete of 0.1, 0.2, and 0.4 % by the volume of concrete. Fifteen different mixture has been prepared to investigate the relation between compressive strength and UPV values and also to investigate on the effect of volume and type of fiber on UPV values.

Keywords: compressive strength, polypropylene fiber, steel fiber, ultrasonic pulse velocity, volume, type of fiber

Procedia PDF Downloads 402
16069 Optimal Harmonic Filters Design of Taiwan High Speed Rail Traction System

Authors: Ying-Pin Chang

Abstract:

This paper presents a method for combining a particle swarm optimization with nonlinear time-varying evolution and orthogonal arrays (PSO-NTVEOA) in the planning of harmonic filters for the high speed railway traction system with specially connected transformers in unbalanced three-phase power systems. The objective is to minimize the cost of the filter, the filters loss, the total harmonic distortion of currents and voltages at each bus simultaneously. An orthogonal array is first conducted to obtain the initial solution set. The set is then treated as the initial training sample. Next, the PSO-NTVEOA method parameters are determined by using matrix experiments with an orthogonal array, in which a minimal number of experiments would have an effect that approximates the full factorial experiments. This PSO-NTVEOA method is then applied to design optimal harmonic filters in Taiwan High Speed Rail (THSR) traction system, where both rectifiers and inverters with IGBT are used. From the results of the illustrative examples, the feasibility of the PSO-NTVEOA to design an optimal passive harmonic filter of THSR system is verified and the design approach can greatly reduce the harmonic distortion. Three design schemes are compared that V-V connection suppressing the 3rd order harmonic, and Scott and Le Blanc connection for the harmonic improvement is better than the V-V connection.

Keywords: harmonic filters, particle swarm optimization, nonlinear time-varying evolution, orthogonal arrays, specially connected transformers

Procedia PDF Downloads 392
16068 Management and Agreement Protocol in Computer Security

Authors: Abdulameer K. Hussain

Abstract:

When dealing with a cryptographic system we note that there are many activities performed by parties of this cryptographic system and the most prominent of these activities is the process of agreement between the parties involved in the cryptographic system on how to deal and perform the cryptographic system tasks to be more secure, more confident and reliable. The most common agreement among parties is a key agreement and other types of agreements. Despite the fact that there is an attempt from some quarters to find other effective agreement methods but these methods are limited to the traditional agreements. This paper presents different parameters to perform more effectively the task of the agreement, including the key alternative, the agreement on the encryption method used and the agreement to prevent the denial of the services. To manage and achieve these goals, this method proposes the existence of an control and monitoring entity to manage these agreements by collecting different statistical information of the opinions of the authorized parties in the cryptographic system. These statistics help this entity to take the proper decision about the agreement factors. This entity is called Agreement Manager (AM).

Keywords: agreement parameters, key agreement, key exchange, security management

Procedia PDF Downloads 421
16067 The Different Roles between Sodium and Potassium Ions in Ion Exchange of WO3/SiO2 Catalysts

Authors: Kritsada Pipitthapan

Abstract:

WO3/SiO2 catalysts were modified by an ion exchange method with sodium hydroxide or potassium hydroxide solution. The performance of the modified catalysts was tested in the metathesis of ethylene and trans-2-butene to propylene. During ion exchange, sodium and potassium ions played different roles. Sodium modified catalysts revealed constant trans-2-butene conversion and propylene selectivity when the concentrations of sodium in the solution were varied. In contrast, potassium modified catalysts showed reduction of the conversion and increase of the selectivity. From these results, potassium hydroxide may affect the transformation of tungsten oxide active species, resulting in the decrease in conversion whereas sodium hydroxide did not. Moreover, the modification of catalysts by this method improved the catalyst stability by lowering the amount of coke deposited on the catalyst surface.

Keywords: acid sites, alkali metal, isomerization, metathesis

Procedia PDF Downloads 251
16066 Application of Nonparametric Geographically Weighted Regression to Evaluate the Unemployment Rate in East Java

Authors: Sifriyani Sifriyani, I Nyoman Budiantara, Sri Haryatmi, Gunardi Gunardi

Abstract:

East Java Province has a first rank as a province that has the most counties and cities in Indonesia and has the largest population. In 2015, the population reached 38.847.561 million, this figure showed a very high population growth. High population growth is feared to lead to increase the levels of unemployment. In this study, the researchers mapped and modeled the unemployment rate with 6 variables that were supposed to influence. Modeling was done by nonparametric geographically weighted regression methods with truncated spline approach. This method was chosen because spline method is a flexible method, these models tend to look for its own estimation. In this modeling, there were point knots, the point that showed the changes of data. The selection of the optimum point knots was done by selecting the most minimun value of Generalized Cross Validation (GCV). Based on the research, 6 variables were declared to affect the level of unemployment in eastern Java. They were the percentage of population that is educated above high school, the rate of economic growth, the population density, the investment ratio of total labor force, the regional minimum wage and the ratio of the number of big industry and medium scale industry from the work force. The nonparametric geographically weighted regression models with truncated spline approach had a coefficient of determination 98.95% and the value of MSE equal to 0.0047.

Keywords: East Java, nonparametric geographically weighted regression, spatial, spline approach, unemployed rate

Procedia PDF Downloads 321
16065 Teachers and Innovations in Information and Communication Technology

Authors: Martina Manenova, Lukas Cirus

Abstract:

This article introduces research focused on elementary school teachers’ approach to innovations in ICT. The diffusion of innovations theory, which was written by E. M. Rogers, captures the processes of innovation adoption. The research method derived from this theory and the Rogers’ questionnaire focused on the diffusion of innovations was used as the basic research method. The research sample consisted of elementary school teachers. The comparison of results with the Rogers’ results shows that among the teachers in the research sample the so-called early majority, as well as the overall division of the data, was rather central (early adopter, early majority, and later majority). The teachers very rarely appeared on the edge positions (innovator, laggard). The obtained results can be applied to teaching practice and used especially in the implementation of new technologies and techniques into the educational process.

Keywords: innovation, diffusion of innovation, information and communication technology, teachers

Procedia PDF Downloads 293
16064 An Efficient Motion Recognition System Based on LMA Technique and a Discrete Hidden Markov Model

Authors: Insaf Ajili, Malik Mallem, Jean-Yves Didier

Abstract:

Human motion recognition has been extensively increased in recent years due to its importance in a wide range of applications, such as human-computer interaction, intelligent surveillance, augmented reality, content-based video compression and retrieval, etc. However, it is still regarded as a challenging task especially in realistic scenarios. It can be seen as a general machine learning problem which requires an effective human motion representation and an efficient learning method. In this work, we introduce a descriptor based on Laban Movement Analysis technique, a formal and universal language for human movement, to capture both quantitative and qualitative aspects of movement. We use Discrete Hidden Markov Model (DHMM) for training and classification motions. We improve the classification algorithm by proposing two DHMMs for each motion class to process the motion sequence in two different directions, forward and backward. Such modification allows avoiding the misclassification that can happen when recognizing similar motions. Two experiments are conducted. In the first one, we evaluate our method on a public dataset, the Microsoft Research Cambridge-12 Kinect gesture data set (MSRC-12) which is a widely used dataset for evaluating action/gesture recognition methods. In the second experiment, we build a dataset composed of 10 gestures(Introduce yourself, waving, Dance, move, turn left, turn right, stop, sit down, increase velocity, decrease velocity) performed by 20 persons. The evaluation of the system includes testing the efficiency of our descriptor vector based on LMA with basic DHMM method and comparing the recognition results of the modified DHMM with the original one. Experiment results demonstrate that our method outperforms most of existing methods that used the MSRC-12 dataset, and a near perfect classification rate in our dataset.

Keywords: human motion recognition, motion representation, Laban Movement Analysis, Discrete Hidden Markov Model

Procedia PDF Downloads 207
16063 Effect of Mesh Size on the Supersonic Viscous Flow Parameters around an Axisymmetric Blunt Body

Authors: Haoui Rabah

Abstract:

The aim of this work is to analyze a viscous flow around the axisymmetric blunt body taken into account the mesh size both in the free stream and into the boundary layer. The resolution of the Navier-Stokes equations is realized by using the finite volume method to determine the flow parameters and detached shock position. The numerical technique uses the Flux Vector Splitting method of Van Leer. Here, adequate time stepping parameter, CFL coefficient and mesh size level are selected to ensure numerical convergence. The effect of the mesh size is significant on the shear stress and velocity profile. The best solution is obtained with using a very fine grid. This study enabled us to confirm that the determination of boundary layer thickness can be obtained only if the size of the mesh is lower than a certain value limits given by our calculations.

Keywords: supersonic flow, viscous flow, finite volume, blunt body

Procedia PDF Downloads 604
16062 Temporal and Spacial Adaptation Strategies in Aerodynamic Simulation of Bluff Bodies Using Vortex Particle Methods

Authors: Dario Milani, Guido Morgenthal

Abstract:

Fluid dynamic computation of wind caused forces on bluff bodies e.g light flexible civil structures or high incidence of ground approaching airplane wings, is one of the major criteria governing their design. For such structures a significant dynamic response may result, requiring the usage of small scale devices as guide-vanes in bridge design to control these effects. The focus of this paper is on the numerical simulation of the bluff body problem involving multiscale phenomena induced by small scale devices. One of the solution methods for the CFD simulation that is relatively successful in this class of applications is the Vortex Particle Method (VPM). The method is based on a grid free Lagrangian formulation of the Navier-Stokes equations, where the velocity field is modeled by particles representing local vorticity. These vortices are being convected due to the free stream velocity as well as diffused. This representation yields the main advantages of low numerical diffusion, compact discretization as the vorticity is strongly localized, implicitly accounting for the free-space boundary conditions typical for this class of FSI problems, and a natural representation of the vortex creation process inherent in bluff body flows. When the particle resolution reaches the Kolmogorov dissipation length, the method becomes a Direct Numerical Simulation (DNS). However, it is crucial to note that any solution method aims at balancing the computational cost against the accuracy achievable. In the classical VPM method, if the fluid domain is discretized by Np particles, the computational cost is O(Np2). For the coupled FSI problem of interest, for example large structures such as long-span bridges, the aerodynamic behavior may be influenced or even dominated by small structural details such as barriers, handrails or fairings. For such geometrically complex and dimensionally large structures, resolving the complete domain with the conventional VPM particle discretization might become prohibitively expensive to compute even for moderate numbers of particles. It is possible to reduce this cost either by reducing the number of particles or by controlling its local distribution. It is also possible to increase the accuracy of the solution without increasing substantially the global computational cost by computing a correction of the particle-particle interaction in some regions of interest. In this paper different strategies are presented in order to extend the conventional VPM method to reduce the computational cost whilst resolving the required details of the flow. The methods include temporal sub stepping to increase the accuracy of the particles convection in certain regions as well as dynamically re-discretizing the particle map to locally control the global and the local amount of particles. Finally, these methods will be applied on a test case and the improvements in the efficiency as well as the accuracy of the proposed extension to the method are presented. The important benefits in terms of accuracy and computational cost of the combination of these methods will be thus presented as long as their relevant applications.

Keywords: adaptation, fluid dynamic, remeshing, substepping, vortex particle method

Procedia PDF Downloads 262
16061 Scheduling Tasks in Embedded Systems Based on NoC Architecture

Authors: D. Dorota

Abstract:

This paper presents a method to generate and schedule task in the architecture of embedded systems based on the simulated annealing. This method takes into account the attribute of divisibility of tasks. A proposal represents the process in the form of trees. Despite the fact that the architecture of Network-on-Chip (NoC) is an interesting alternative to a bus architecture based on multi-processors systems, it requires a lot of work that ensures the optimization of communication. This paper proposes an effective approach to generate dedicated NoC topology solving communication problems. Network NoC is generated taking into account the energy consumption and resource issues. Ultimately generated is minimal, dedicated NoC topology. The proposed solution is assumed to be a simple router design and the minimum number of lines.

Keywords: Network-on-Chip, NoC-based embedded systems, scheduling task in embedded systems, simulated annealing

Procedia PDF Downloads 377
16060 The Optimal Indirect Vector Controller Design via an Adaptive Tabu Search Algorithm

Authors: P. Sawatnatee, S. Udomsuk, K-N. Areerak, K-L. Areerak, A. Srikaew

Abstract:

The paper presents how to design the indirect vector control of three-phase induction motor drive systems using the artificial intelligence technique called the adaptive tabu search. The results from the simulation and the experiment show that the drive system with the controller designed from the proposed method can provide the best output speed response compared with those of the conventional method. The controller design using the proposed technique can be used to create the software package for engineers to achieve the optimal controller design of the induction motor speed control based on the indirect vector concept.

Keywords: indirect vector control, induction motor, adaptive tabu search, control design, artificial intelligence

Procedia PDF Downloads 398
16059 Multi-Criteria Based Robust Markowitz Model under Box Uncertainty

Authors: Pulak Swain, A. K. Ojha

Abstract:

Portfolio optimization is based on dealing with the problems of efficient asset allocation. Risk and Expected return are two conflicting criteria in such problems, where the investor prefers the return to be high and the risk to be low. Using multi-objective approach we can solve those type of problems. However the information which we have for the input parameters are generally ambiguous and the input values can fluctuate around some nominal values. We can not ignore the uncertainty in input values, as they can affect the asset allocation drastically. So we use Robust Optimization approach to the problems where the input parameters comes under box uncertainty. In this paper, we solve the multi criteria robust problem with the help of  E- constraint method.

Keywords: portfolio optimization, multi-objective optimization, ϵ - constraint method, box uncertainty, robust optimization

Procedia PDF Downloads 139
16058 Ghost Frequency Noise Reduction through Displacement Deviation Analysis

Authors: Paua Ketan, Bhagate Rajkumar, Adiga Ganesh, M. Kiran

Abstract:

Low gear noise is an important sound quality feature in modern passenger cars. Annoying gear noise from the gearbox is influenced by the gear design, gearbox shaft layout, manufacturing deviations in the components, assembly errors and the mounting arrangement of the complete gearbox. Geometrical deviations in the form of profile and lead errors are often present on the flanks of the inspected gears. Ghost frequencies of a gear are very challenging to identify in standard gear measurement and analysis process due to small wavelengths involved. In this paper, gear whine noise occurring at non-integral multiples of gear mesh frequency of passenger car gearbox is investigated and the root cause is identified using the displacement deviation analysis (DDA) method. DDA method is applied to identify ghost frequency excitations on the flanks of gears arising out of generation grinding. Frequency identified through DDA correlated with the frequency of vibration and noise on the end-of-line machine as well as vehicle level measurements. With the application of DDA method along with standard lead profile measurement, gears with ghost frequency geometry deviations were identified on the production line to eliminate defective parts and thereby eliminate ghost frequency noise from a vehicle. Further, displacement deviation analysis can be used in conjunction with the manufacturing process simulation to arrive at suitable countermeasures for arresting the ghost frequency.

Keywords: displacement deviation analysis, gear whine, ghost frequency, sound quality

Procedia PDF Downloads 146
16057 Investigation of Clustering Algorithms Used in Wireless Sensor Networks

Authors: Naim Karasekreter, Ugur Fidan, Fatih Basciftci

Abstract:

Wireless sensor networks are networks in which more than one sensor node is organized among themselves. The working principle is based on the transfer of the sensed data over the other nodes in the network to the central station. Wireless sensor networks concentrate on routing algorithms, energy efficiency and clustering algorithms. In the clustering method, the nodes in the network are divided into clusters using different parameters and the most suitable cluster head is selected from among them. The data to be sent to the center is sent per cluster, and the cluster head is transmitted to the center. With this method, the network traffic is reduced and the energy efficiency of the nodes is increased. In this study, clustering algorithms were examined in terms of clustering performances and cluster head selection characteristics to try to identify weak and strong sides. This work is supported by the Project 17.Kariyer.123 of Afyon Kocatepe University BAP Commission.

Keywords: wireless sensor networks (WSN), clustering algorithm, cluster head, clustering

Procedia PDF Downloads 513
16056 Modelling Spatial Dynamics of Terrorism

Authors: André Python

Abstract:

To this day, terrorism persists as a worldwide threat, exemplified by the recent deadly attacks in January 2015 in Paris and the ongoing massacres perpetrated by ISIS in Iraq and Syria. In response to this threat, states deploy various counterterrorism measures, the cost of which could be reduced through effective preventive measures. In order to increase the efficiency of preventive measures, policy-makers may benefit from accurate predictive models that are able to capture the complex spatial dynamics of terrorism occurring at a local scale. Despite empirical research carried out at country-level that has confirmed theories explaining the diffusion processes of terrorism across space and time, scholars have failed to assess diffusion’s theories on a local scale. Moreover, since scholars have not made the most of recent statistical modelling approaches, they have been unable to build up predictive models accurate in both space and time. In an effort to address these shortcomings, this research suggests a novel approach to systematically assess the theories of terrorism’s diffusion on a local scale and provide a predictive model of the local spatial dynamics of terrorism worldwide. With a focus on the lethal terrorist events that occurred after 9/11, this paper addresses the following question: why and how does lethal terrorism diffuse in space and time? Based on geolocalised data on worldwide terrorist attacks and covariates gathered from 2002 to 2013, a binomial spatio-temporal point process is used to model the probability of terrorist attacks on a sphere (the world), the surface of which is discretised in the form of Delaunay triangles and refined in areas of specific interest. Within a Bayesian framework, the model is fitted through an integrated nested Laplace approximation - a recent fitting approach that computes fast and accurate estimates of posterior marginals. Hence, for each location in the world, the model provides a probability of encountering a lethal terrorist attack and measures of volatility, which inform on the model’s predictability. Diffusion processes are visualised through interactive maps that highlight space-time variations in the probability and volatility of encountering a lethal attack from 2002 to 2013. Based on the previous twelve years of observation, the location and lethality of terrorist events in 2014 are statistically accurately predicted. Throughout the global scope of this research, local diffusion processes such as escalation and relocation are systematically examined: the former process describes an expansion from high concentration areas of lethal terrorist events (hotspots) to neighbouring areas, while the latter is characterised by changes in the location of hotspots. By controlling for the effect of geographical, economical and demographic variables, the results of the model suggest that the diffusion processes of lethal terrorism are jointly driven by contagious and non-contagious factors that operate on a local scale – as predicted by theories of diffusion. Moreover, by providing a quantitative measure of predictability, the model prevents policy-makers from making decisions based on highly uncertain predictions. Ultimately, this research may provide important complementary tools to enhance the efficiency of policies that aim to prevent and combat terrorism.

Keywords: diffusion process, terrorism, spatial dynamics, spatio-temporal modeling

Procedia PDF Downloads 351
16055 Assessment of Exhaust Emissions and Fuel Consumption from Means of Transport in Agriculture

Authors: Jerzy Merkisz, Piotr Lijewski, Pawel Fuc, Maciej Siedlecki, Andrzej Ziolkowski, Sylwester Weymann

Abstract:

The paper discusses the problem of load transport using farm tractors and road tractor units. This type of carriage of goods is often done with farm vehicles. The tests were performed with the PEMS equipment (Portable Emission Measurement System) under actual traffic conditions. The vehicles carried a load of 20000 kg. This research method is one of the most desired because it provides reliable information on the actual vehicle emissions and fuel consumption (carbon balance method). For the tests, a route was selected that simulated a trip from a small town to a food-processing facility located in a city. The analysis of the obtained results gave a clear answer as to what vehicles need to be used for the carriage of this type of cargo in terms of exhaust emissions and fuel consumption.

Keywords: emission, transport, fuel consumption, PEMS

Procedia PDF Downloads 530
16054 Estimation of Soil Erosion and Sediment Yield for ONG River Using GIS

Authors: Sanjay Kumar Behera, Kanhu Charan Patra

Abstract:

A GIS-based method has been applied for the determination of soil erosion and sediment yield in a small watershed in Ong River basin, Odisha, India. The method involves spatial disintegration of the catchment into homogenous grid cells to capture the catchment heterogeneity. The gross soil erosion in each cell was calculated using Universal Soil Loss Equation (USLE) by carefully determining its various parameters. The concept of sediment delivery ratio is used to route surface erosion from each of the discretized cells to the catchment outlet. The process of sediment delivery from grid cells to the catchment outlet is represented by the topographical characteristics of the cells. The effect of DEM resolution on sediment yield is analyzed using two different resolutions of DEM. The spatial discretization of the catchment and derivation of the physical parameters related to erosion in the cell are performed through GIS techniques.

Keywords: DEM, GIS, sediment delivery ratio, sediment yield, soil erosion

Procedia PDF Downloads 449
16053 Conservativeness of Probabilistic Constrained Optimal Control Method for Unknown Probability Distribution

Authors: Tomoaki Hashimoto

Abstract:

In recent decades, probabilistic constrained optimal control problems have attracted much attention in many research field. Although probabilistic constraints are generally intractable in an optimization problem, several tractable methods haven been proposed to handle probabilistic constraints. In most methods, probabilistic constraints are reduced to deterministic constraints that are tractable in an optimization problem. However, there is a gap between the transformed deterministic constraints in case of known and unknown probability distribution. This paper examines the conservativeness of probabilistic constrained optimization method with the unknown probability distribution. The objective of this paper is to provide a quantitative assessment of the conservatism for tractable constraints in probabilistic constrained optimization with the unknown probability distribution.

Keywords: optimal control, stochastic systems, discrete time systems, probabilistic constraints

Procedia PDF Downloads 581
16052 Comparison of Tribological Properties of TiO₂, ZrO₂ and TiO₂–ZrO₂ Composite Films Prepared by Sol–Gel Method

Authors: O. Çomaklı, M. Yazıcı, T. Yetim, A. F. Yetim, A. Çelik

Abstract:

In this study, TiO₂, ZrO₂, and TiO₂–ZrO₂ composite films were coated on Cp-Ti substrates by sol-gel method. Structures of uncoated and coated samples were investigated by X-ray diffraction and SEM. XRD data identified anatase phase in TiO₂ coated samples and tetragonal zirconia phase in ZrO₂ coated samples while both of anatase and tetragonal zirconia phases in TiO₂–ZrO₂ composite films. The mechanical and wear properties of samples were investigated using micro hardness, pin-on-disk tribotester, and 3D profilometer. The best wear resistance was obtained from TiO₂–ZrO₂ composite films. This can be attributed to their high surface hardness, low surface roughness and high thickness of the film.

Keywords: sol-gel, TiO₂, ZrO₂, TiO₂–ZrO₂, composite films, wear

Procedia PDF Downloads 263
16051 Vibration Frequencies Analysis of Nanoporous Graphene Membrane

Authors: Haw-Long Lee, Win-Jin Chang, Yu-Ching Yang

Abstract:

In this study, we use the atomic-scale finite element method to investigate the vibrational behavior of the armchair- and zigzag-structured nanoporous graphene layers with different size under the SFSF and CFFF boundary conditions. The fundamental frequencies computed for the graphene layers without pore are compared with the results of previous studies. We observe very good correspondence of our results with that of the other studies in all the considered cases. For the armchair- and zigzag-structured nanoporous graphene layers under the SFSF and CFFF boundary conditions, the frequencies decrease as the size of the nanopore increase. When the positions of the pore are symmetric with respect to the center of the graphene, the frequency of the zigzag pore graphene is higher than that of the armchair one.

Keywords: atomic-scale finite element method, graphene, nanoporous, natural frequency

Procedia PDF Downloads 361
16050 Optimization of Fused Deposition Modeling 3D Printing Process via Preprocess Calibration Routine Using Low-Cost Thermal Sensing

Authors: Raz Flieshman, Adam Michael Altenbuchner, Jörg Krüger

Abstract:

This paper presents an approach to optimizing the Fused Deposition Modeling (FDM) 3D printing process through a preprocess calibration routine of printing parameters. The core of this method involves the use of a low-cost thermal sensor capable of measuring tempera-tures within the range of -20 to 500 degrees Celsius for detailed process observation. The calibration process is conducted by printing a predetermined path while varying the process parameters through machine instructions (g-code). This enables the extraction of critical thermal, dimensional, and surface properties along the printed path. The calibration routine utilizes computer vision models to extract features and metrics from the thermal images, in-cluding temperature distribution, layer adhesion quality, surface roughness, and dimension-al accuracy and consistency. These extracted properties are then analyzed to optimize the process parameters to achieve the desired qualities of the printed material. A significant benefit of this calibration method is its potential to create printing parameter profiles for new polymer and composite materials, thereby enhancing the versatility and application range of FDM 3D printing. The proposed method demonstrates significant potential in enhancing the precision and reliability of FDM 3D printing, making it a valuable contribution to the field of additive manufacturing.

Keywords: FDM 3D printing, preprocess calibration, thermal sensor, process optimization, additive manufacturing, computer vision, material profiles

Procedia PDF Downloads 41
16049 A Semiparametric Approach to Estimate the Mode of Continuous Multivariate Data

Authors: Tiee-Jian Wu, Chih-Yuan Hsu

Abstract:

Mode estimation is an important task, because it has applications to data from a wide variety of sources. We propose a semi-parametric approach to estimate the mode of an unknown continuous multivariate density function. Our approach is based on a weighted average of a parametric density estimate using the Box-Cox transform and a non-parametric kernel density estimate. Our semi-parametric mode estimate improves both the parametric- and non-parametric- mode estimates. Specifically, our mode estimate solves the non-consistency problem of parametric mode estimates (at large sample sizes) and reduces the variability of non-parametric mode estimates (at small sample sizes). The performance of our method at practical sample sizes is demonstrated by simulation examples and two real examples from the fields of climatology and image recognition.

Keywords: Box-Cox transform, density estimation, mode seeking, semiparametric method

Procedia PDF Downloads 285
16048 CsPbBr₃@MOF-5-Based Single Drop Microextraction for in-situ Fluorescence Colorimetric Detection of Dechlorination Reaction

Authors: Yanxue Shang, Jingbin Zeng

Abstract:

Chlorobenzene homologues (CBHs) are a category of environmental pollutants that can not be ignored. They can stay in the environment for a long period and are potentially carcinogenic. The traditional degradation method of CBHs is dechlorination followed by sample preparation and analysis. This is not only time-consuming and laborious, but the detection and analysis processes are used in conjunction with large-scale instruments. Therefore, this can not achieve rapid and low-cost detection. Compared with traditional sensing methods, colorimetric sensing is simpler and more convenient. In recent years, chromaticity sensors based on fluorescence have attracted more and more attention. Compared with sensing methods based on changes in fluorescence intensity, changes in color gradients are easier to recognize by the naked eye. Accordingly, this work proposes to use single drop microextraction (SDME) technology to solve the above problems. After the dechlorination reaction was completed, the organic droplet extracts Cl⁻ and realizes fluorescence colorimetric sensing at the same time. This method was integrated sample processing and visual in-situ detection, simplifying the detection process. As a fluorescence colorimetric sensor material, CsPbBr₃ was encapsulated in MOF-5 to construct CsPbBr₃@MOF-5 fluorescence colorimetric composite. Then the fluorescence colorimetric sensor was constructed by dispersing the composite in SDME organic droplets. When the Br⁻ in CsPbBr₃ exchanges with Cl⁻ produced by the dechlorination reactions, it is converted into CsPbCl₃. The fluorescence color of the single droplet of SDME will change from green to blue emission, thereby realizing visual observation. Therein, SDME can enhance the concentration and enrichment of Cl⁻ and instead of sample pretreatment. The fluorescence color change of CsPbBr₃@MOF-5 can replace the detection process of large-scale instruments to achieve real-time rapid detection. Due to the absorption ability of MOF-5, it can not only improve the stability of CsPbBr₃, but induce the adsorption of Cl⁻. Simultaneously, accelerate the exchange of Br- and Cl⁻ in CsPbBr₃ and the detection process of Cl⁻. The absorption process was verified by density functional theory (DFT) calculations. This method exhibits exceptional linearity for Cl⁻ in the range of 10⁻² - 10⁻⁶ M (10000 μM - 1 μM) with a limit of detection of 10⁻⁷ M. Whereafter, the dechlorination reactions of different kinds of CBHs were also carried out with this method, and all had satisfactory detection ability. Also verified the accuracy by gas chromatography (GC), and it was found that the SDME we developed in this work had high credibility. In summary, the in-situ visualization method of dechlorination reaction detection was a combination of sample processing and fluorescence colorimetric sensing. Thus, the strategy researched herein represents a promising method for the visual detection of dechlorination reactions and can be extended for applications in environments, chemical industries, and foods.

Keywords: chlorobenzene homologues, colorimetric sensor, metal halide perovskite, metal-organic frameworks, single drop microextraction

Procedia PDF Downloads 143
16047 Modelling and Simulation of Aero-Elastic Vibrations Using System Dynamic Approach

Authors: Cosmas Pandit Pagwiwoko, Ammar Khaled Abdelaziz Abdelsamia

Abstract:

Flutter as a phenomenon of flow-induced and self-excited vibration has to be recognized considering its harmful effect on the structure especially in a stage of aircraft design. This phenomenon is also important for a wind energy harvester based on the fluttering surface due to its effective operational velocity range. This multi-physics occurrence can be presented by two governing equations in both fluid and structure simultaneously in respecting certain boundary conditions on the surface of the body. In this work, the equations are resolved separately by two distinct solvers, one-time step of each domain. The modelling and simulation of this flow-structure interaction in ANSYS show the effectiveness of this loosely coupled method in representing flutter phenomenon however the process is time-consuming for design purposes. Therefore, another technique using the same weak coupled aero-structure is proposed by using system dynamics approach. In this technique, the aerodynamic forces were calculated using singularity function for a range of frequencies and certain natural mode shapes are transformed into time domain by employing an approximation model of fraction rational function in Laplace variable. The representation of structure in a multi-degree-of-freedom coupled with a transfer function of aerodynamic forces can then be simulated in time domain on a block-diagram platform such as Simulink MATLAB. The dynamic response of flutter at certain velocity can be evaluated with another established flutter calculation in frequency domain k-method. In this method, a parameter of artificial structural damping is inserted in the equation of motion to assure the energy balance of flow and vibrating structure. The simulation in time domain is particularly interested as it enables to apply the structural non-linear factors accurately. Experimental tests on a fluttering airfoil in the wind tunnel are also conducted to validate the method.

Keywords: flutter, flow-induced vibration, flow-structure interaction, non-linear structure

Procedia PDF Downloads 315