Search results for: error minimization
1481 Modeling Binomial Dependent Distribution of the Values: Synthesis Tables of Probabilities of Errors of the First and Second Kind of Biometrics-Neural Network Authentication System
Authors: B. S.Akhmetov, S. T. Akhmetova, D. N. Nadeyev, V. Yu. Yegorov, V. V. Smogoonov
Abstract:
Estimated probabilities of errors of the first and second kind for nonideal biometrics-neural transducers 256 outputs, the construction of nomograms based error probability of 'own' and 'alien' from the mathematical expectation and standard deviation of the normalized measures Hamming.Keywords: modeling, errors, probability, biometrics, neural network, authentication
Procedia PDF Downloads 4831480 Validity of a Timing System in the Alpine Ski Field: A Magnet-Based Timing System Using the Magnetometer Built into an Inertial Measurement Units
Authors: Carla Pérez-Chirinos Buxadé, Bruno Fernández-Valdés, Mónica Morral-Yepes, Sílvia Tuyà Viñas, Josep Maria Padullés Riu, Gerard Moras Feliu
Abstract:
There is a long way to explore all the possible applications inertial measurement units (IMUs) have in the sports field. The aim of this study was to evaluate the validity of a new application on the use of these wearable sensors, specifically it was to evaluate a magnet-based timing system (M-BTS) for timing gate-to-gate in an alpine ski slalom using the magnetometer embedded in an IMU. This was a validation study. The criterion validity of time measured by the M-BTS was assessed using the 95% error range against actual time obtained from photocells. The experiment was carried out with first-and second-year junior skiers performing a ski slalom on a ski training slope. Eight alpine skiers (17.4 ± 0.8 years, 176.4 ± 4.9 cm, 67.7 ± 2.0 kg, 128.8 ± 26.6 slalom FIS-Points) participated in the study. An IMU device was attached to the skier’s lower back. Skiers performed a 40-gate slalom from which four gates were assessed. The M-BTS consisted of placing four bar magnets buried into the snow surface on the inner side of each gate’s turning pole; the magnetometer built into the IMU detected the peak-shaped magnetic field when passing near the magnets at a certain speed. Four magnetic peaks were detected. The time compressed between peaks was calculated. Three inter-gate times were obtained for each system: photocells and M-BTS. The total time was defined as the time sum of the inter-gate times. The 95% error interval for the total time was 0.050 s for the ski slalom. The M-BTS is valid for timing gate-to-gate in an alpine ski slalom. Inter-gate times can provide additional data for analyzing a skier’s performance, such as asymmetries between left and right foot.Keywords: gate crossing time, inertial measurement unit, timing system, wearable sensor
Procedia PDF Downloads 1851479 Assessing Environmental Urban Sustainability Using Multivariate Analysis: A Case of Nagpur, India
Authors: Anusha Vaddiraj Pallapu
Abstract:
Measuring urban sustainable development is at the forefront in contributing to overall sustainability, and it refers to attaining social equity, environmental protection and minimizing the impacts of urbanization. Assessing performance of urban issues ranging from larger consumption of natural resources by humans in terms of lifestyle to creating a polluted nearby environment, social and even economic dimensions of sustainability major issues observed such as water quality, transportation, management of solid waste and traffic pollution. However, relying on the framework of the project to do the goals of sustainable development or minimization of urban impacts through management practices is not enough to deal with the present urban issues. The aim of the sustainability is to know how severely the resources are depleted because of human consumption and how issues are characterized. The paper aims to assign benchmarks for the selected sustainability indicators for research, and analysis is done through multivariate analysis in Indian context a case of Nagpur city to identify the play role of each urban issues in the overall sustainability. The main objectives of this paper are to examine the indicators over by time basis on various scenarios and how benchmarking is used, what and which categories of values should be considered as the performance of indicators function.Keywords: environmental sustainability indicators, principal component analysis, urban sustainability, urban clusters, benchmarking
Procedia PDF Downloads 3451478 Integration of Load Introduction Elements into Fabrics
Authors: Jan Schwennen, Harlad Schmid, Juergen Fleischer
Abstract:
Lightweight design plays an important role in the automotive industry. Especially the combination of metal and CFRP shows great potential for future vehicle concepts. This requires joining technologies that are cost-efficient and appropriate for the materials involved. Previous investigations show that integrating load introduction elements during CFRP part manufacturing offers great advantages in mechanical performance. However, it is not yet clear how to integrate the elements in an automated process without harming the fiber structure. In this paper, a test rig is build up to investigate the effect of different parameters during insert integration experimentally. After a short description of the experimental equipment, preliminary tests are performed to determine a set of important process parameters. Based on that, the planning of design of experiments is given. The interpretation and evaluation of the test results show that with a minimization of the insert diameter and the peak angle less harm on the fiber structure can be achieved. Furthermore, a maximization of the die diameter above the insert shows a positive effect on the fiber structure. At the end of this paper, a theoretical description of alternative peak shaping is given and then the results get validated on the basis of an industrial reference part.Keywords: CFRP, fabrics, insert, load introduction element, integration
Procedia PDF Downloads 2441477 Orbit Determination from Two Position Vectors Using Finite Difference Method
Authors: Akhilesh Kumar, Sathyanarayan G., Nirmala S.
Abstract:
An unusual approach is developed to determine the orbit of satellites/space objects. The determination of orbits is considered a boundary value problem and has been solved using the finite difference method (FDM). Only positions of the satellites/space objects are known at two end times taken as boundary conditions. The technique of finite difference has been used to calculate the orbit between end times. In this approach, the governing equation is defined as the satellite's equation of motion with a perturbed acceleration. Using the finite difference method, the governing equations and boundary conditions are discretized. The resulting system of algebraic equations is solved using Tri Diagonal Matrix Algorithm (TDMA) until convergence is achieved. This methodology test and evaluation has been done using all GPS satellite orbits from National Geospatial-Intelligence Agency (NGA) precise product for Doy 125, 2023. Towards this, two hours of twelve sets have been taken into consideration. Only positions at the end times of each twelve sets are considered boundary conditions. This algorithm is applied to all GPS satellites. Results achieved using FDM compared with the results of NGA precise orbits. The maximum RSS error for the position is 0.48 [m] and the velocity is 0.43 [mm/sec]. Also, the present algorithm is applied on the IRNSS satellites for Doy 220, 2023. The maximum RSS error for the position is 0.49 [m], and for velocity is 0.28 [mm/sec]. Next, a simulation has been done for a Highly Elliptical orbit for DOY 63, 2023, for the duration of 6 hours. The RSS of difference in position is 0.92 [m] and velocity is 1.58 [mm/sec] for the orbital speed of more than 5km/sec. Whereas the RSS of difference in position is 0.13 [m] and velocity is 0.12 [mm/sec] for the orbital speed less than 5km/sec. Results show that the newly created method is reliable and accurate. Further applications of the developed methodology include missile and spacecraft targeting, orbit design (mission planning), space rendezvous and interception, space debris correlation, and navigation solutions.Keywords: finite difference method, grid generation, NavIC system, orbit perturbation
Procedia PDF Downloads 861476 Information Theoretic Approach for Beamforming in Wireless Communications
Authors: Syed Khurram Mahmud, Athar Naveed, Shoaib Arif
Abstract:
Beamforming is a signal processing technique extensively utilized in wireless communications and radars for desired signal intensification and interference signal minimization through spatial selectivity. In this paper, we present a method for calculation of optimal weight vectors for smart antenna array, to achieve a directive pattern during transmission and selective reception in interference prone environment. In proposed scheme, Mutual Information (MI) extrema are evaluated through an energy constrained objective function, which is based on a-priori information of interference source and desired array factor. Signal to Interference plus Noise Ratio (SINR) performance is evaluated for both transmission and reception. In our scheme, MI is presented as an index to identify trade-off between information gain, SINR, illumination time and spatial selectivity in an energy constrained optimization problem. The employed method yields lesser computational complexity, which is presented through comparative analysis with conventional methods in vogue. MI based beamforming offers enhancement of signal integrity in degraded environment while reducing computational intricacy and correlating key performance indicators.Keywords: beamforming, interference, mutual information, wireless communications
Procedia PDF Downloads 2821475 Reducing Support Structures in Design for Additive Manufacturing: A Neural Networks Approach
Authors: Olivia Borgue, Massimo Panarotto, Ola Isaksson
Abstract:
This article presents a neural networks-based strategy for reducing the need for support structures when designing for additive manufacturing (AM). Additive manufacturing is a relatively new and immature industrial technology, and the information to make confident decisions when designing for AM is limited. This lack of information impacts especially the early stages of engineering design, for instance, it is difficult to actively consider the support structures needed for manufacturing a part. This difficulty is related to the challenge of designing a product geometry accounting for customer requirements, manufacturing constraints and minimization of support structure. The approach presented in this article proposes an automatized geometry modification technique for reducing the use of the support structures while designing for AM. This strategy starts with a neural network-based strategy for shape recognition to achieve product classification, using an STL file of the product as input. Based on the classification, an automatic part geometry modification based on MATLAB© is implemented. At the end of the process, the strategy presents different geometry modification alternatives depending on the type of product to be designed. The geometry alternatives are then evaluated adopting a QFD-like decision support tool.Keywords: additive manufacturing, engineering design, geometry modification optimization, neural networks
Procedia PDF Downloads 2551474 Forecasting Optimal Production Program Using Profitability Optimization by Genetic Algorithm and Neural Network
Authors: Galal H. Senussi, Muamar Benisa, Sanja Vasin
Abstract:
In our business field today, one of the most important issues for any enterprises is cost minimization and profit maximization. Second issue is how to develop a strong and capable model that is able to give us desired forecasting of these two issues. Many researches deal with these issues using different methods. In this study, we developed a model for multi-criteria production program optimization, integrated with Artificial Neural Network. The prediction of the production cost and profit per unit of a product, dealing with two obverse functions at same time can be extremely difficult, especially if there is a great amount of conflict information about production parameters. Feed-Forward Neural Networks are suitable for generalization, which means that the network will generate a proper output as a result to input it has never seen. Therefore, with small set of examples the network will adjust its weight coefficients so the input will generate a proper output. This essential characteristic is of the most important abilities enabling this network to be used in variety of problems spreading from engineering to finance etc. From our results as we will see later, Feed-Forward Neural Networks has a strong ability and capability to map inputs into desired outputs.Keywords: project profitability, multi-objective optimization, genetic algorithm, Pareto set, neural networks
Procedia PDF Downloads 4461473 Design and Simulation of All Optical Fiber to the Home Network
Authors: Rahul Malhotra
Abstract:
Fiber based access networks can deliver performance that can support the increasing demands for high speed connections. One of the new technologies that have emerged in recent years is Passive Optical Networks. This paper is targeted to show the simultaneous delivery of triple play service (data, voice and video). The comparative investigation and suitability of various data rates is presented. It is demonstrated that as we increase the data rate, number of users to be accommodated decreases due to increase in bit error rate.Keywords: BER, PON, TDMPON, GPON, CWDM, OLT, ONT
Procedia PDF Downloads 5581472 The Design of Multiple Detection Parallel Combined Spread Spectrum Communication System
Authors: Lixin Tian, Wei Xue
Abstract:
Many jobs in society go underground, such as mine mining, tunnel construction and subways, which are vital to the development of society. Once accidents occur in these places, the interruption of traditional wired communication is not conducive to the development of rescue work. In order to realize the positioning, early warning and command functions of underground personnel and improve rescue efficiency, it is necessary to develop and design an emergency ground communication system. It is easy to be subjected to narrowband interference when performing conventional underground communication. Spreading communication can be used for this problem. However, general spread spectrum methods such as direct spread communication are inefficient, so it is proposed to use parallel combined spread spectrum (PCSS) communication to improve efficiency. The PCSS communication not only has the anti-interference ability and the good concealment of the traditional spread spectrum system, but also has a relatively high frequency band utilization rate and a strong information transmission capability. So, this technology has been widely used in practice. This paper presents a PCSS communication model-multiple detection parallel combined spread spectrum (MDPCSS) communication system. In this paper, the principle of MDPCSS communication system is described, that is, the sequence at the transmitting end is processed in blocks and cyclically shifted to facilitate multiple detection at the receiving end. The block diagrams of the transmitter and receiver of the MDPCSS communication system are introduced. At the same time, the calculation formula of the system bit error rate (BER) is introduced, and the simulation and analysis of the BER of the system are completed. By comparing with the common parallel PCSS communication, we can draw a conclusion that it is indeed possible to reduce the BER and improve the system performance. Furthermore, the influence of different pseudo-code lengths selected on the system BER is simulated and analyzed, and the conclusion is that the larger the pseudo-code length is, the smaller the system error rate is.Keywords: cyclic shift, multiple detection, parallel combined spread spectrum, PN code
Procedia PDF Downloads 1401471 Wasting Human and Computer Resources
Authors: Mária Csernoch, Piroska Biró
Abstract:
The legends about “user-friendly” and “easy-to-use” birotical tools (computer-related office tools) have been spreading and misleading end-users. This approach has led us to the extremely high number of incorrect documents, causing serious financial losses in the creating, modifying, and retrieving processes. Our research proved that there are at least two sources of this underachievement: (1) The lack of the definition of the correctly edited, formatted documents. Consequently, end-users do not know whether their methods and results are correct or not. They are not aware of their ignorance. They are so ignorant that their ignorance does not allow them to realize their lack of knowledge. (2) The end-users’ problem-solving methods. We have found that in non-traditional programming environments end-users apply, almost exclusively, surface approach metacognitive methods to carry out their computer related activities, which are proved less effective than deep approach methods. Based on these findings we have developed deep approach methods which are based on and adapted from traditional programming languages. In this study, we focus on the most popular type of birotical documents, the text-based documents. We have provided the definition of the correctly edited text, and based on this definition, adapted the debugging method known in programming. According to the method, before the realization of text editing, a thorough debugging of already existing texts and the categorization of errors are carried out. With this method in advance to real text editing users learn the requirements of text-based documents and also of the correctly formatted text. The method has been proved much more effective than the previously applied surface approach methods. The advantages of the method are that the real text handling requires much less human and computer sources than clicking aimlessly in the GUI (Graphical User Interface), and the data retrieval is much more effective than from error-prone documents.Keywords: deep approach metacognitive methods, error-prone birotical documents, financial losses, human and computer resources
Procedia PDF Downloads 3831470 The Effectiveness of Sulfate Reducing Bacteria in Minimizing Methane and Sludge Production from Palm Oil Mill Effluent (POME)
Authors: K. Abdul Halim, E. L. Yong
Abstract:
Palm oil industry is a major revenue earner in Malaysia, despite the growth of the industry is synonymous with a massive production of agro-industrial wastewater. Through the oil extraction processes, palm oil mill effluent (POME) contributes to the largest liquid wastes generated. Due to the high amount of organic compound, POME can cause inland water pollution if discharged untreated into the water course as well as affect the aquatic ecosystem. For more than 20 years, Malaysia adopted the conventional biological treatment known as lagoon system that apply biological treatment. Besides having difficulties in complying with the standard, a large build up area is needed and retention time is higher. Although anaerobic digester is more favorable, this process comes along with enormous volumes of sludge and methane gas, demanding attention from the mill operators. In order to reduce the sludge production, denitrifiers are to be removed first. Sulfate reducing bacteria has shown the capability to inhibit the growth of methanogens. This is expected to substantially reduce both the sludge and methane production in anaerobic digesters. In this paper, the effectiveness of sulfate reducing bacteria in minimizing sludge and methane will be examined.Keywords: methane reduction, palm oil mill effluent, sludge minimization, sulfate reducing bacteria, sulfate reduction
Procedia PDF Downloads 4331469 Fixed Points of Contractive-Like Operators by a Faster Iterative Process
Authors: Safeer Hussain Khan
Abstract:
In this paper, we prove a strong convergence result using a recently introduced iterative process with contractive-like operators. This improves and generalizes corresponding results in the literature in two ways: the iterative process is faster, operators are more general. In the end, we indicate that the results can also be proved with the iterative process with error terms.Keywords: contractive-like operator, iterative process, fixed point, strong convergence
Procedia PDF Downloads 4351468 Design of Multi-Loop Controller for Minimization of Energy Consumption in the Distillation Column
Authors: Vinayambika S. Bhat, S. Shanmuga Priya, I. Thirunavukkarasu, Shreeranga Bhat
Abstract:
An attempt has been made to design a decoupling controller for systems with more inputs more outputs with dead time in it. The de-coupler is designed for the chemical process industry 3×3 plant transfer function with dead time. The Quantitative Feedback Theory (QFT) based controller has also been designed here for the 2×2 distillation column transfer function. The developed control techniques were simulated using the MATLAB/Simulink. Also, the stability of the process was analyzed, together with the presence of various perturbations in it. Time domain specifications like setting time along with overshoot and oscillations were analyzed to prove the efficiency of the de-coupler method. The load disturbance rejection was tested along with its performance. The QFT control technique was synthesized based on the stability and performance specifications in the presence of uncertainty in time constant of the plant transfer function through sequential loop shaping technique. Further, the energy efficiency of the distillation column was improved by proper tuning of the controller. A distillation column consumes 3% of the total energy consumption of the world. A suitable control technique is very important from an economic point of view. The real time implementation of the process is under process in our laboratory.Keywords: distillation, energy, MIMO process, time delay, robust stability
Procedia PDF Downloads 4151467 Second Order Optimality Conditions in Nonsmooth Analysis on Riemannian Manifolds
Authors: Seyedehsomayeh Hosseini
Abstract:
Much attention has been paid over centuries to understanding and solving the problem of minimization of functions. Compared to linear programming and nonlinear unconstrained optimization problems, nonlinear constrained optimization problems are much more difficult. Since the procedure of finding an optimizer is a search based on the local information of the constraints and the objective function, it is very important to develop techniques using geometric properties of the constraints and the objective function. In fact, differential geometry provides a powerful tool to characterize and analyze these geometric properties. Thus, there is clearly a link between the techniques of optimization on manifolds and standard constrained optimization approaches. Furthermore, there are manifolds that are not defined as constrained sets in R^n an important example is the Grassmann manifolds. Hence, to solve optimization problems on these spaces, intrinsic methods are used. In a nondifferentiable problem, the gradient information of the objective function generally cannot be used to determine the direction in which the function is decreasing. Therefore, techniques of nonsmooth analysis are needed to deal with such a problem. As a manifold, in general, does not have a linear structure, the usual techniques, which are often used in nonsmooth analysis on linear spaces, cannot be applied and new techniques need to be developed. This paper presents necessary and sufficient conditions for a strict local minimum of extended real-valued, nonsmooth functions defined on Riemannian manifolds.Keywords: Riemannian manifolds, nonsmooth optimization, lower semicontinuous functions, subdifferential
Procedia PDF Downloads 3621466 Spectrophotometric Determination of Phenylephrine Hydrochloride by Coupling with Diazotized 2,4-Dinitroaniline
Authors: Sulaiman Gafar Muhamad
Abstract:
A rapid spectrophotometric method for the micro-determination of phenylephrine-HCl (PHE) has been developed. The proposed method involves the coupling of phenylephrine-HCl with diazotized 2,4-dinitroaniline in alkaline medium at λmax 455 nm. Under the present optimum condition, Beer’s law was obeyed in the range of 1.0-20 μg/ml of PHE with molar absorptivity of 1.915 ×104 l. mol-1.cm-1, with a relative error of 0.015 and a relative standard deviation of 0.024%. The current method has been applied successfully to estimate phenylephrine-HCl in pharmaceutical preparations (nose drop and syrup).Keywords: diazo-coupling, 2, 4-dinitroaniline, phenylephrine-HCl, spectrophotometry
Procedia PDF Downloads 2591465 Optimization of Cutting Parameters on Delamination Using Taguchi Method during Drilling of GFRP Composites
Authors: Vimanyu Chadha, Ranganath M. Singari
Abstract:
Drilling composite materials is a frequently practiced machining process during assembling in various industries such as automotive and aerospace. However, drilling of glass fiber reinforced plastic (GFRP) composites is significantly affected by damage tendency of these materials under cutting forces such as thrust force and torque. The aim of this paper is to investigate the influence of the various cutting parameters such as cutting speed and feed rate; subsequently also to study the influence of number of layers on delamination produced while drilling a GFRP composite. A plan of experiments, based on Taguchi techniques, was instituted considering drilling with prefixed cutting parameters in a hand lay-up GFRP material. The damage induced associated with drilling GFRP composites were measured. Moreover, Analysis of Variance (ANOVA) was performed to obtain minimization of delamination influenced by drilling parameters and number layers. The optimum drilling factor combination was obtained by using the analysis of signal-to-noise ratio. The conclusion revealed that feed rate was the most influential factor on the delamination. The best results of the delamination were obtained with composites with a greater number of layers at lower cutting speeds and feed rates.Keywords: analysis of variance, delamination, design optimization, drilling, glass fiber reinforced plastic composites, Taguchi method
Procedia PDF Downloads 2601464 Signal Integrity Performance Analysis in Capacitive and Inductively Coupled Very Large Scale Integration Interconnect Models
Authors: Mudavath Raju, Bhaskar Gugulothu, B. Rajendra Naik
Abstract:
The rapid advances in Very Large Scale Integration (VLSI) technology has resulted in the reduction of minimum feature size to sub-quarter microns and switching time in tens of picoseconds or even less. As a result, the degradation of high-speed digital circuits due to signal integrity issues such as coupling effects, clock feedthrough, crosstalk noise and delay uncertainty noise. Crosstalk noise in VLSI interconnects is a major concern and reduction in VLSI interconnect has become more important for high-speed digital circuits. It is the most effectively considered in Deep Sub Micron (DSM) and Ultra Deep Sub Micron (UDSM) technology. Increasing spacing in-between aggressor and victim line is one of the technique to reduce the crosstalk. Guard trace or shield insertion in-between aggressor and victim is also one of the prominent options for the minimization of crosstalk. In this paper, far end crosstalk noise is estimated with mutual inductance and capacitance RLC interconnect model. Also investigated the extent of crosstalk in capacitive and inductively coupled interconnects to minimizes the same through shield insertion technique.Keywords: VLSI, interconnects, signal integrity, crosstalk, shield insertion, guard trace, deep sub micron
Procedia PDF Downloads 1881463 Modelling Causal Effects from Complex Longitudinal Data via Point Effects of Treatments
Authors: Xiaoqin Wang, Li Yin
Abstract:
Background and purpose: In many practices, one estimates causal effects arising from a complex stochastic process, where a sequence of treatments are assigned to influence a certain outcome of interest, and there exist time-dependent covariates between treatments. When covariates are plentiful and/or continuous, statistical modeling is needed to reduce the huge dimensionality of the problem and allow for the estimation of causal effects. Recently, Wang and Yin (Annals of statistics, 2020) derived a new general formula, which expresses these causal effects in terms of the point effects of treatments in single-point causal inference. As a result, it is possible to conduct the modeling via point effects. The purpose of the work is to study the modeling of these causal effects via point effects. Challenges and solutions: The time-dependent covariates often have influences from earlier treatments as well as on subsequent treatments. Consequently, the standard parameters – i.e., the mean of the outcome given all treatments and covariates-- are essentially all different (null paradox). Furthermore, the dimension of the parameters is huge (curse of dimensionality). Therefore, it can be difficult to conduct the modeling in terms of standard parameters. Instead of standard parameters, we have use point effects of treatments to develop likelihood-based parametric approach to the modeling of these causal effects and are able to model the causal effects of a sequence of treatments by modeling a small number of point effects of individual treatment Achievements: We are able to conduct the modeling of the causal effects from a sequence of treatments in the familiar framework of single-point causal inference. The simulation shows that our method achieves not only an unbiased estimate for the causal effect but also the nominal level of type I error and a low level of type II error for the hypothesis testing. We have applied this method to a longitudinal study of COVID-19 mortality among Scandinavian countries and found that the Swedish approach performed far worse than the other countries' approach for COVID-19 mortality and the poor performance was largely due to its early measure during the initial period of the pandemic.Keywords: causal effect, point effect, statistical modelling, sequential causal inference
Procedia PDF Downloads 2071462 Optrix: Energy Aware Cross Layer Routing Using Convex Optimization in Wireless Sensor Networks
Authors: Ali Shareef, Aliha Shareef, Yifeng Zhu
Abstract:
Energy minimization is of great importance in wireless sensor networks in extending the battery lifetime. One of the key activities of nodes in a WSN is communication and the routing of their data to a centralized base-station or sink. Routing using the shortest path to the sink is not the best solution since it will cause nodes along this path to fail prematurely. We propose a cross-layer energy efficient routing protocol Optrix that utilizes a convex formulation to maximize the lifetime of the network as a whole. We further propose, Optrix-BW, a novel convex formulation with bandwidth constraint that allows the channel conditions to be accounted for in routing. By considering this key channel parameter we demonstrate that Optrix-BW is capable of congestion control. Optrix is implemented in TinyOS, and we demonstrate that a relatively large topology of 40 nodes can converge to within 91% of the optimal routing solution. We describe the pitfalls and issues related with utilizing a continuous form technique such as convex optimization with discrete packet based communication systems as found in WSNs. We propose a routing controller mechanism that allows for this transformation. We compare Optrix against the Collection Tree Protocol (CTP) and we found that Optrix performs better in terms of convergence to an optimal routing solution, for load balancing and network lifetime maximization than CTP.Keywords: wireless sensor network, Energy Efficient Routing
Procedia PDF Downloads 3931461 Security System for Safe Transmission of Medical Image
Authors: Mohammed Jamal Al-Mansor, Kok Beng Gan
Abstract:
This paper develops an optimized embedding of payload in medical image by using genetic optimization. The goal is to preserve region of interest from being distorted because of the watermark. By using this developed system there is no need of manual defining of region of interest through experts as the system will apply the genetic optimization to select the parts of image that can carry the watermark with guaranteeing less distortion. The experimental results assure that genetic based optimization is useful for performing steganography with less mean square error percentage.Keywords: AES, DWT, genetic algorithm, watermarking
Procedia PDF Downloads 4121460 Discontinuous Galerkin Method for Higher-Order Ordinary Differential Equations
Authors: Helmi Temimi
Abstract:
In this paper, we study the super-convergence properties of the discontinuous Galerkin (DG) method applied to one-dimensional mth-order ordinary differential equations without introducing auxiliary variables. We found that nth−derivative of the DG solution exhibits an optimal O (hp+1−n) convergence rates in the L2-norm when p-degree piecewise polynomials with p≥1 are used. We further found that the odd-derivatives and the even derivatives are super convergent, respectively, at the upwind and downwind endpoints.Keywords: discontinuous, galerkin, superconvergence, higherorder, error, estimates
Procedia PDF Downloads 4791459 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second
Authors: P. V. Pramila , V. Mahesh
Abstract:
Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients esulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF 25, PEF,FEF 25-75, FEF50, and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF 25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects). It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.Keywords: FEV, multivariate adaptive regression splines pulmonary function test, random forest
Procedia PDF Downloads 3121458 Cationic Surfactants Influence on the Fouling Phenomenon Control in Ultrafiltration of Latex Contaminated Water and Wastewater
Authors: Amira Abdelrasoul, Huu Doan, Ali Lohi
Abstract:
The goal of the present study was to minimize the ultrafiltration fouling of latex effluent using Cetyltrimethyl ammonium bromide (CTAB) as a cationic surfactant. Hydrophilic Polysulfone and Ultrafilic flat heterogeneous membranes, with MWCO of 60,000 and 100,000, respectively, as well as hydrophobic Polyvinylidene Difluoride with MWCO of 100,000, were used under a constant flow rate and cross-flow mode in ultrafiltration of latex solution. In addition, a Polycarbonate flat membrane with uniform pore size of 0.05 µm was also used. The effect of CTAB on the latex particle size distribution was investigated at different concentrations, various treatment times, and diverse agitation duration. The effects of CTAB on the zeta potential of latex particles and membrane surfaces were also investigated. The results obtained indicated that the particle size distribution of treated latex effluent showed noticeable shifts in the peaks toward a larger size range due to the aggregation of particles. As a consequence, the mass of fouling contributing to pore blocking and the irreversible fouling were significantly reduced. The optimum results occurred with the addition of CTAB at the critical micelle concentration of 0.36 g/L for 10 minutes with minimal agitation. Higher stirring rate had a negative effect on membrane fouling minimization.Keywords: cationic surfactant, latex particles, membrane fouling, ultrafiltration, zeta potential
Procedia PDF Downloads 5291457 A Higher Order Shear and Normal Deformation Theory for Functionally Graded Sandwich Beam
Authors: R. Bennai, H. Ait Atmane, Jr., A. Tounsi
Abstract:
In this work, a new analytical approach using a refined theory of hyperbolic shear deformation of a beam was developed to study the free vibration of graduated sandwiches beams under different boundary conditions. The effects of transverse shear strains and the transverse normal deformation are considered. The constituent materials of the beam are supposed gradually variable depending the height direction based on a simple power distribution law in terms of the volume fractions of the constituents; the two materials with which we worked are metals and ceramics. The core layer is taken homogeneous and made of an isotropic material; while the banks layers consist of FGM materials with a homogeneous fraction compared to the middle layer. Movement equations are obtained by the energy minimization principle. Analytical solutions of free vibration and buckling are obtained for sandwich beams under different support conditions; these conditions are taken into account by incorporating new form functions. In the end, illustrative examples are presented to show the effects of changes in different parameters such as (material graduation, the stretching effect of the thickness, boundary conditions and thickness ratio - length) on the vibration free and buckling of an FGM sandwich beams.Keywords: functionally graded sandwich beam, refined shear deformation theory, stretching effect, free vibration
Procedia PDF Downloads 2481456 Measuring the Quality of Business Education: Employment Readiness Assessment
Authors: Gulbakhyt Sultanova
Abstract:
Business education institutions assess the progress of their students by giving them grades for courses completed and calculating a Grade Point Average (GPA). Whether the participation in these courses has led to the development of competences enabling graduates to successfully compete in the labor market should be measured using a new index: Employment Readiness Assessment (ERA). The higher the ERA, the higher the quality of education at a business school. This is applied, empirical research conducted by using a method of linear optimization. The aim of research is to identify factors which lead to the minimization of the deviation of GPA from ERA as well as to the maximization of ERA. ERA is composed of three components resulting from testing proficiency in Business English, testing work and personal skills, and job interview simulation. The quality of education is improving if GPA approximates ERA and ERA increases. Factors which have had a positive effect on quality enhancement are academic mobility of students and staff, practical-oriented courses taught by staff with work experience, and research-based courses taught by staff with research experience. ERA is a better index to measure the quality of business education than traditional indexes such as GPA due to its greater accuracy in assessing the level of graduates’ competences demanded in the labor market. Optimizing the educational process in pursuit of quality enhancement, ERA has to be used in parallel with GPA to find out which changes worked and resulted in improvement.Keywords: assessment and evaluation, competence evaluation, education quality, employment readiness
Procedia PDF Downloads 4461455 Optimization of a Flux Switching Permanent Magnet Machine Using Laminated Segmented Rotor
Authors: Seyedmilad Kazemisangdehi, Seyedmehdi Kazemisangdehi
Abstract:
Flux switching permanent magnet machines are considered for wide range of applications because of their outstanding merits including high torque/power densities, high efficiency, simple and robust rotor structure. Therefore, several topologies have been proposed like the PM exited flux switching machine, hybrid excited flux switching type, and so on. Recently, a novel laminated segmented rotor flux switching permanent magnet machine was introduced. It features flux barriers on rotor structure to enhance the performances of machine including torque ripple reduction and also torque and efficiency improvements at the same time. This is while, the design of barriers was not optimized by the authors. Therefore, in this paper three coefficients regarding the position of the barriers are considered for optimization. The effect of each coefficient on the performance of this machine is investigated by finite element method and finally an optimized design of flux barriers based on these three coefficients is proposed from different points of view including electromagnetic torque maximization and cogging torque/torque ripple minimization. At optimum design from maximum developed torque aspect, this machine generates 0.65 Nm torque higher than that of the not-optimized design with an almost 0.4 % improvement in efficiency.Keywords: finite element analysis, FSPM, laminated segmented rotor flux switching permanent magnet machine, optimization
Procedia PDF Downloads 2321454 Heat-Induced Uncertainty of Industrial Computed Tomography Measuring a Stainless Steel Cylinder
Authors: Verena M. Moock, Darien E. Arce Chávez, Mariana M. Espejel González, Leopoldo Ruíz-Huerta, Crescencio García-Segundo
Abstract:
Uncertainty analysis in industrial computed tomography is commonly related to metrological trace tools, which offer precision measurements of external part features. Unfortunately, there is no such reference tool for internal measurements to profit from the unique imaging potential of X-rays. Uncertainty approximations for computed tomography are still based on general aspects of the industrial machine and do not adapt to acquisition parameters or part characteristics. The present study investigates the impact of the acquisition time on the dimensional uncertainty measuring a stainless steel cylinder with a circular tomography scan. The authors develop the figure difference method for X-ray radiography to evaluate the volumetric differences introduced within the projected absorption maps of the metal workpiece. The dimensional uncertainty is dominantly influenced by photon energy dissipated as heat causing the thermal expansion of the metal, as monitored by an infrared camera within the industrial tomograph. With the proposed methodology, we are able to show evolving temperature differences throughout the tomography acquisition. This is an early study showing that the number of projections in computer tomography induces dimensional error due to energy absorption. The error magnitude would depend on the thermal properties of the sample and the acquisition parameters by placing apparent non-uniform unwanted volumetric expansion. We introduce infrared imaging for the experimental display of metrological uncertainty in a particular metal part of symmetric geometry. We assess that the current results are of fundamental value to reach the balance between the number of projections and uncertainty tolerance when performing analysis with X-ray dimensional exploration in precision measurements with industrial tomography.Keywords: computed tomography, digital metrology, infrared imaging, thermal expansion
Procedia PDF Downloads 1241453 Investigation of Delivery of Triple Play Data in GE-PON Fiber to the Home Network
Authors: Ashima Anurag Sharma
Abstract:
Optical fiber based networks can deliver performance that can support the increasing demands for high speed connections. One of the new technologies that have emerged in recent years is Passive Optical Networks. This research paper is targeted to show the simultaneous delivery of triple play service (data, voice, and video). The comparison between various data rates is presented. It is demonstrated that as we increase the data rate, number of users to be decreases due to increase in bit error rate.Keywords: BER, PON, TDMPON, GPON, CWDM, OLT, ONT
Procedia PDF Downloads 5291452 Use of Artificial Neural Networks to Estimate Evapotranspiration for Efficient Irrigation Management
Authors: Adriana Postal, Silvio C. Sampaio, Marcio A. Villas Boas, Josué P. Castro
Abstract:
This study deals with the estimation of reference evapotranspiration (ET₀) in an agricultural context, focusing on efficient irrigation management to meet the growing interest in the sustainable management of water resources. Given the importance of water in agriculture and its scarcity in many regions, efficient use of this resource is essential to ensure food security and environmental sustainability. The methodology used involved the application of artificial intelligence techniques, specifically Multilayer Perceptron (MLP) Artificial Neural Networks (ANNs), to predict ET₀ in the state of Paraná, Brazil. The models were trained and validated with meteorological data from the Brazilian National Institute of Meteorology (INMET), together with data obtained from a producer's weather station in the western region of Paraná. Two optimizers (SGD and Adam) and different meteorological variables, such as temperature, humidity, solar radiation, and wind speed, were explored as inputs to the models. Nineteen configurations with different input variables were tested; amidst them, configuration 9, with 8 input variables, was identified as the most efficient of all. Configuration 10, with 4 input variables, was considered the most effective, considering the smallest number of variables. The main conclusions of this study show that MLP ANNs are capable of accurately estimating ET₀, providing a valuable tool for irrigation management in agriculture. Both configurations (9 and 10) showed promising performance in predicting ET₀. The validation of the models with cultivator data underlined the practical relevance of these tools and confirmed their generalization ability for different field conditions. The results of the statistical metrics, including Mean Absolute Error (MAE), Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Coefficient of Determination (R²), showed excellent agreement between the model predictions and the observed data, with MAE as low as 0.01 mm/day and 0.03 mm/day, respectively. In addition, the models achieved an R² between 0.99 and 1, indicating a satisfactory fit to the real data. This agreement was also confirmed by the Kolmogorov-Smirnov test, which evaluates the agreement of the predictions with the statistical behavior of the real data and yields values between 0.02 and 0.04 for the producer data. In addition, the results of this study suggest that the developed technique can be applied to other locations by using specific data from these sites to further improve ET₀ predictions and thus contribute to sustainable irrigation management in different agricultural regions. The study has some limitations, such as the use of a single ANN architecture and two optimizers, the validation with data from only one producer, and the possible underestimation of the influence of seasonality and local climate variability. An irrigation management application using the most efficient models from this study is already under development. Future research can explore different ANN architectures and optimization techniques, validate models with data from multiple producers and regions, and investigate the model's response to different seasonal and climatic conditions.Keywords: agricultural technology, neural networks in agriculture, water efficiency, water use optimization
Procedia PDF Downloads 51