Search results for: minimum root mean square (RMS) error matching algorithm
7470 Development of a Multi-Variate Model for Matching Plant Nitrogen Requirements with Supply for Reducing Losses in Dairy Systems
Authors: Iris Vogeler, Rogerio Cichota, Armin Werner
Abstract:
Dairy farms are under pressure to increase productivity while reducing environmental impacts. Effective fertiliser management practices are critical to achieve this. Determination of optimum nitrogen (N) fertilisation rates which maximise pasture growth and minimise N losses is challenging due to variability in plant requirements and likely near-future supply of N by the soil. Remote sensing can be used for mapping N nutrition status of plants and to rapidly assess the spatial variability within a field. An algorithm is, however, lacking which relates the N status of the plants to the expected yield response to additions of N. The aim of this simulation study was to develop a multi-variate model for determining N fertilisation rate for a target percentage of the maximum achievable yield based on the pasture N concentration (ii) use of an algorithm for guiding fertilisation rates, and (iii) evaluation of the model regarding pasture yield and N losses, including N leaching, denitrification and volatilisation. A simulation study was carried out using the Agricultural Production Systems Simulator (APSIM). The simulations were done for an irrigated ryegrass pasture in the Canterbury region of New Zealand. A multi-variate model was developed and used to determine monthly required N fertilisation rates based on pasture N content prior to fertilisation and targets of 50, 75, 90 and 100% of the potential monthly yield. These monthly optimised fertilisation rules were evaluated by running APSIM for a ten-year period to provide yield and N loss estimates from both nonurine and urine affected areas. Comparison with typical fertilisation rates of 150 and 400 kg N/ha/year was also done. Assessment of pasture yield and leaching from fertiliser and urine patches indicated a large reduction in N losses when N fertilisation rates were controlled by the multi-variate model. However, the reduction in leaching losses was much smaller when taking into account the effects of urine patches. The proposed approach based on biophysical modelling to develop a multi-variate model for determining optimum N fertilisation rates dependent on pasture N content is very promising. Further analysis, under different environmental conditions and validation is required before the approach can be used to help adjust fertiliser management practices to temporal and spatial N demand based on the nitrogen status of the pasture.Keywords: APSIM modelling, optimum N fertilization rate, pasture N content, ryegrass pasture, three dimensional surface response function.
Procedia PDF Downloads 1307469 Blind Super-Resolution Reconstruction Based on PSF Estimation
Authors: Osama A. Omer, Amal Hamed
Abstract:
Successful blind image Super-Resolution algorithms require the exact estimation of the Point Spread Function (PSF). In the absence of any prior information about the imagery system and the true image; this estimation is normally done by trial and error experimentation until an acceptable restored image quality is obtained. Multi-frame blind Super-Resolution algorithms often have disadvantages of slow convergence and sensitiveness to complex noises. This paper presents a Super-Resolution image reconstruction algorithm based on estimation of the PSF that yields the optimum restored image quality. The estimation of PSF is performed by the knife-edge method and it is implemented by measuring spreading of the edges in the reproduced HR image itself during the reconstruction process. The proposed image reconstruction approach is using L1 norm minimization and robust regularization based on a bilateral prior to deal with different data and noise models. A series of experiment results show that the proposed method can outperform other previous work robustly and efficiently.Keywords: blind, PSF, super-resolution, knife-edge, blurring, bilateral, L1 norm
Procedia PDF Downloads 3657468 A Fast Community Detection Algorithm
Authors: Chung-Yuan Huang, Yu-Hsiang Fu, Chuen-Tsai Sun
Abstract:
Community detection represents an important data-mining tool for analyzing and understanding real-world complex network structures and functions. We believe that at least four criteria determine the appropriateness of a community detection algorithm: (a) it produces useable normalized mutual information (NMI) and modularity results for social networks, (b) it overcomes resolution limitation problems associated with synthetic networks, (c) it produces good NMI results and performance efficiency for Lancichinetti-Fortunato-Radicchi (LFR) benchmark networks, and (d) it produces good modularity and performance efficiency for large-scale real-world complex networks. To our knowledge, no existing community detection algorithm meets all four criteria. In this paper, we describe a simple hierarchical arc-merging (HAM) algorithm that uses network topologies and rule-based arc-merging strategies to identify community structures that satisfy the criteria. We used five well-studied social network datasets and eight sets of LFR benchmark networks to validate the ground-truth community correctness of HAM, eight large-scale real-world complex networks to measure its performance efficiency, and two synthetic networks to determine its susceptibility to resolution limitation problems. Our results indicate that the proposed HAM algorithm is capable of providing satisfactory performance efficiency and that HAM-identified communities were close to ground-truth communities in social and LFR benchmark networks while overcoming resolution limitation problems.Keywords: complex network, social network, community detection, network hierarchy
Procedia PDF Downloads 2277467 Prediction of the Aerodynamic Stall of a Helicopter’s Main Rotor Using a Computational Fluid Dynamics Analysis
Authors: Assel Thami Lahlou, Soufiane Stouti, Ismail Lagrat, Hamid Mounir, Oussama Bouazaoui
Abstract:
The purpose of this research work is to predict the helicopter from stalling by finding the minimum and maximum values that the pitch angle can take in order to fly in a hover state condition. The stall of a helicopter in hover occurs when the pitch angle is too small to generate the thrust required to support its weight or when the critical angle of attack that gives maximum lift is reached or exceeded. In order to find the minimum pitch angle, a 3D CFD simulation was done in this work using ANSYS FLUENT as the CFD solver. We started with a small value of the pitch angle θ, and we kept increasing its value until we found the thrust coefficient required to fly in a hover state and support the weight of the helicopter. For the CFD analysis, the Multiple Reference Frame (MRF) method with k-ε turbulent model was used to study the 3D flow around the rotor for θmin. On the other hand, a 2D simulation of the airfoil NACA 0012 was executed with a velocity inlet Vin=ΩR/2 to visualize the flow at the location span R/2 of the disk rotor using the Spallart-Allmaras turbulent model. Finding the critical angle of attack at this position will give us the ability to predict the stall in hover flight. The results obtained will be exposed later in the article. This study was so useful in analyzing the limitations of the helicopter’s main rotor and thus, in predicting accidents that can lead to a lot of damage.Keywords: aerodynamic, CFD, helicopter, stall, blades, main rotor, minimum pitch angle, maximum pitch angle
Procedia PDF Downloads 817466 Fuzzy Logic Control for Flexible Joint Manipulator: An Experimental Implementation
Authors: Sophia Fry, Mahir Irtiza, Alexa Hoffman, Yousef Sardahi
Abstract:
This study presents an intelligent control algorithm for a flexible robotic arm. Fuzzy control is used to control the motion of the arm to maintain the arm tip at the desired position while reducing vibration and increasing the system speed of response. The Fuzzy controller (FC) is based on adding the tip angular position to the arm deflection angle and using their sum as a feedback signal to the control algorithm. This reduces the complexity of the FC in terms of the input variables, number of membership functions, fuzzy rules, and control structure. Also, the design of the fuzzy controller is model-free and uses only our knowledge about the system. To show the efficacy of the FC, the control algorithm is implemented on the flexible joint manipulator (FJM) developed by Quanser. The results show that the proposed control method is effective in terms of response time, overshoot, and vibration amplitude.Keywords: fuzzy logic control, model-free control, flexible joint manipulators, nonlinear control
Procedia PDF Downloads 1187465 Digital Watermarking Based on Visual Cryptography and Histogram
Authors: R. Rama Kishore, Sunesh
Abstract:
Nowadays, robust and secure watermarking algorithm and its optimization have been need of the hour. A watermarking algorithm is presented to achieve the copy right protection of the owner based on visual cryptography, histogram shape property and entropy. In this, both host image and watermark are preprocessed. Host image is preprocessed by using Butterworth filter, and watermark is with visual cryptography. Applying visual cryptography on water mark generates two shares. One share is used for embedding the watermark, and the other one is used for solving any dispute with the aid of trusted authority. Usage of histogram shape makes the process more robust against geometric and signal processing attacks. The combination of visual cryptography, Butterworth filter, histogram, and entropy can make the algorithm more robust, imperceptible, and copy right protection of the owner.Keywords: digital watermarking, visual cryptography, histogram, butter worth filter
Procedia PDF Downloads 3587464 Indian Road Traffic Flow Analysis Using Blob Tracking from Video Sequences
Authors: Balaji Ganesh Rajagopal, Subramanian Appavu alias Balamurugan, Ayyalraj Midhun Kumar, Krishnan Nallaperumal
Abstract:
Intelligent Transportation System is an Emerging area to solve multiple transportation problems. Several forms of inputs are needed in order to solve ITS problems. Advanced Traveler Information System (ATIS) is a core and important ITS area of this modern era. This involves travel time forecasting, efficient road map analysis and cost based path selection, Detection of the vehicle in the dynamic conditions and Traffic congestion state forecasting. This Article designs and provides an algorithm for traffic data generation which can be used for the above said ATIS application. By inputting the real world traffic situation in the form of video sequences, the algorithm determines the Traffic density in terms of congestion, number of vehicles in a given path which can be fed for various ATIS applications. The Algorithm deduces the key frame from the video sequences and follows the Blob detection, Identification and Tracking using connected components algorithm to determine the correlation between the vehicles moving in the real road scene.Keywords: traffic transportation, traffic density estimation, blob identification and tracking, relative velocity of vehicles, correlation between vehicles
Procedia PDF Downloads 5107463 Assessment of Image Databases Used for Human Skin Detection Methods
Authors: Saleh Alshehri
Abstract:
Human skin detection is a vital step in many applications. Some of the applications are critical especially those related to security. This leverages the importance of a high-performance detection algorithm. To validate the accuracy of the algorithm, image databases are usually used. However, the suitability of these image databases is still questionable. It is suggested that the suitability can be measured mainly by the span the database covers of the color space. This research investigates the validity of three famous image databases.Keywords: image databases, image processing, pattern recognition, neural networks
Procedia PDF Downloads 2717462 Neural Network and Support Vector Machine for Prediction of Foot Disorders Based on Foot Analysis
Authors: Monireh Ahmadi Bani, Adel Khorramrouz, Lalenoor Morvarid, Bagheri Mahtab
Abstract:
Background:- Foot disorders are common in musculoskeletal problems. Plantar pressure distribution measurement is one the most important part of foot disorders diagnosis for quantitative analysis. However, the association of plantar pressure and foot disorders is not clear. With the growth of dataset and machine learning methods, the relationship between foot disorders and plantar pressures can be detected. Significance of the study:- The purpose of this study was to predict the probability of common foot disorders based on peak plantar pressure distribution and center of pressure during walking. Methodologies:- 2323 participants were assessed in a foot therapy clinic between 2015 and 2021. Foot disorders were diagnosed by an experienced physician and then they were asked to walk on a force plate scanner. After the data preprocessing, due to the difference in walking time and foot size, we normalized the samples based on time and foot size. Some of force plate variables were selected as input to a deep neural network (DNN), and the probability of any each foot disorder was measured. In next step, we used support vector machine (SVM) and run dataset for each foot disorder (classification of yes or no). We compared DNN and SVM for foot disorders prediction based on plantar pressure distributions and center of pressure. Findings:- The results demonstrated that the accuracy of deep learning architecture is sufficient for most clinical and research applications in the study population. In addition, the SVM approach has more accuracy for predictions, enabling applications for foot disorders diagnosis. The detection accuracy was 71% by the deep learning algorithm and 78% by the SVM algorithm. Moreover, when we worked with peak plantar pressure distribution, it was more accurate than center of pressure dataset. Conclusion:- Both algorithms- deep learning and SVM will help therapist and patients to improve the data pool and enhance foot disorders prediction with less expense and error after removing some restrictions properly.Keywords: deep neural network, foot disorder, plantar pressure, support vector machine
Procedia PDF Downloads 3587461 On the Accuracy of Basic Modal Displacement Method Considering Various Earthquakes
Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar
Abstract:
Time history seismic analysis is supposed to be the most accurate method to predict the seismic demand of structures. On the other hand, the required computational time of this method toward achieving the result is its main deficiency. While being applied in optimization process, in which the structure must be analyzed thousands of time, reducing the required computational time of seismic analysis of structures makes the optimization algorithms more practical. Apparently, the invented approximate methods produce some amount of errors in comparison with exact time history analysis but the recently proposed method namely, Complete Quadratic Combination (CQC) and Sum Root of the Sum of Squares (SRSS) drastically reduces the computational time by combination of peak responses in each mode. In the present research, the Basic Modal Displacement (BMD) method is introduced and applied towards estimation of seismic demand of main structure. Seismic demand of sampled structure is estimated by calculation of modal displacement of basic structure (in which the modal displacement has been calculated). Shear steel sampled structures are selected as case studies. The error applying the introduced method is calculated by comparison of the estimated seismic demands with exact time history dynamic analysis. The efficiency of the proposed method is demonstrated by application of three types of earthquakes (in view of time of peak ground acceleration).Keywords: time history dynamic analysis, basic modal displacement, earthquake-induced demands, shear steel structures
Procedia PDF Downloads 3557460 Scheduling Nodes Activity and Data Communication for Target Tracking in Wireless Sensor Networks
Authors: AmirHossein Mohajerzadeh, Mohammad Alishahi, Saeed Aslishahi, Mohsen Zabihi
Abstract:
In this paper, we consider sensor nodes with the capability of measuring the bearings (relative angle to the target). We use geometric methods to select a set of observer nodes which are responsible for collecting data from the target. Considering the characteristics of target tracking applications, it is clear that significant numbers of sensor nodes are usually inactive. Therefore, in order to minimize the total network energy consumption, a set of sensor nodes, called sentinel, is periodically selected for monitoring, controlling the environment and transmitting data through the network. The other nodes are inactive. Furthermore, the proposed algorithm provides a joint scheduling and routing algorithm to transmit data between network nodes and the fusion center (FC) in which not only provides an efficient way to estimate the target position but also provides an efficient target tracking. Performance evaluation confirms the superiority of the proposed algorithm.Keywords: coverage, routing, scheduling, target tracking, wireless sensor networks
Procedia PDF Downloads 3787459 Proportional and Integral Controller-Based Direct Current Servo Motor Speed Characterization
Authors: Adel Salem Bahakeem, Ahmad Jamal, Mir Md. Maruf Morshed, Elwaleed Awad Khidir
Abstract:
Direct Current (DC) servo motors, or simply DC motors, play an important role in many industrial applications such as manufacturing of plastics, precise positioning of the equipment, and operating computer-controlled systems where speed of feed control, maintaining the position, and ensuring to have a constantly desired output is very critical. These parameters can be controlled with the help of control systems such as the Proportional Integral Derivative (PID) controller. The aim of the current work is to investigate the effects of Proportional (P) and Integral (I) controllers on the steady state and transient response of the DC motor. The controller gains are varied to observe their effects on the error, damping, and stability of the steady and transient motor response. The current investigation is conducted experimentally on a servo trainer CE 110 using analog PI controller CE 120 and theoretically using Simulink in MATLAB. Both experimental and theoretical work involves varying integral controller gain to obtain the response to a steady-state input, varying, individually, the proportional and integral controller gains to obtain the response to a step input function at a certain frequency, and theoretically obtaining the proportional and integral controller gains for desired values of damping ratio and response frequency. Results reveal that a proportional controller helps reduce the steady-state and transient error between the input signal and output response and makes the system more stable. In addition, it also speeds up the response of the system. On the other hand, the integral controller eliminates the error but tends to make the system unstable with induced oscillations and slow response to eliminate the error. From the current work, it is desired to achieve a stable response of the servo motor in terms of its angular velocity subjected to steady-state and transient input signals by utilizing the strengths of both P and I controllers.Keywords: DC servo motor, proportional controller, integral controller, controller gain optimization, Simulink
Procedia PDF Downloads 1107458 Arbuscular Mycorrhizal Symbiosis in Trema orientalis: Effect of a Naturally-Occurring Symbiosis Receptor Kinase Mutant Allele
Authors: Yuda Purwana Roswanjaya, Wouter Kohlen, Rene Geurts
Abstract:
The Trema genus represents a group of fast-growing tropical tree species within the Cannabaceae. Interestingly, five species nested in this lineage -known as Parasponia- can establish rhizobium nitrogen-fixing root nodules, similar to those found in legumes. Parasponia and legumes use a conserved genetic network to control root nodule formation, among which are genes also essential for mycorrhizal symbiosis (the so-called common symbiotic pathway). However, Trema species lost several genes that function exclusively in nodulation, suggesting a loss-of the nodulation trait in Trema. Strikingly, in a Trema orientalis population found in Malaysian Borneo we identified a truncated SYMBIOSIS RECEPTOR KINASE (SYMRK) mutant allele lacking a large portion of the c-terminal kinase domain. In legumes this gene is essential for nodulation and mycorrhization. This raises the question whether Trema orientalis can still be mycorrhized. To answer this question, we established quantitative mycorrhization assay for Parasponia andersonii and Trema orientalis. Plants were grown in closed pots on half strength Hoagland medium containing 20 µM potassium phosphate in sterilized sand and inoculated with 125 spores of Rhizopagus irregularis (Agronutrion-DAOM197198). Mycorrhization efficiency was determined by analyzing the frequency of mycorrhiza (%F), the intensity of the mycorrhizal colonization (%M) and the arbuscule abundance (%A) in the root system. Trema orientalis RG33 can be mycorrhized, though with lower efficiency compared to Parasponia andersonii. From this we conclude that a functional SYMRK kinase domain is not essential for Trema orientalis mycorrhization. In ongoing experiments, we aim to investigate the role of SYMRK in Parasponia andersonii mycorrhization and nodulation. For this two Parasponia andersonii symrk CRISPR-Cas9 mutant alleles were created. One mimicking the TorSYMRKRG33 allele by deletion of exon 13-15, and a full Parasponia andersonii SYMRK knockout.Keywords: endomycorrhization, Parasponia andersonii, symbiosis receptor kinase (SYMRK), Trema orientalis
Procedia PDF Downloads 1647457 Representativity Based Wasserstein Active Regression
Authors: Benjamin Bobbia, Matthias Picard
Abstract:
In recent years active learning methodologies based on the representativity of the data seems more promising to limit overfitting. The presented query methodology for regression using the Wasserstein distance measuring the representativity of our labelled dataset compared to the global distribution. In this work a crucial use of GroupSort Neural Networks is made therewith to draw a double advantage. The Wasserstein distance can be exactly expressed in terms of such neural networks. Moreover, one can provide explicit bounds for their size and depth together with rates of convergence. However, heterogeneity of the dataset is also considered by weighting the Wasserstein distance with the error of approximation at the previous step of active learning. Such an approach leads to a reduction of overfitting and high prediction performance after few steps of query. After having detailed the methodology and algorithm, an empirical study is presented in order to investigate the range of our hyperparameters. The performances of this method are compared, in terms of numbers of query needed, with other classical and recent query methods on several UCI datasets.Keywords: active learning, Lipschitz regularization, neural networks, optimal transport, regression
Procedia PDF Downloads 807456 The Properties of Risk-based Approaches to Asset Allocation Using Combined Metrics of Portfolio Volatility and Kurtosis: Theoretical and Empirical Analysis
Authors: Maria Debora Braga, Luigi Riso, Maria Grazia Zoia
Abstract:
Risk-based approaches to asset allocation are portfolio construction methods that do not rely on the input of expected returns for the asset classes in the investment universe and only use risk information. They include the Minimum Variance Strategy (MV strategy), the traditional (volatility-based) Risk Parity Strategy (SRP strategy), the Most Diversified Portfolio Strategy (MDP strategy) and, for many, the Equally Weighted Strategy (EW strategy). All the mentioned approaches were based on portfolio volatility as a reference risk measure but in 2023, the Kurtosis-based Risk Parity strategy (KRP strategy) and the Minimum Kurtosis strategy (MK strategy) were introduced. Understandably, they used the fourth root of the portfolio-fourth moment as a proxy for portfolio kurtosis to work with a homogeneous function of degree one. This paper contributes mainly theoretically and methodologically to the framework of risk-based asset allocation approaches with two steps forward. First, a new and more flexible objective function considering a linear combination (with positive coefficients that sum to one) of portfolio volatility and portfolio kurtosis is used to alternatively serve a risk minimization goal or a homogeneous risk distribution goal. Hence, the new basic idea consists in extending the achievement of typical risk-based approaches’ goals to a combined risk measure. To give the rationale behind operating with such a risk measure, it is worth remembering that volatility and kurtosis are expressions of uncertainty, to be read as dispersion of returns around the mean and that both preserve adherence to a symmetric framework and consideration for the entire returns distribution as well, but also that they differ from each other in that the former captures the “normal” / “ordinary” dispersion of returns, while the latter is able to catch the huge dispersion. Therefore, the combined risk metric that uses two individual metrics focused on the same phenomena but differently sensitive to its intensity allows the asset manager to express, in the context of an objective function by varying the “relevance coefficient” associated with the individual metrics, alternatively, a wide set of plausible investment goals for the portfolio construction process while serving investors differently concerned with tail risk and traditional risk. Since this is the first study that also implements risk-based approaches using a combined risk measure, it becomes of fundamental importance to investigate the portfolio effects triggered by this innovation. The paper also offers a second contribution. Until the recent advent of the MK strategy and the KRP strategy, efforts to highlight interesting properties of risk-based approaches were inevitably directed towards the traditional MV strategy and SRP strategy. Previous literature established an increasing order in terms of portfolio volatility, starting from the MV strategy, through the SRP strategy, arriving at the EQ strategy and provided the mathematical proof for the “equalization effect” concerning marginal risks when the MV strategy is considered, and concerning risk contributions when the SRP strategy is considered. Regarding the validity of similar conclusions when referring to the MK strategy and KRP strategy, the development of a theoretical demonstration is still pending. This paper fills this gap.Keywords: risk parity, portfolio kurtosis, risk diversification, asset allocation
Procedia PDF Downloads 657455 Simultaneous Electrochemical Detection of Chromium(III), Arsenic(III), and Mercury (II) In Water Using Anodic Stripping Voltammetry
Authors: V. Sai Geethika, Sai Snehitha Yadavalli, Swati Ghosh Acharyya
Abstract:
This study involves a single element and simultaneous electrochemical detection of heavy metal ions through square wave anodic stripping voltammetry. A glassy carbon electrode was used to detect and quantify heavy metals such as As(III), Hg(II), Cr(VI) ions in water. Under optimized conditions, peak separation was obtained by varying concentrations, scan rates, and temperatures. As (III), Hg (II), Cr (III) were simultaneously detected with GCE. Several analytical methods, such as inductively coupled plasma mass spectroscopy (ICP-MS), atomic absorption spectroscopy (AAS), were used previously to detect heavy metal ions, which are authentic but are not good enough for online monitoring due to the bulkiness of the equipment. The study provides a good alternative that is simple, more efficient, and low-cost, involving a portable potentiostat. Heavy metals having different oxidation states can be detected by anodic stripping voltammetry. This method can be easily integrated with electronics. Square wave Anodic stripping voltammetry is used with a potential range of -2.5 V – 2.5 V for single ion detection by a three-electrode cell consisting of silver/silver chloride(Ag/AgCl) as reference and platinum (Pt) counter and glassy carbon (GCE) working electrodes. All three ions are optimized by varying the parameters like concentration, scan rate, pH, temperature, and all these optimized parameters were used for studying the effects of simultaneous detection. The procedure involves preparing an electrolyte using deionized water, cleaning the surface of GCE, depositing the ions by applying the redox potentials obtained from cyclic voltammetry (CV), and then detecting by applying oxidizing potential, i.e., stripping voltage. So this includes ASV techniques such as open-circuit voltage (OCV), chronoamperometry (CA), and square wave voltammetry (SWV). Firstly, the concentration of the ions varied from 50 ppb to 5000 ppb, and an optimum concentration was determined where the three ions were detected. A concentration of 400 ppb was used while varying the temperatures in the range of 25°C – 45°C. Optimum peak intensity was obtained at a temperature of 30°C with a low scan rate of 0.005 V-s⁻¹. All the parameters were optimized, and several effects have been noticed while three ions As(II), Cr(III), Hg(II) were detected alone and simultaneously.Keywords: Arsenic(III), Chromium(III), glassy carbon electrode, Mercury (II), square wave anodic stripping voltammetry
Procedia PDF Downloads 857454 Entropy Generation Analyze Due to the Steady Natural Convection of Newtonian Fluid in a Square Enclosure
Authors: T. T. Naas, Y. Lasbet, C. Kezrane
Abstract:
The thermal control in many systems is widely accomplished applying mixed convection process due to its low cost, reliability and easy maintenance. Typical applications include the aircraft electronic equipment, rotating-disc heat exchangers, turbo machinery, and nuclear reactors, etc. Natural convection in an inclined square enclosure heated via wall heater has been studied numerically. Finite volume method is used for solving momentum and energy equations in the form of stream function–vorticity. The right and left walls are kept at a constant temperature, while the other parts are adiabatic. The range of the inclination angle covers a whole revolution. The method is validated for a vertical cavity. A general power law dependence of the Nusselt number with respect to the Rayleigh number with the coefficient and exponent as functions of the inclination angle is presented. For a fixed Rayleigh number, the inclination angle increases or decreases is found.Keywords: natural convection in enclosure, inclined enclosure, Nusselt number, entropy generation analyze
Procedia PDF Downloads 2607453 Crow Search Algorithm-Based Task Offloading Strategies for Fog Computing Architectures
Authors: Aniket Ganvir, Ritarani Sahu, Suchismita Chinara
Abstract:
The rapid digitization of various aspects of life is leading to the creation of smart IoT ecosystems, where interconnected devices generate significant amounts of valuable data. However, these IoT devices face constraints such as limited computational resources and bandwidth. Cloud computing emerges as a solution by offering ample resources for offloading tasks efficiently despite introducing latency issues, especially for time-sensitive applications like fog computing. Fog computing (FC) addresses latency concerns by bringing computation and storage closer to the network edge, minimizing data travel distance, and enhancing efficiency. Offloading tasks to fog nodes or the cloud can conserve energy and extend IoT device lifespan. The offloading process is intricate, with tasks categorized as full or partial, and its optimization presents an NP-hard problem. Traditional greedy search methods struggle to address the complexity of task offloading efficiently. To overcome this, the efficient crow search algorithm (ECSA) has been proposed as a meta-heuristic optimization algorithm. ECSA aims to effectively optimize computation offloading, providing solutions to this challenging problem.Keywords: IoT, fog computing, task offloading, efficient crow search algorithm
Procedia PDF Downloads 587452 Profit-Based Artificial Neural Network (ANN) Trained by Migrating Birds Optimization: A Case Study in Credit Card Fraud Detection
Authors: Ashkan Zakaryazad, Ekrem Duman
Abstract:
A typical classification technique ranks the instances in a data set according to the likelihood of belonging to one (positive) class. A credit card (CC) fraud detection model ranks the transactions in terms of probability of being fraud. In fact, this approach is often criticized, because firms do not care about fraud probability but about the profitability or costliness of detecting a fraudulent transaction. The key contribution in this study is to focus on the profit maximization in the model building step. The artificial neural network proposed in this study works based on profit maximization instead of minimizing the error of prediction. Moreover, some studies have shown that the back propagation algorithm, similar to other gradient–based algorithms, usually gets trapped in local optima and swarm-based algorithms are more successful in this respect. In this study, we train our profit maximization ANN using the Migrating Birds optimization (MBO) which is introduced to literature recently.Keywords: neural network, profit-based neural network, sum of squared errors (SSE), MBO, gradient descent
Procedia PDF Downloads 4757451 A Design of Elliptic Curve Cryptography Processor based on SM2 over GF(p)
Authors: Shiji Hu, Lei Li, Wanting Zhou, DaoHong Yang
Abstract:
The data encryption, is the foundation of today’s communication. On this basis, how to improve the speed of data encryption and decryption is always a problem that scholars work for. In this paper, we proposed an elliptic curve crypto processor architecture based on SM2 prime field. In terms of hardware implementation, we optimized the algorithms in different stages of the structure. In finite field modulo operation, we proposed an optimized improvement of Karatsuba-Ofman multiplication algorithm, and shorten the critical path through pipeline structure in the algorithm implementation. Based on SM2 recommended prime field, a fast modular reduction algorithm is used to reduce 512-bit wide data obtained from the multiplication unit. The radix-4 extended Euclidean algorithm was used to realize the conversion between affine coordinate system and Jacobi projective coordinate system. In the parallel scheduling of point operations on elliptic curves, we proposed a three-level parallel structure of point addition and point double based on the Jacobian projective coordinate system. Combined with the scalar multiplication algorithm, we added mutual pre-operation to the point addition and double point operation to improve the efficiency of the scalar point multiplication. The proposed ECC hardware architecture was verified and implemented on Xilinx Virtex-7 and ZYNQ-7 platforms, and each 256-bit scalar multiplication operation took 0.275ms. The performance for handling scalar multiplication is 32 times that of CPU(dual-core ARM Cortex-A9).Keywords: Elliptic curve cryptosystems, SM2, modular multiplication, point multiplication.
Procedia PDF Downloads 987450 2.4 GHz 0.13µM Multi Biased Cascode Power Amplifier for ISM Band Wireless Applications
Authors: Udayan Patankar, Shashwati Bhagat, Vilas Nitneware, Ants Koel
Abstract:
An ISM band power amplifier is a type of electronic amplifier used to convert a low-power radio-frequency signal into a larger signal of significant power, typically used for driving the antenna of a transmitter. Due to drastic changes in telecommunication generations may lead to the requirements of improvements. Rapid changes in communication lead to the wide implementation of WLAN technology for its excellent characteristics, such as high transmission speed, long communication distance, and high reliability. Many applications such as WLAN, Bluetooth, and ZigBee, etc. were evolved with 2.4GHz to 5 GHz ISM Band, in which the power amplifier (PA) is a key building block of RF transmitters. There are many manufacturing processes available to manufacture a power amplifier for desired power output, but the major problem they have faced is about the power it consumed for its proper working, as many of them are fabricated on the GaN HEMT, Bi COMS process. In this paper we present a CMOS Base two stage cascode design of power amplifier working on 2.4GHz ISM frequency band. To lower the costs and allow full integration of a complete System-on-Chip (SoC) we have chosen 0.13µm low power CMOS technology for design. While designing a power amplifier, it is a real task to achieve higher power efficiency with minimum resources. This design showcase the Multi biased Cascode methodology to implement a two-stage CMOS power amplifier using ADS and LTSpice simulating tool. Main source is maximum of 2.4V which is internally distributed into different biasing point VB driving and VB driven as required for distinct stages of two stage RF power amplifier. It shows maximum power added efficiency near about 70.195% whereas its Power added efficiency calculated at 1 dB compression point is 44.669 %. Biased MOSFET is used to reduce total dc current as this circuit is designed for different wireless applications comes under 2.4GHz ISM Band.Keywords: RFIC, PAE, RF CMOS, impedance matching
Procedia PDF Downloads 2247449 Asymmetric Linkages Between Global Sustainable Index (Green Bond) and Cryptocurrency Markets with Portfolio Implications
Authors: Faheem Ur Rehman, Muhammad Khalil Khan, Miao Qing
Abstract:
This study investigated the asymmetric links and portfolio strategies between green bonds and the markets of three different cryptocurrencies, i.e., green, Islamic, and conventional, using data from January 1, 2018, to April 8, 2022, and employing asymmetric TVP-VAR model to quantify risk spillovers in the network analysis. In addition, we use the minimum variance, minimum correlation, and minimum connectedness methodologies to assess the portfolio implications. The results of the asymmetric dynamic connectedness index (TCI) model show that by adopting cryptocurrencies for digital finance, risk spillovers are found to be reduced. The findings of net directional connectedness demonstrate that during the study period, green bonds consistently get return spillovers from all other network variables. Positive return spillovers are bigger in magnitude than negative ones. These results imply that the influence of the green bond market on the cryptocurrency markets is decreasing. Positive return spillovers generate higher connectedness values for (HG, BNB, and TRX) coins and persistent net recipients in the specific network. On the other hand, Cardano and ADA coins are persistent net transmitters in the system. XLM and MIOTA's responsibilities shift over time, and there is evidence of asymmetry when both positive and negative returns are considered. According to the pairwise portfolio weights, BNB vs. BTC has the largest portfolio weights in the system, followed by BNB vs. Ethereum, suggesting the best investment strategies in the network.Keywords: asymmetric TVP-VAR, global sustainable index, cryptocurrency, portfolios
Procedia PDF Downloads 787448 Influence of the Molar Concentration and Substrate Temperature on Fluorine-Doped Zinc Oxide Thin Films Chemically Sprayed
Authors: J. Ramirez, A. Maldonado, M. de la L. Olvera
Abstract:
The effect of both the molar concentration of the starting solution and the substrate temperature on the electrical, morphological, structural and optical properties of chemically sprayed fluorine-doped zinc oxide (ZnO:F) thin films deposited on glass substrates, is analyzed in this work. All the starting solutions employed were aged for ten days before the deposition. The results show that as the molar concentration increases, a decrease in the electrical resistivity values is obtained, reaching the minimum in films deposited from a 0.4 M solution at 500°C. A further increase in the molar concentration leads to a very slight increase in the resistivity. On the other hand, as the substrate temperature is increased, the resistivity decreases and a tendency towards to minimum value is evidenced; taking the molar concentration as parameter, minimum values are reached at 500°C. The attain of ZnO:F thin films, with a resistivity as low as 7.8×10-3 Ώcm (sheet resistance of 130 Ώ/☐ and film thickness of 600 nm) measured in as-deposited films is reported here for the first time. The concurrent effect of the high molar concentration of the starting solution, the substrate temperature values used, and the ageing of the starting solution, which might cause polymerization of the zinc ions with the fluorine species, enhance the electrical properties. The structure of the films is polycrystalline, with a (002) preferential growth. Molar concentration rules the surface morphology as at low concentration an hexagonal and porous structure is developed changing to a uniform compact and small grain size surface in the films deposited with the high molar concentrations.Keywords: zinc oxide, chemical spray, thin films, TCO
Procedia PDF Downloads 5037447 Designing a Learning Table and Game Cards for Preschoolers for Disaster Risk Reduction (DRR) on Earthquake
Authors: Mehrnoosh Mirzaei
Abstract:
Children are among the most vulnerable at the occurrence of natural disasters such as earthquakes. Most of the management and measures which are considered for both before and during an earthquake are neither suitable nor efficient for this age group and cannot be applied. On the other hand, due to their age, it is hard to educate and train children to learn and understand the concept of earthquake risk mitigation as matters like earthquake prevention and safe places during an earthquake are not easily perceived. To our knowledge, children’s awareness of such concepts via their own world with the help of games is the best training method in this case. In this article, the researcher has tried to consider the child an active element before and during the earthquake. With training, provided by adults before the incidence of an earthquake, the child has the ability to learn disaster risk reduction (DRR). The focus of this research is on learning risk reduction behavior and regarding children as an individual element. The information of this article has been gathered from library resources, observations and the drawings of 10 children aged 5 whose subject was their conceptual definition of an earthquake who were asked to illustrate their conceptual definition of an earthquake; the results of 20 questionnaires filled in by preschoolers along with information gathered by interviewing them. The design of the suitable educational game, appropriate for the needs of this age group, has been made based on the theory of design with help of the user and the priority of children’s learning needs. The final result is a package of a game which is comprised of a learning table and matching cards showing sign marks for safe and unsafe places which introduce the safe behaviors and safe locations before and during the earthquake. These educational games can be used both in group contexts in kindergartens and on an individual basis at home, and they help in earthquake risk reduction.Keywords: disaster education, earthquake sign marks, learning table, matching card, risk reduction behavior
Procedia PDF Downloads 2577446 Application of Additive Manufacturing for Production of Optimum Topologies
Authors: Mahdi Mottahedi, Peter Zahn, Armin Lechler, Alexander Verl
Abstract:
Optimal topology of components leads to the maximum stiffness with the minimum material use. For the generation of these topologies, normally algorithms are employed, which tackle manufacturing limitations, at the cost of the optimal result. The global optimum result with penalty factor one, however, cannot be fabricated with conventional methods. In this article, an additive manufacturing method is introduced, in order to enable the production of global topology optimization results. For a benchmark, topology optimization with higher and lower penalty factors are performed. Different algorithms are employed in order to interpret the results of topology optimization with lower factors in many microstructure layers. These layers are then joined to form the final geometry. The algorithms’ benefits are then compared experimentally and numerically for the best interpretation. The findings demonstrate that by implementation of the selected algorithm, the stiffness of the components produced with this method is higher than what could have been produced by conventional techniques.Keywords: topology optimization, additive manufacturing, 3D-printer, laminated object manufacturing
Procedia PDF Downloads 3397445 Ordinal Regression with Fenton-Wilkinson Order Statistics: A Case Study of an Orienteering Race
Authors: Joonas Pääkkönen
Abstract:
In sports, individuals and teams are typically interested in final rankings. Final results, such as times or distances, dictate these rankings, also known as places. Places can be further associated with ordered random variables, commonly referred to as order statistics. In this work, we introduce a simple, yet accurate order statistical ordinal regression function that predicts relay race places with changeover-times. We call this function the Fenton-Wilkinson Order Statistics model. This model is built on the following educated assumption: individual leg-times follow log-normal distributions. Moreover, our key idea is to utilize Fenton-Wilkinson approximations of changeover-times alongside an estimator for the total number of teams as in the notorious German tank problem. This original place regression function is sigmoidal and thus correctly predicts the existence of a small number of elite teams that significantly outperform the rest of the teams. Our model also describes how place increases linearly with changeover-time at the inflection point of the log-normal distribution function. With real-world data from Jukola 2019, a massive orienteering relay race, the model is shown to be highly accurate even when the size of the training set is only 5% of the whole data set. Numerical results also show that our model exhibits smaller place prediction root-mean-square-errors than linear regression, mord regression and Gaussian process regression.Keywords: Fenton-Wilkinson approximation, German tank problem, log-normal distribution, order statistics, ordinal regression, orienteering, sports analytics, sports modeling
Procedia PDF Downloads 1247444 Low Density Parity Check Codes
Authors: Kassoul Ilyes
Abstract:
The field of error correcting codes has been revolutionized by the introduction of iteratively decoded codes. Among these, LDPC codes are now a preferred solution thanks to their remarkable performance and low complexity. The binary version of LDPC codes showed even better performance, although it’s decoding introduced greater complexity. This thesis studies the performance of binary LDPC codes using simplified weighted decisions. Information is transported between a transmitter and a receiver by digital transmission systems, either by propagating over a radio channel or also by using a transmission medium such as the transmission line. The purpose of the transmission system is then to carry the information from the transmitter to the receiver as reliably as possible. These codes have not generated enough interest within the coding theory community. This forgetfulness will last until the introduction of Turbo-codes and the iterative principle. Then it was proposed to adopt Pearl's Belief Propagation (BP) algorithm for decoding these codes. Subsequently, Luby introduced irregular LDPC codes characterized by a parity check matrix. And finally, we study simplifications on binary LDPC codes. Thus, we propose a method to make the exact calculation of the APP simpler. This method leads to simplifying the implementation of the system.Keywords: LDPC, parity check matrix, 5G, BER, SNR
Procedia PDF Downloads 1547443 Wireless FPGA-Based Motion Controller Design by Implementing 3-Axis Linear Trajectory
Authors: Kiana Zeighami, Morteza Ozlati Moghadam
Abstract:
Designing a high accuracy and high precision motion controller is one of the important issues in today’s industry. There are effective solutions available in the industry but the real-time performance, smoothness and accuracy of the movement can be further improved. This paper discusses a complete solution to carry out the movement of three stepper motors in three dimensions. The objective is to provide a method to design a fully integrated System-on-Chip (SOC)-based motion controller to reduce the cost and complexity of production by incorporating Field Programmable Gate Array (FPGA) into the design. In the proposed method the FPGA receives its commands from a host computer via wireless internet communication and calculates the motion trajectory for three axes. A profile generator module is designed to realize the interpolation algorithm by translating the position data to the real-time pulses. This paper discusses an approach to implement the linear interpolation algorithm, since it is one of the fundamentals of robots’ movements and it is highly applicable in motion control industries. Along with full profile trajectory, the triangular drive is implemented to eliminate the existence of error at small distances. To integrate the parallelism and real-time performance of FPGA with the power of Central Processing Unit (CPU) in executing complex and sequential algorithms, the NIOS II soft-core processor was added into the design. This paper presents different operating modes such as absolute, relative positioning, reset and velocity modes to fulfill the user requirements. The proposed approach was evaluated by designing a custom-made FPGA board along with a mechanical structure. As a result, a precise and smooth movement of stepper motors was observed which proved the effectiveness of this approach.Keywords: 3-axis linear interpolation, FPGA, motion controller, micro-stepping
Procedia PDF Downloads 2087442 Employing Visual Culture to Enhance Initial Adult Maltese Language Acquisition
Authors: Jacqueline Żammit
Abstract:
Recent research indicates that the utilization of right-brain strategies holds significant implications for the acquisition of language skills. Nevertheless, the utilization of visual culture as a means to stimulate these strategies and amplify language retention among adults engaging in second language (L2) learning remains a relatively unexplored area. This investigation delves into the impact of visual culture on activating right-brain processes during the initial stages of language acquisition, particularly in the context of teaching Maltese as a second language (ML2) to adult learners. By employing a qualitative research approach, this study convenes a focus group comprising twenty-seven educators to delve into a range of visual culture techniques integrated within language instruction. The collected data is subjected to thematic analysis using NVivo software. The findings underscore a variety of impactful visual culture techniques, encompassing activities such as drawing, sketching, interactive matching games, orthographic mapping, memory palace strategies, wordless picture books, picture-centered learning methodologies, infographics, Face Memory Game, Spot the Difference, Word Search Puzzles, the Hidden Object Game, educational videos, the Shadow Matching technique, Find the Differences exercises, and color-coded methodologies. These identified techniques hold potential for application within ML2 classes for adult learners. Consequently, this study not only provides insights into optimizing language learning through specific visual culture strategies but also furnishes practical recommendations for enhancing language competencies and skills.Keywords: visual culture, right-brain strategies, second language acquisition, maltese as a second language, visual aids, language-based activities
Procedia PDF Downloads 617441 Curvature Based-Methods for Automatic Coarse and Fine Registration in Dimensional Metrology
Authors: Rindra Rantoson, Hichem Nouira, Nabil Anwer, Charyar Mehdi-Souzani
Abstract:
Multiple measurements by means of various data acquisition systems are generally required to measure the shape of freeform workpieces for accuracy, reliability and holisticity. The obtained data are aligned and fused into a common coordinate system within a registration technique involving coarse and fine registrations. Standardized iterative methods have been established for fine registration such as Iterative Closest Points (ICP) and its variants. For coarse registration, no conventional method has been adopted yet despite a significant number of techniques which have been developed in the literature to supply an automatic rough matching between data sets. Two main issues are addressed in this paper: the coarse registration and the fine registration. For coarse registration, two novel automated methods based on the exploitation of discrete curvatures are presented: an enhanced Hough Transformation (HT) and an improved Ransac Transformation. The use of curvature features in both methods aims to reduce computational cost. For fine registration, a new variant of ICP method is proposed in order to reduce registration error using curvature parameters. A specific distance considering the curvature similarity has been combined with Euclidean distance to define the distance criterion used for correspondences searching. Additionally, the objective function has been improved by combining the point-to-point (P-P) minimization and the point-to-plane (P-Pl) minimization with automatic weights. These ones are determined from the preliminary calculated curvature features at each point of the workpiece surface. The algorithms are applied on simulated and real data performed by a computer tomography (CT) system. The obtained results reveal the benefit of the proposed novel curvature-based registration methods.Keywords: discrete curvature, RANSAC transformation, hough transformation, coarse registration, ICP variant, point-to-point and point-to-plane minimization combination, computer tomography
Procedia PDF Downloads 424