Search results for: equation error
3330 Comparison of the Distillation Curve Obtained Experimentally with the Curve Extrapolated by a Commercial Simulator
Authors: Lívia B. Meirelles, Erika C. A. N. Chrisman, Flávia B. de Andrade, Lilian C. M. de Oliveira
Abstract:
True Boiling Point distillation (TBP) is one of the most common experimental techniques for the determination of petroleum properties. This curve provides information about the performance of petroleum in terms of its cuts. The experiment is performed in a few days. Techniques are used to determine the properties faster with a software that calculates the distillation curve when a little information about crude oil is known. In order to evaluate the accuracy of distillation curve prediction, eight points of the TBP curve and specific gravity curve (348 K and 523 K) were inserted into the HYSYS Oil Manager, and the extended curve was evaluated up to 748 K. The methods were able to predict the curve with the accuracy of 0.6%-9.2% error (Software X ASTM), 0.2%-5.1% error (Software X Spaltrohr).Keywords: distillation curve, petroleum distillation, simulation, true boiling point curve
Procedia PDF Downloads 4413329 A Hyperexponential Approximation to Finite-Time and Infinite-Time Ruin Probabilities of Compound Poisson Processes
Authors: Amir T. Payandeh Najafabadi
Abstract:
This article considers the problem of evaluating infinite-time (or finite-time) ruin probability under a given compound Poisson surplus process by approximating the claim size distribution by a finite mixture exponential, say Hyperexponential, distribution. It restates the infinite-time (or finite-time) ruin probability as a solvable ordinary differential equation (or a partial differential equation). Application of our findings has been given through a simulation study.Keywords: ruin probability, compound poisson processes, mixture exponential (hyperexponential) distribution, heavy-tailed distributions
Procedia PDF Downloads 3413328 Theoretical Analysis of the Solid State and Optical Characteristics of Calcium Sulpide Thin Film
Authors: Emmanuel Ifeanyi Ugwu
Abstract:
Calcium Sulphide which is one of Chalcogenide group of thin films has been analyzed in this work using a theoretical approach in which a scalar wave was propagated through the material thin film medium deposited on a glass substrate with the assumption that the dielectric medium has homogenous reference dielectric constant term, and a perturbed dielectric function, representing the deposited thin film medium on the surface of the glass substrate as represented in this work. These were substituted into a defined scalar wave equation that was solved first of all by transforming it into Volterra equation of second type and solved using the method of separation of variable on scalar wave and subsequently, Green’s function technique was introduced to obtain a model equation of wave propagating through the thin film that was invariably used in computing the propagated field, for different input wavelengths representing UV, Visible and Near-infrared regions of field considering the influence of the dielectric constants of the thin film on the propagating field. The results obtained were used in turn to compute the band gaps, solid state and optical properties of the thin film.Keywords: scalar wave, dielectric constant, calcium sulphide, solid state, optical properties
Procedia PDF Downloads 1183327 Mathematical and Numerical Analysis of a Reaction Diffusion System of Lambda-Omega Type
Authors: Hassan Al Salman, Ahmed Al Ghafli
Abstract:
In this study we consider a nonlinear in time finite element approximation of a reaction diffusion system of lambda-omega type. We use a fixed point theorem to prove existence of the approximations. Then, we derive some essential stability estimates and discuss the uniqueness of the approximations. Also, we prove an optimal error bound in time for d=1, 2 and 3 space dimensions. Finally, we present some numerical experiments to verify the theoretical results.Keywords: reaction diffusion system, finite element approximation, fixed point theorem, an optimal error bound
Procedia PDF Downloads 5333326 Effect of pH-Dependent Surface Charge on the Electroosmotic Flow through Nanochannel
Authors: Partha P. Gopmandal, Somnath Bhattacharyya, Naren Bag
Abstract:
In this article, we have studied the effect of pH-regulated surface charge on the electroosmotic flow (EOF) through nanochannel filled with binary symmetric electrolyte solution. The channel wall possesses either an acidic or a basic functional group. Going beyond the widely employed Debye-Huckel linearization, we develop a mathematical model based on Nernst-Planck equation for the charged species, Poisson equation for the induced potential, Stokes equation for fluid flow. A finite volume based numerical algorithm is adopted to study the effect of key parameters on the EOF. We have computed the coupled governing equations through the finite volume method and our results found to be in good agreement with the analytical solution obtained from the corresponding linear model based on low surface charge condition or strong electrolyte solution. The influence of the surface charge density, reaction constant of the functional groups, bulk pH, and concentration of the electrolyte solution on the overall flow rate is studied extensively. We find the effect of surface charge diminishes with the increase in electrolyte concentration. In addition for strong electrolyte, the surface charge becomes independent of pH due to complete dissociation of the functional groups.Keywords: electroosmosis, finite volume method, functional group, surface charge
Procedia PDF Downloads 4193325 Aggregate Supply Response of Some Livestock Commodities in Algeria: Cointegration- Vector Error Correction Model Approach
Authors: Amine M. Benmehaia, Amine Oulmane
Abstract:
The supply response of agricultural commodities to changes in price incentives is an important issue for the success of any policy reform in the agricultural sector. This study aims to quantify the responsiveness of producers of some livestock commodities to price incentives in Algerian context. Time series analysis is used on annual data for a period of 52 years (1966-2018). Both co-integration and vector error correction model (VECM) are used through the Nerlove model of partial adjustment. The study attempts to determine the long-run and short-run relationships along with the magnitudes of disequilibria in the selected commodities. Results show that the short-run price elasticities are low in cow and sheep meat sectors (8.7 and 8% respectively), while their respective long-run elasticities are 16.5 and 10.5, whereas eggs and milk have very high short-run price elasticities (82 and 90% respectively) with long-run elasticities of 40 and 46 respectively. The error correction coefficient, reflecting the speed of adjustment towards the long-run equilibrium, is statistically significant and have the expected negative sign. Its estimates are 12.7 for cow meat, 33.5 for sheep meat, 46.7 for eggs and 8.4 for milk. It seems that cow meat and milk producers have a weak feedback of about 12.7% and 8.4% respectively of the previous year's disequilibrium from the long-run price elasticity, whereas sheep meat and eggs producers adjust to correct long run disequilibrium with a high speed of adjustment (33.5% and 46.7 % respectively). The implication of this is that much more in-depth research is needed to identify those factors that affect agricultural supply and to describe the effect of factors that shift supply in response to price incentives. This could provide valuable information for government in the use of appropriate policy measures.Keywords: Algeria, cointegration, livestock, supply response, vector error correction model
Procedia PDF Downloads 1413324 Continuous Differential Evolution Based Parameter Estimation Framework for Signal Models
Authors: Ammara Mehmood, Aneela Zameer, Muhammad Asif Zahoor Raja, Muhammad Faisal Fateh
Abstract:
In this work, the strength of bio-inspired computational intelligence based technique is exploited for parameter estimation for the periodic signals using Continuous Differential Evolution (CDE) by defining an error function in the mean square sense. Multidimensional and nonlinear nature of the problem emerging in sinusoidal signal models along with noise makes it a challenging optimization task, which is dealt with robustness and effectiveness of CDE to ensure convergence and avoid trapping in local minima. In the proposed scheme of Continuous Differential Evolution based Signal Parameter Estimation (CDESPE), unknown adjustable weights of the signal system identification model are optimized utilizing CDE algorithm. The performance of CDESPE model is validated through statistics based various performance indices on a sufficiently large number of runs in terms of estimation error, mean squared error and Thiel’s inequality coefficient. Efficacy of CDESPE is examined by comparison with the actual parameters of the system, Genetic Algorithm based outcomes and from various deterministic approaches at different signal-to-noise ratio (SNR) levels.Keywords: parameter estimation, bio-inspired computing, continuous differential evolution (CDE), periodic signals
Procedia PDF Downloads 3023323 Mixed Number Algebra and Its Application
Authors: Md. Shah Alam
Abstract:
Mushfiq Ahmad has defined a Mixed Number, which is the sum of a scalar and a Cartesian vector. He has also defined the elementary group operations of Mixed numbers i.e. the norm of Mixed numbers, the product of two Mixed numbers, the identity element and the inverse. It has been observed that Mixed Number is consistent with Pauli matrix algebra and a handy tool to work with Dirac electron theory. Its use as a mathematical method in Physics has been studied. (1) We have applied Mixed number in Quantum Mechanics: Mixed Number version of Displacement operator, Vector differential operator, and Angular momentum operator has been developed. Mixed Number method has also been applied to Klein-Gordon equation. (2) We have applied Mixed number in Electrodynamics: Mixed Number version of Maxwell’s equation, the Electric and Magnetic field quantities and Lorentz Force has been found. (3) An associative transformation of Mixed Number numbers fulfilling Lorentz invariance requirement is developed. (4) We have applied Mixed number algebra as an extension of Complex number. Mixed numbers and the Quaternions have isomorphic correspondence, but they are different in algebraic details. The multiplication of unit Mixed number and the multiplication of unit Quaternions are different. Since Mixed Number has properties similar to those of Pauli matrix algebra, Mixed Number algebra is a more convenient tool to deal with Dirac equation.Keywords: mixed number, special relativity, quantum mechanics, electrodynamics, pauli matrix
Procedia PDF Downloads 3643322 Cellular Traffic Prediction through Multi-Layer Hybrid Network
Authors: Supriya H. S., Chandrakala B. M.
Abstract:
Deep learning based models have been recently successful adoption for network traffic prediction. However, training a deep learning model for various prediction tasks is considered one of the critical tasks due to various reasons. This research work develops Multi-Layer Hybrid Network (MLHN) for network traffic prediction and analysis; MLHN comprises the three distinctive networks for handling the different inputs for custom feature extraction. Furthermore, an optimized and efficient parameter-tuning algorithm is introduced to enhance parameter learning. MLHN is evaluated considering the “Big Data Challenge” dataset considering the Mean Absolute Error, Root Mean Square Error and R^2as metrics; furthermore, MLHN efficiency is proved through comparison with a state-of-art approach.Keywords: MLHN, network traffic prediction
Procedia PDF Downloads 883321 Profitability Assessment of Granite Aggregate Production and the Development of a Profit Assessment Model
Authors: Melodi Mbuyi Mata, Blessing Olamide Taiwo, Afolabi Ayodele David
Abstract:
The purpose of this research is to create empirical models for assessing the profitability of granite aggregate production in Akure, Ondo state aggregate quarries. In addition, an artificial neural network (ANN) model and multivariate predicting models for granite profitability were developed in the study. A formal survey questionnaire was used to collect data for the study. The data extracted from the case study mine for this study includes granite marketing operations, royalty, production costs, and mine production information. The following methods were used to achieve the goal of this study: descriptive statistics, MATLAB 2017, and SPSS16.0 software in analyzing and modeling the data collected from granite traders in the study areas. The ANN and Multi Variant Regression models' prediction accuracy was compared using a coefficient of determination (R²), Root mean square error (RMSE), and mean square error (MSE). Due to the high prediction error, the model evaluation indices revealed that the ANN model was suitable for predicting generated profit in a typical quarry. More quarries in Nigeria's southwest region and other geopolitical zones should be considered to improve ANN prediction accuracy.Keywords: national development, granite, profitability assessment, ANN models
Procedia PDF Downloads 1013320 Improved Performance Scheme for Joint Transmission in Downlink Coordinated Multi-Point Transmission
Authors: Young-Su Ryu, Su-Hyun Jung, Myoung-Jin Kim, Hyoung-Kyu Song
Abstract:
In this paper, improved performance scheme for joint transmission is proposed in downlink (DL) coordinated multi-point(CoMP) in case of constraint transmission power. This scheme is that serving transmission point (TP) request a joint transmission to inter-TP and selects one pre-coding technique according to channel state information(CSI) from user equipment(UE). The simulation results show that the bit error rate(BER) and throughput performances of the proposed scheme provide high spectral efficiency and reliable data at the cell edge.Keywords: CoMP, joint transmission, minimum mean square error, zero-forcing, zero-forcing dirty paper coding
Procedia PDF Downloads 5533319 Matching Law in Autoshaped Choice in Neural Networks
Authors: Giselle Maggie Fer Castañeda, Diego Iván González
Abstract:
The objective of this work was to study the autoshaped choice behavior in the Donahoe, Burgos and Palmer (DBP) neural network model and analyze it under the matching law. Autoshaped choice can be viewed as a form of economic behavior defined as the preference between alternatives according to their relative outcomes. The Donahoe, Burgos and Palmer (DBP) model is a connectionist proposal that unifies operant and Pavlovian conditioning. This model has been used for more than three decades as a neurobehavioral explanation of conditioning phenomena, as well as a generator of predictions suitable for experimental testing with non-human animals and humans. The study consisted of different simulations in which, in each one, a ratio of reinforcement was established for two alternatives, and the responses (i.e., activations) in each of them were measured. Choice studies with animals have demonstrated that the data generally conform closely to the generalized matching law equation, which states that the response ratio equals proportionally to the reinforcement ratio; therefore, it was expected to find similar results with the neural networks of the Donahoe, Burgos and Palmer (DBP) model since these networks have simulated and predicted various conditioning phenomena. The results were analyzed by the generalized matching law equation, and it was observed that under some contingencies, the data from the networks adjusted approximately to what was established by the equation. Implications and limitations are discussed.Keywords: matching law, neural networks, computational models, behavioral sciences
Procedia PDF Downloads 743318 Solving Momentum and Energy Equation by Using Differential Transform Techniques
Authors: Mustafa Ekici
Abstract:
Natural convection is a basic process which is important in a wide variety of practical applications. In essence, a heated fluid expands and rises from buoyancy due to decreased density. Numerous papers have been written on natural or mixed convection in vertical ducts heated on the side. These equations have been proved to be valuable tools for the modelling of many phenomena such as fluid dynamics. Finding solutions to such equations or system of equations are in general not an easy task. We propose a method, which is called differential transform method, of solving a non-linear equations and compare the results with some of the other techniques. Illustrative examples shows that the results are in good agreement.Keywords: differential transform method, momentum, energy equation, boundry value problem
Procedia PDF Downloads 4613317 Identification of Architectural Design Error Risk Factors in Construction Projects Using IDEF0 Technique
Authors: Sahar Tabarroki, Ahad Nazari
Abstract:
The design process is one of the most key project processes in the construction industry. Although architects have the responsibility to produce complete, accurate, and coordinated documents, architectural design is accompanied by many errors. A design error occurs when the constraints and requirements of the design are not satisfied. Errors are potentially costly and time-consuming to correct if not caught early during the design phase, and they become expensive in either construction documents or in the construction phase. The aim of this research is to identify the risk factors of architectural design errors, so identification of risks is necessary. First, a literature review in the design process was conducted and then a questionnaire was designed to identify the risks and risk factors. The questions in the form of the questionnaire were based on the “similar service description of study and supervision of architectural works” published by “Vice Presidency of Strategic Planning & Supervision of I.R. Iran” as the base of architects’ tasks. Second, the top 10 risks of architectural activities were identified. To determine the positions of possible causes of risks with respect to architectural activities, these activities were located in a design process modeled by the IDEF0 technique. The research was carried out by choosing a case study, checking the design drawings, interviewing its architect and client, and providing a checklist in order to identify the concrete examples of architectural design errors. The results revealed that activities such as “defining the current and future requirements of the project”, “studies and space planning,” and “time and cost estimation of suggested solution” has a higher error risk than others. Moreover, the most important causes include “unclear goals of a client”, “time force by a client”, and “lack of knowledge of architects about the requirements of end-users”. For error detecting in the case study, lack of criteria, standards and design criteria, and lack of coordination among them, was a barrier, anyway, “lack of coordination between architectural design and electrical and mechanical facility”, “violation of the standard dimensions and sizes in space designing”, “design omissions” were identified as the most important design errors.Keywords: architectural design, design error, risk management, risk factor
Procedia PDF Downloads 1303316 A Dynamic Symplectic Manifold Analysis for Wave Propagation in Porous Media
Authors: K. I. M. Guerra, L. A. P. Silva, J. C. Leal
Abstract:
This study aims to understand with more amplitude and clarity the behavior of a porous medium where a pressure wave travels, translated into relative displacements inside the material, using mathematical tools derived from topology and symplectic geometry. The paper starts with a given partial differential equation based on the continuity and conservation theorems to describe the traveling wave through the porous body. A solution for this equation is proposed after all boundary, and initial conditions are fixed, and it’s accepted that the solution lies in a manifold U of purely spatial dimensions and that is embedded in the Real n-dimensional manifold, with spatial and kinetic dimensions. It’s shown that the U manifold of lower dimensions than IRna, where it is embedded, inherits properties of the vector spaces existing inside the topology it lies on. Then, a second manifold (U*), embedded in another space called IRnb of stress dimensions, is proposed and there’s a non-degenerative function that maps it into the U manifold. This relation is proved as a transformation in between two corresponding admissible solutions of the differential equation in distinct dimensions and properties, leading to a more visual and intuitive understanding of the whole dynamic process of a stress wave through a porous medium and also highlighting the dimensional invariance of Terzaghi’s theory for any coordinate system.Keywords: poremechanics, soil dynamics, symplectic geometry, wave propagation
Procedia PDF Downloads 2963315 Feature Location Restoration for Under-Sampled Photoplethysmogram Using Spline Interpolation
Authors: Hangsik Shin
Abstract:
The purpose of this research is to restore the feature location of under-sampled photoplethysmogram using spline interpolation and to investigate feasibility for feature shape restoration. We obtained 10 kHz-sampled photoplethysmogram and decimated it to generate under-sampled dataset. Decimated dataset has 5 kHz, 2.5 k Hz, 1 kHz, 500 Hz, 250 Hz, 25 Hz and 10 Hz sampling frequency. To investigate the restoration performance, we interpolated under-sampled signals with 10 kHz, then compared feature locations with feature locations of 10 kHz sampled photoplethysmogram. Features were upper and lower peak of photplethysmography waveform. Result showed that time differences were dramatically decreased by interpolation. Location error was lesser than 1 ms in both feature types. In 10 Hz sampled cases, location error was also deceased a lot, however, they were still over 10 ms.Keywords: peak detection, photoplethysmography, sampling, signal reconstruction
Procedia PDF Downloads 3683314 Estimation of Longitudinal Dispersion Coefficient Using Tracer Data
Authors: K. Ebrahimi, Sh. Shahid, M. Mohammadi Ghaleni, M. H. Omid
Abstract:
The longitudinal dispersion coefficient is a crucial parameter for 1-D water quality analysis of riverine flows. So far, different types of empirical equations for estimation of the coefficient have been developed, based on various case studies. The main objective of this paper is to develop an empirical equation for estimation of the coefficient for a riverine flow. For this purpose, a set of tracer experiments was conducted, involving salt tracer, at three sections located in downstream of a lengthy canal. Tracer data were measured in three mixing lengths along the canal including; 45, 75 and 100m. According to the results, the obtained coefficients from new developed empirical equation gave an encouraging level of agreement with the theoretical values.Keywords: coefficients, dispersion, river, tracer, water quality
Procedia PDF Downloads 3893313 Globally Attractive Mild Solutions for Non-Local in Time Subdiffusion Equations of Neutral Type
Authors: Jorge Gonzalez Camus, Carlos Lizama
Abstract:
In this work is proved the existence of at least one globally attractive mild solution to the Cauchy problem, for fractional evolution equation of neutral type, involving the fractional derivate in Caputo sense. An almost sectorial operator on a Banach space X and a kernel belonging to a large class appears in the equation, which covers many relevant cases from physics applications, in particular, the important case of time - fractional evolution equations of neutral type. The main tool used in this work was the Hausdorff measure of noncompactness and fixed point theorems, specifically Darbo-type. Initially, the equation is a Cauchy problem, involving a fractional derivate in Caputo sense. Then, is formulated the equivalent integral version, and defining a convenient functional, using the analytic integral resolvent operator, and verifying the hypothesis of the fixed point theorem of Darbo type, give us the existence of mild solution for the initial problem. Furthermore, each mild solution is globally attractive, a property that is desired in asymptotic behavior for that solution.Keywords: attractive mild solutions, integral Volterra equations, neutral type equations, non-local in time equations
Procedia PDF Downloads 1583312 Precipitation Intensity: Duration Based Threshold Analysis for Initiation of Landslides in Upper Alaknanda Valley
Authors: Soumiya Bhattacharjee, P. K. Champati Ray, Shovan L. Chattoraj, Mrinmoy Dhara
Abstract:
The entire Himalayan range is globally renowned for rainfall-induced landslides. The prime focus of the study is to determine rainfall based threshold for initiation of landslides that can be used as an important component of an early warning system for alerting stake holders. This research deals with temporal dimension of slope failures due to extreme rainfall events along the National Highway-58 from Karanprayag to Badrinath in the Garhwal Himalaya, India. Post processed 3-hourly rainfall intensity data and its corresponding duration from daily rainfall data available from Tropical Rainfall Measuring Mission (TRMM) were used as the prime source of rainfall data. Landslide event records from Border Road Organization (BRO) and some ancillary landslide inventory data for 2013 and 2014 have been used to determine Intensity Duration (ID) based rainfall threshold. The derived governing threshold equation, I= 4.738D-0.025, has been considered for prediction of landslides of the study region. This equation was validated with an accuracy of 70% landslides during August and September 2014. The derived equation was considered for further prediction of landslides of the study region. From the obtained results and validation, it can be inferred that this equation can be used for initiation of landslides in the study area to work as a part of an early warning system. Results can significantly improve with ground based rainfall estimates and better database on landslide records. Thus, the study has demonstrated a very low cost method to get first-hand information on possibility of impending landslide in any region, thereby providing alert and better preparedness for landslide disaster mitigation.Keywords: landslide, intensity-duration, rainfall threshold, TRMM, slope, inventory, early warning system
Procedia PDF Downloads 2733311 InfoMiracles in the Qur’an and a Mathematical Proof to the Existence of God
Authors: Mohammad Mahmoud Mandurah
Abstract:
The existence of InfoMiracles in scripture is evidence that the scripture has a divine origin. It is also evidence to the existence of God. An InfoMiracle is an information-based miracle. The basic component of an InfoMiracle is a piece of information that could not be obtained by a human except through a divine channel. The existence of a sufficient number of convincing InfoMiracles in a scripture necessitates the existence of the divine source to these InfoMiracles. A mathematical equation is developed to prove that the Qur’an has a divine origin, and hence, prove the existence of God. The equation depends on a single variable only, which is the number of InfoMiracles in the Qur’an. The Qur’an is rich with InfoMiracles. It is shown that the existence of less than 30 InfoMiracles in the Qur’an is sufficient proof to the existence of God and that the Qur’an is a revelation from God.Keywords: InfoMiracle, God, mathematical proof, miracle, probability
Procedia PDF Downloads 2173310 Maximum Initial Input Allowed to Iterative Learning Control Set-up Using Singular Values
Authors: Naser Alajmi, Ali Alobaidly, Mubarak Alhajri, Salem Salamah, Muhammad Alsubaie
Abstract:
Iterative Learning Control (ILC) known to be a controlling tool to overcome periodic disturbances for repetitive systems. This technique is required to let the error signal tends to zero as the number of operation increases. The learning process that lies within this context is strongly dependent on the initial input which if selected properly tends to let the learning process be more effective compared to the case where a system starts from blind. ILC uses previous recorded execution data to update the following execution/trial input such that a reference trajectory is followed to a high accuracy. Error convergence in ILC is generally highly dependent on the input applied to a plant for trial $1$, thus a good choice of initial starting input signal would make learning faster and as a consequence the error tends to zero faster as well. In the work presented within, an upper limit based on the Singular Values Principle (SV) is derived for the initial input signal applied at trial $1$ such that the system follow the reference in less number of trials without responding aggressively or exceeding the working envelope where a system is required to move within in a robot arm, for example. Simulation results presented illustrate the theory introduced within this paper.Keywords: initial input, iterative learning control, maximum input, singular values
Procedia PDF Downloads 2413309 Lamb Waves Propagation in Elastic-Viscoelastic Three-Layer Adhesive Joints
Authors: Pezhman Taghipour Birgani, Mehdi Shekarzadeh
Abstract:
In this paper, the propagation of lamb waves in three-layer joints is investigated using global matrix method. Theoretical boundary value problem in three-layer adhesive joints with perfect bond and traction free boundary conditions on their outer surfaces is solved to find a combination of frequencies and modes with the lowest attenuation. The characteristic equation is derived by applying continuity and boundary conditions in three-layer joints using global matrix method. Attenuation and phase velocity dispersion curves are obtained with numerical solution of this equation by a computer code for a three-layer joint, including an aluminum repair patch bonded to the aircraft aluminum skin by a layer of viscoelastic epoxy adhesive. To validate the numerical solution results of the characteristic equation, wave structure curves are plotted for a special mode in two different frequencies in the adhesive joint. The purpose of present paper is to find a combination of frequencies and modes with minimum attenuation in high and low frequencies. These frequencies and modes are recognizable by transducers in inspections with Lamb waves because of low attenuation level.Keywords: three-layer adhesive joints, viscoelastic, lamb waves, global matrix method
Procedia PDF Downloads 3933308 The Non-Existence of Perfect 2-Error Correcting Lee Codes of Word Length 7 over Z
Authors: Catarina Cruz, Ana Breda
Abstract:
Tiling problems have been capturing the attention of many mathematicians due to their real-life applications. In this study, we deal with tilings of Zⁿ by Lee spheres, where n is a positive integer number, being these tilings related with error correcting codes on the transmission of information over a noisy channel. We focus our attention on the question ‘for what values of n and r does the n-dimensional Lee sphere of radius r tile Zⁿ?’. It seems that the n-dimensional Lee sphere of radius r does not tile Zⁿ for n ≥ 3 and r ≥ 2. Here, we prove that is not possible to tile Z⁷ with Lee spheres of radius 2 presenting a proof based on a combinatorial method and faithful to the geometric idea of the problem. The non-existence of such tilings has been studied by several authors being considered the most difficult cases those in which the radius of the Lee spheres is equal to 2. The relation between these tilings and error correcting codes is established considering the center of a Lee sphere as a codeword and the other elements of the sphere as words which are decoded by the central codeword. When the Lee spheres of radius r centered at elements of a set M ⊂ Zⁿ tile Zⁿ, M is a perfect r-error correcting Lee code of word length n over Z, denoted by PL(n, r). Our strategy to prove the non-existence of PL(7, 2) codes are based on the assumption of the existence of such code M. Without loss of generality, we suppose that O ∈ M, where O = (0, ..., 0). In this sense and taking into account that we are dealing with Lee spheres of radius 2, O covers all words which are distant two or fewer units from it. By the definition of PL(7, 2) code, each word which is distant three units from O must be covered by a unique codeword of M. These words have to be covered by codewords which dist five units from O. We prove the non-existence of PL(7, 2) codes showing that it is not possible to cover all the referred words without superposition of Lee spheres whose centers are distant five units from O, contradicting the definition of PL(7, 2) code. We achieve this contradiction by combining the cardinality of particular subsets of codewords which are distant five units from O. There exists an extensive literature on codes in the Lee metric. Here, we present a new approach to prove the non-existence of PL(7, 2) codes.Keywords: Golomb-Welch conjecture, Lee metric, perfect Lee codes, tilings
Procedia PDF Downloads 1603307 Assessment of Time-variant Work Stress for Human Error Prevention
Authors: Hyeon-Kyo Lim, Tong-Il Jang, Yong-Hee Lee
Abstract:
For an operator in a nuclear power plant, human error is one of the most dreaded factors that may result in unexpected accidents. The possibility of human errors may be low, but the risk of them would be unimaginably enormous. Thus, for accident prevention, it is quite indispensable to analyze the influence of any factors which may raise the possibility of human errors. During the past decades, not a few research results showed that performance of human operators may vary over time due to lots of factors. Among them, stress is known to be an indirect factor that may cause human errors and result in mental illness. Until now, not a few assessment tools have been developed to assess stress level of human workers. However, it still is questionable to utilize them for human performance anticipation which is related with human error possibility, because they were mainly developed from the viewpoint of mental health rather than industrial safety. Stress level of a person may go up or down with work time. In that sense, if they would be applicable in the safety aspect, they should be able to assess the variation resulted from work time at least. Therefore, this study aimed to compare their applicability for safety purpose. More than 10 kinds of work stress tools were analyzed with reference to assessment items, assessment and analysis methods, and follow-up measures which are known to close related factors with work stress. The results showed that most tools mainly focused their weights on some common organizational factors such as demands, supports, and relationships, in sequence. Their weights were broadly similar. However, they failed to recommend practical solutions. Instead, they merely advised to set up overall counterplans in PDCA cycle or risk management activities which would be far from practical human error prevention. Thus, it was concluded that application of stress assessment tools mainly developed for mental health seemed to be impractical for safety purpose with respect to human performance anticipation, and that development of a new assessment tools would be inevitable if anyone wants to assess stress level in the aspect of human performance variation and accident prevention. As a consequence, as practical counterplans, this study proposed a new scheme for assessment of work stress level of a human operator that may vary over work time which is closely related with the possibility of human errors.Keywords: human error, human performance, work stress, assessment tool, time-variant, accident prevention
Procedia PDF Downloads 6723306 Banking Sector Development and Economic Growth: Evidence from the State of Qatar
Authors: Fekri Shawtari
Abstract:
The banking sector plays a very crucial role in the economic development of the country. As a financial intermediary, it has assigned a great role in the economic growth and stability. This paper aims to examine the empirically the relationship between banking industry and economic growth in state of Qatar. We adopt the VAR vector error correction model (VECM) along with Granger causality to address the issue over the long-run and short-run between the banking sector and economic growth. It is expected that the results will give policy directions to the policymakers to make strategies that are conducive toward boosting development to achieve the targeted economic growth in current situation.Keywords: economic growth, banking sector, Qatar, vector error correction model, VECM
Procedia PDF Downloads 1703305 Nonhomogeneous Linear Second Order Differential Equations and Resonance through Geogebra Program
Authors: F. Maass, P. Martin, J. Olivares
Abstract:
The aim of this work is the application of the program GeoGebra in teaching the study of nonhomogeneous linear second order differential equations with constant coefficients. Different kind of functions or forces will be considered in the right hand side of the differential equations, in particular, the emphasis will be placed in the case of trigonometrical functions producing the resonance phenomena. In order to obtain this, the frequencies of the trigonometrical functions will be changed. Once the resonances appear, these have to be correlationated with the roots of the second order algebraic equation determined by the coefficients of the differential equation. In this way, the physics and engineering students will understand resonance effects and its consequences in the simplest way. A large variety of examples will be shown, using different kind of functions for the nonhomogeneous part of the differential equations.Keywords: education, geogebra, ordinary differential equations, resonance
Procedia PDF Downloads 2453304 The Value of Store Choice Criteria on Perceived Patronage Intentions
Authors: Susana Marques
Abstract:
Research on how store environment cues influence consumers’ store choice decision criteria, such as store operations, product quality, monetary price, store image and sales promotion, is sparse. Especially absent research on the simultaneous impact of multiple store environment cues. The authors propose a comprehensive store choice model that includes: three types of store environment cues as exogenous constructs; various store choice criteria as possible mediating constructs, and store patronage intentions as an endogenous construct. On the basis of testing with a sample of 561 customers of hypermarkets, the model is partially supported. This study used structural equation modelling to test the proposed model.Keywords: store choice, store patronage, structural equation modelling, retailing
Procedia PDF Downloads 2723303 Basket Option Pricing under Jump Diffusion Models
Authors: Ali Safdari-Vaighani
Abstract:
Pricing financial contracts on several underlying assets received more and more interest as a demand for complex derivatives. The option pricing under asset price involving jump diffusion processes leads to the partial integral differential equation (PIDEs), which is an extension of the Black-Scholes PDE with a new integral term. The aim of this paper is to show how basket option prices in the jump diffusion models, mainly on the Merton model, can be computed using RBF based approximation methods. For a test problem, the RBF-PU method is applied for numerical solution of partial integral differential equation arising from the two-asset European vanilla put options. The numerical result shows the accuracy and efficiency of the presented method.Keywords: basket option, jump diffusion, radial basis function, RBF-PUM
Procedia PDF Downloads 3543302 Virtual Assessment of Measurement Error in the Fractional Flow Reserve
Authors: Keltoum Chahour, Mickael Binois
Abstract:
Due to a lack of standardization during the invasive fractional flow reserve (FFR) procedure, the index is subject to many sources of uncertainties. In this paper, we investigate -through simulation- the effect of the (FFR) device position and configuration on the obtained value of the (FFR) fraction. For this purpose, we use computational fluid dynamics (CFD) in a 3D domain corresponding to a diseased arterial portion. The (FFR) pressure captor is introduced inside it with a given length and coefficient of bending to capture the (FFR) value. To get over the computational limitations, basically, the time of the simulation is about 2h 15min for one (FFR) value; we generate a Gaussian Process (GP) model for (FFR) prediction. The (GP) model indicates good accuracy and demonstrates the effective error in the measurement created by the random configuration of the pressure captor.Keywords: fractional flow reserve, Gaussian processes, computational fluid dynamics, drift
Procedia PDF Downloads 1343301 Modeling Visual Memorability Assessment with Autoencoders Reveals Characteristics of Memorable Images
Authors: Elham Bagheri, Yalda Mohsenzadeh
Abstract:
Image memorability refers to the phenomenon where certain images are more likely to be remembered by humans than others. It is a quantifiable and intrinsic attribute of an image. Understanding how visual perception and memory interact is important in both cognitive science and artificial intelligence. It reveals the complex processes that support human cognition and helps to improve machine learning algorithms by mimicking the brain's efficient data processing and storage mechanisms. To explore the computational underpinnings of image memorability, this study examines the relationship between an image's reconstruction error, distinctiveness in latent space, and its memorability score. A trained autoencoder is used to replicate human-like memorability assessment inspired by the visual memory game employed in memorability estimations. This study leverages a VGG-based autoencoder that is pre-trained on the vast ImageNet dataset, enabling it to recognize patterns and features that are common to a wide and diverse range of images. An empirical analysis is conducted using the MemCat dataset, which includes 10,000 images from five broad categories: animals, sports, food, landscapes, and vehicles, along with their corresponding memorability scores. The memorability score assigned to each image represents the probability of that image being remembered by participants after a single exposure. The autoencoder is finetuned for one epoch with a batch size of one, attempting to create a scenario similar to human memorability experiments where memorability is quantified by the likelihood of an image being remembered after being seen only once. The reconstruction error, which is quantified as the difference between the original and reconstructed images, serves as a measure of how well the autoencoder has learned to represent the data. The reconstruction error of each image, the error reduction, and its distinctiveness in latent space are calculated and correlated with the memorability score. Distinctiveness is measured as the Euclidean distance between each image's latent representation and its nearest neighbor within the autoencoder's latent space. Different structural and perceptual loss functions are considered to quantify the reconstruction error. The results indicate that there is a strong correlation between the reconstruction error and the distinctiveness of images and their memorability scores. This suggests that images with more unique distinct features that challenge the autoencoder's compressive capacities are inherently more memorable. There is also a negative correlation between the reduction in reconstruction error compared to the autoencoder pre-trained on ImageNet, which suggests that highly memorable images are harder to reconstruct, probably due to having features that are more difficult to learn by the autoencoder. These insights suggest a new pathway for evaluating image memorability, which could potentially impact industries reliant on visual content and mark a step forward in merging the fields of artificial intelligence and cognitive science. The current research opens avenues for utilizing neural representations as instruments for understanding and predicting visual memory.Keywords: autoencoder, computational vision, image memorability, image reconstruction, memory retention, reconstruction error, visual perception
Procedia PDF Downloads 90