Search results for: generalized autoregressive score model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18620

Search results for: generalized autoregressive score model

15560 Multi-Stage Multi-Period Production Planning in Wire and Cable Industry

Authors: Mahnaz Hosseinzadeh, Shaghayegh Rezaee Amiri

Abstract:

This paper presents a methodology for serial production planning problem in wire and cable manufacturing process that addresses the problem of input-output imbalance in different consecutive stations, hoping to minimize the halt of machines in each stage. To this end, a linear Goal Programming (GP) model is developed, in which four main categories of constraints as per the number of runs per machine, machines’ sequences, acceptable inventories of machines at the end of each period, and the necessity of fulfillment of the customers’ orders are considered. The model is formulated based upon on the real data obtained from IKO TAK Company, an important supplier of wire and cable for oil and gas and automotive industries in Iran. By solving the model in GAMS software the optimal number of runs, end-of-period inventories, and the possible minimum idle time for each machine are calculated. The application of the numerical results in the target company has shown the efficiency of the proposed model and the solution in decreasing the lead time of the end product delivery to the customers by 20%. Accordingly, the developed model could be easily applied in wire and cable companies for the aim of optimal production planning to reduce the halt of machines in manufacturing stages.

Keywords: goal programming approach, GP, production planning, serial manufacturing process, wire and cable industry

Procedia PDF Downloads 154
15559 A Mathematical Model of Power System State Estimation for Power Flow Solution

Authors: F. Benhamida, A. Graa, L. Benameur, I. Ziane

Abstract:

The state estimation of the electrical power system operation state is very important for supervising task. With the nonlinearity of the AC power flow model, the state estimation problem (SEP) is a nonlinear mathematical problem with many local optima. This paper treat the mathematical model for the SEP and the monitoring of the nonlinear systems of great dimensions with an application on power electrical system, the modelling, the analysis and state estimation synthesis in order to supervise the power system behavior. in fact, it is very difficult, to see impossible, (for reasons of accessibility, techniques and/or of cost) to measure the excessive number of the variables of state in a large-sized system. It is thus important to develop software sensors being able to produce a reliable estimate of the variables necessary for the diagnosis and also for the control.

Keywords: power system, state estimation, robustness, observability

Procedia PDF Downloads 512
15558 Methodology: A Review in Modelling and Predictability of Embankment in Soft Ground

Authors: Bhim Kumar Dahal

Abstract:

Transportation network development in the developing country is in rapid pace. The majority of the network belongs to railway and expressway which passes through diverse topography, landform and geological conditions despite the avoidance principle during route selection. Construction of such networks demand many low to high embankment which required improvement in the foundation soil. This paper is mainly focused on the various advanced ground improvement techniques used to improve the soft soil, modelling approach and its predictability for embankments construction. The ground improvement techniques can be broadly classified in to three groups i.e. densification group, drainage and consolidation group and reinforcement group which are discussed with some case studies.  Various methods were used in modelling of the embankments from simple 1-dimensional to complex 3-dimensional model using variety of constitutive models. However, the reliability of the predictions is not found systematically improved with the level of sophistication.  And sometimes the predictions are deviated more than 60% to the monitored value besides using same level of erudition. This deviation is found mainly due to the selection of constitutive model, assumptions made during different stages, deviation in the selection of model parameters and simplification during physical modelling of the ground condition. This deviation can be reduced by using optimization process, optimization tools and sensitivity analysis of the model parameters which will guide to select the appropriate model parameters.

Keywords: cement, improvement, physical properties, strength

Procedia PDF Downloads 170
15557 Lie Symmetry Treatment for Pricing Options with Transactions Costs under the Fractional Black-Scholes Model

Authors: B. F. Nteumagne, E. Pindza, E. Mare

Abstract:

We apply Lie symmetries analysis to price and hedge options in the fractional Brownian framework. The reputation of Lie groups is well spread in the area of Mathematical sciences and lately, in Finance. In the presence of transactions costs and under fractional Brownian motions, analytical solutions become difficult to obtain. Lie symmetries analysis allows us to simplify the problem and obtain new analytical solution. In this paper, we investigate the use of symmetries to reduce the partial differential equation obtained and obtain the analytical solution. We then proposed a hedging procedure and calibration technique for these types of options, and test the model on real market data. We show the robustness of our methodology by its application to the pricing of digital options.

Keywords: fractional brownian model, symmetry, transaction cost, option pricing

Procedia PDF Downloads 393
15556 Study of the Process of Climate Change According to Data Simulation Using LARS-WG Software during 2010-2030: Case Study of Semnan Province

Authors: Leila Rashidian

Abstract:

Temperature rise on Earth has had harmful effects on the Earth's surface and has led to change in precipitation patterns all around the world. The present research was aimed to study the process of climate change according to the data simulation in future and compare these parameters with current situation in the studied stations in Semnan province including Garmsar, Shahrood and Semnan. In this regard, LARS-WG software, HADCM3 model and A2 scenario were used for the 2010-2030 period. In this model, climatic parameters such as maximum and minimum temperature, precipitation and radiation were used daily. The obtained results indicated that there will be a 4.4% increase in precipitation in Semnan province compared with the observed data, and in general, there will be a 1.9% increase in temperature. This temperature rise has significant impact on precipitation patterns. Most of precipitation will be raining (torrential rains in some cases). According to the results, from west to east, the country will experience more temperature rise and will be warmer.

Keywords: climate change, Semnan province, Lars.WG model, climate parameters, HADCM₃ model

Procedia PDF Downloads 247
15555 Prediction of Oil Recovery Factor Using Artificial Neural Network

Authors: O. P. Oladipo, O. A. Falode

Abstract:

The determination of Recovery Factor is of great importance to the reservoir engineer since it relates reserves to the initial oil in place. Reserves are the producible portion of reservoirs and give an indication of the profitability of a field Development. The core objective of this project is to develop an artificial neural network model using selected reservoir data to predict Recovery Factors (RF) of hydrocarbon reservoirs and compare the model with a couple of the existing correlations. The type of Artificial Neural Network model developed was the Single Layer Feed Forward Network. MATLAB was used as the network simulator and the network was trained using the supervised learning method, Afterwards, the network was tested with input data never seen by the network. The results of the predicted values of the recovery factors of the Artificial Neural Network Model, API Correlation for water drive reservoirs (Sands and Sandstones) and Guthrie and Greenberger Correlation Equation were obtained and compared. It was noted that the coefficient of correlation of the Artificial Neural Network Model was higher than the coefficient of correlations of the other two correlation equations, thus making it a more accurate prediction tool. The Artificial Neural Network, because of its accurate prediction ability is helpful in the correct prediction of hydrocarbon reservoir factors. Artificial Neural Network could be applied in the prediction of other Petroleum Engineering parameters because it is able to recognise complex patterns of data set and establish a relationship between them.

Keywords: recovery factor, reservoir, reserves, artificial neural network, hydrocarbon, MATLAB, API, Guthrie, Greenberger

Procedia PDF Downloads 434
15554 Image Segmentation: New Methods

Authors: Flaurence Benjamain, Michel Casperance

Abstract:

We present in this paper, first, a comparative study of three mathematical theories to achieve the fusion of information sources. This study aims to identify the characteristics inherent in theories of possibilities, belief functions (DST) and plausible and paradoxical reasoning to establish a strategy of choice that allows us to adopt the most appropriate theory to solve a problem of fusion in order, taking into account the acquired information and imperfections that accompany them. Using the new theory of plausible and paradoxical reasoning, also called Dezert-Smarandache Theory (DSmT), to fuse information multi-sources needs, at first step, the generation of the composites events witch is, in general, difficult. Thus, we present in this paper a new approach to construct pertinent paradoxical classes based on gray levels histograms, which also allows to reduce the cardinality of the hyper-powerset. Secondly, we developed a new technique for order and coding generalized focal elements. This method is exploited, in particular, to calculate the cardinality of Dezert and Smarandache. Then, we give an experimentation of classification of a remote sensing image that illustrates the given methods and we compared the result obtained by the DSmT with that resulting from the use of the DST and theory of possibilities.

Keywords: segmentation, image, approach, vision computing

Procedia PDF Downloads 268
15553 Construction of Submerged Aquatic Vegetation Index through Global Sensitivity Analysis of Radiative Transfer Model

Authors: Guanhua Zhou, Zhongqi Ma

Abstract:

Submerged aquatic vegetation (SAV) in wetlands can absorb nitrogen and phosphorus effectively to prevent the eutrophication of water. It is feasible to monitor the distribution of SAV through remote sensing, but for the reason of weak vegetation signals affected by water body, traditional terrestrial vegetation indices are not applicable. This paper aims at constructing SAV index to enhance the vegetation signals and distinguish SAV from water body. The methodology is as follows: (1) select the bands sensitive to the vegetation parameters based on global sensitivity analysis of SAV canopy radiative transfer model; (2) take the soil line concept as reference, analyze the distribution of SAV and water reflectance simulated by SAV canopy model and semi-analytical water model in the two-dimensional space built by different sensitive bands; (3)select the band combinations which have better separation performance between SAV and water, and use them to build the SAVI indices in the form of normalized difference vegetation index(NDVI); (4)analyze the sensitivity of indices to the water and vegetation parameters, choose the one more sensitive to vegetation parameters. It is proved that index formed of the bands with central wavelengths in 705nm and 842nm has high sensitivity to chlorophyll content in leaves while it is less affected by water constituents. The model simulation shows a general negative, little correlation of SAV index with increasing water depth. Moreover, the index enhances capabilities in separating SAV from water compared to NDVI. The SAV index is expected to have potential in parameter inversion of wetland remote sensing.

Keywords: global sensitivity analysis, radiative transfer model, submerged aquatic vegetation, vegetation indices

Procedia PDF Downloads 252
15552 Empirical Modeling of Air Dried Rubberwood Drying System

Authors: S. Khamtree, T. Ratanawilai, C. Nuntadusit

Abstract:

Rubberwood is a crucial commercial timber in Southern Thailand. All processes in a rubberwood production depend on the knowledge and expertise of the technicians, especially the drying process. This research aims to develop an empirical model for drying kinetics in rubberwood. During the experiment, the temperature of the hot air and the average air flow velocity were kept at 80-100 °C and 1.75 m/s, respectively. The moisture content in the samples was determined less than 12% in the achievement of drying basis. The drying kinetic was simulated using an empirical solver. The experimental results illustrated that the moisture content was reduced whereas the drying temperature and time were increased. The coefficient of the moisture ratio between the empirical and the experimental model was tested with three statistical parameters, R-square (), Root Mean Square Error (RMSE) and Chi-square (χ²) to predict the accuracy of the parameters. The experimental moisture ratio had a good fit with the empirical model. Additionally, the results indicated that the drying of rubberwood using the Henderson and Pabis model revealed the suitable level of agreement. The result presented an excellent estimation (= 0.9963) for the moisture movement compared to the other models. Therefore, the empirical results were valid and can be implemented in the future experiments.

Keywords: empirical models, rubberwood, moisture ratio, hot air drying

Procedia PDF Downloads 259
15551 A Stochastic Diffusion Process Based on the Two-Parameters Weibull Density Function

Authors: Meriem Bahij, Ahmed Nafidi, Boujemâa Achchab, Sílvio M. A. Gama, José A. O. Matos

Abstract:

Stochastic modeling concerns the use of probability to model real-world situations in which uncertainty is present. Therefore, the purpose of stochastic modeling is to estimate the probability of outcomes within a forecast, i.e. to be able to predict what conditions or decisions might happen under different situations. In the present study, we present a model of a stochastic diffusion process based on the bi-Weibull distribution function (its trend is proportional to the bi-Weibull probability density function). In general, the Weibull distribution has the ability to assume the characteristics of many different types of distributions. This has made it very popular among engineers and quality practitioners, who have considered it the most commonly used distribution for studying problems such as modeling reliability data, accelerated life testing, and maintainability modeling and analysis. In this work, we start by obtaining the probabilistic characteristics of this model, as the explicit expression of the process, its trends, and its distribution by transforming the diffusion process in a Wiener process as shown in the Ricciaardi theorem. Then, we develop the statistical inference of this model using the maximum likelihood methodology. Finally, we analyse with simulated data the computational problems associated with the parameters, an issue of great importance in its application to real data with the use of the convergence analysis methods. Overall, the use of a stochastic model reflects only a pragmatic decision on the part of the modeler. According to the data that is available and the universe of models known to the modeler, this model represents the best currently available description of the phenomenon under consideration.

Keywords: diffusion process, discrete sampling, likelihood estimation method, simulation, stochastic diffusion process, trends functions, bi-parameters weibull density function

Procedia PDF Downloads 303
15550 A Multi Objective Reliable Location-Inventory Capacitated Disruption Facility Problem with Penalty Cost Solve with Efficient Meta Historic Algorithms

Authors: Elham Taghizadeh, Mostafa Abedzadeh, Mostafa Setak

Abstract:

Logistics network is expected that opened facilities work continuously for a long time horizon without any failure; but in real world problems, facilities may face disruptions. This paper studies a reliable joint inventory location problem to optimize cost of facility locations, customers’ assignment, and inventory management decisions when facilities face failure risks and doesn’t work. In our model we assume when a facility is out of work, its customers may be reassigned to other operational facilities otherwise they must endure high penalty costs associated with losing service. For defining the model closer to real world problems, the model is proposed based on p-median problem and the facilities are considered to have limited capacities. We define a new binary variable (Z_is) for showing that customers are not assigned to any facilities. Our problem involve a bi-objective model; the first one minimizes the sum of facility construction costs and expected inventory holding costs, the second one function that mention for the first one is minimizes maximum expected customer costs under normal and failure scenarios. For solving this model we use NSGAII and MOSS algorithms have been applied to find the pareto- archive solution. Also Response Surface Methodology (RSM) is applied for optimizing the NSGAII Algorithm Parameters. We compare performance of two algorithms with three metrics and the results show NSGAII is more suitable for our model.

Keywords: joint inventory-location problem, facility location, NSGAII, MOSS

Procedia PDF Downloads 520
15549 Owner/Managers’ External Financing Used and Preference towards Islamic Banking

Authors: Khalid Hassan Abdesamed, Kalsom Abd Wahab

Abstract:

Economic development and growth are significantly linked to the consistent and sustainable sector of small and medium enterprises (SMEs). Banks are the frontrunners in financing and advising SMEs. The main objective of the study is to assess the tendency of SMEs to use the Islamic bank. Model was developed using quantitative method with a hypothetical-deductive testing approach. Model (N = 364) used primary data on the tendency of SMEs to use Islamic banks gathered from questionnaire. It is found by Mann-Whitney test that the tendency to use Islamic bank varies between those firms which consider formal financing with the ones relying on informal financing with the latter tends more to use Islamic bank. This study can serve academic researchers, policy makers, and developing countries as a model of SMEs’ desirability to Islamic banking.

Keywords: formal financing, informal financing, Islamic bank, SMEs

Procedia PDF Downloads 345
15548 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks

Authors: Mst Shapna Akter, Hossain Shahriar

Abstract:

One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.

Keywords: cyber security, vulnerability detection, neural networks, feature extraction

Procedia PDF Downloads 81
15547 Site Selection of CNG Station by Using FUZZY-AHP Model (Case Study: Gas Zone 4, Tehran City Iran)

Authors: Hamidrza Joodaki

Abstract:

The most complex issue in urban land use planning is site selection that needs to assess the verity of elements and factors. Multi Criteria Decision Making (MCDM) methods are the best approach to deal with complex problems. In this paper, combination of the analytical hierarchy process (AHP) model and FUZZY logic was used as MCDM methods to select the best site for gas station in the 4th gas zone of Tehran. The first and the most important step in FUZZY-AHP model is selection of criteria and sub-criteria. Population, accessibility, proximity and natural disasters were considered as the main criteria in this study. After choosing the criteria, they were weighted based on AHP by EXPERT CHOICE software, and FUZZY logic was used to enhance accuracy and to approach the reality. After these steps, criteria layers were produced and weighted based on FUZZY-AHP model in GIS. Finally, through ARC GIS software, the layers were integrated and the 4th gas zone in TEHRAN was selected as the best site to locate gas station.

Keywords: multiple criteria decision making (MCDM), analytic hierarchy process (AHP), FUZZY logic, geographic information system (GIS)

Procedia PDF Downloads 352
15546 Discrete Element Simulations of Composite Ceramic Powders

Authors: Julia Cristina Bonaldo, Christophe L. Martin, Severine Romero Baivier, Stephane Mazerat

Abstract:

Alumina refractories are commonly used in steel and foundry industries. These refractories are prepared through a powder metallurgy route. They are a mixture of hard alumina particles and graphite platelets embedded into a soft carbonic matrix (binder). The powder can be cold pressed isostatically or uniaxially, depending on the application. The compact is then fired to obtain the final product. The quality of the product is governed by the microstructure of the composite and by the process parameters. The compaction behavior and the mechanical properties of the fired product depend greatly on the amount of each phase, on their morphology and on the initial microstructure. In order to better understand the link between these parameters and the macroscopic behavior, we use the Discrete Element Method (DEM) to simulate the compaction process and the fracture behavior of the fired composite. These simulations are coupled with well-designed experiments. Four mixes with various amounts of Al₂O₃ and binder were tested both experimentally and numerically. In DEM, each particle is modelled and the interactions between particles are taken into account through appropriate contact or bonding laws. Here, we model a bimodal mixture of large Al₂O₃ and small Al₂O₃ covered with a soft binder. This composite is itself mixed with graphite platelets. X-ray tomography images are used to analyze the morphologies of the different components. Large Al₂O₃ particles and graphite platelets are modelled in DEM as sets of particles bonded together. The binder is modelled as a soft shell that covers both large and small Al₂O₃ particles. When two particles with binder indent each other, they first interact through this soft shell. Once a critical indentation is reached (towards the end of compaction), hard Al₂O₃ - Al₂O₃ contacts appear. In accordance with experimental data, DEM simulations show that the amount of Al₂O₃ and the amount of binder play a major role for the compaction behavior. The graphite platelets bend and break during the compaction, also contributing to the macroscopic stress. Firing step is modeled in DEM by ascribing bonds to particles which contact each other after compaction. The fracture behavior of the compacted mixture is also simulated and compared with experimental data. Both diametrical tests (Brazilian tests) and triaxial tests are carried out. Again, the link between the amount of Al₂O₃ particles and the fracture behavior is investigated. The methodology described here can be generalized to other particulate materials that are used in the ceramic industry.

Keywords: cold compaction, composites, discrete element method, refractory materials, x-ray tomography

Procedia PDF Downloads 136
15545 Analysis of the Occurrence of Hydraulic Fracture Phenomena in Roudbar Lorestan Dam

Authors: Masoud Ghaemi, MohammadJafar Hedayati, Faezeh Yousefzadeh, Hoseinali Heydarzadeh

Abstract:

According to the statistics of the International Committee on Large Dams, internal erosion and piping (scour) are major causes of the destruction of earth-fill dams. If such dams are constructed in narrow valleys, the valley walls will increase the arching of the dam body due to the transfer of vertical and horizontal stresses, so the occurrence of hydraulic fracturing in these embankments is more likely. Roudbar Dam in Lorestan is a clay-core pebble earth-fill dam constructed in a relatively narrow valley in western Iran. Three years after the onset of impoundment, there has been a fall in dam behavior. Evaluation of the dam behavior based on the data recorded on the instruments installed inside the dam body and foundation confirms the occurrence of internal erosion in the lower and adjacent parts of the core on the left support (abutment). The phenomenon of hydraulic fracturing is one of the main causes of the onset of internal erosion in this dam. Accordingly, the main objective of this paper is to evaluate the validity of this hypothesis. To evaluate the validity of this hypothesis, the dam behavior during construction and impoundment has been first simulated with a three-dimensional numerical model. Then, using validated empirical equations, the safety factor of the occurrence of hydraulic fracturing phenomenon upstream of the dam score was calculated. Then, using the artificial neural network, the failure time of the given section was predicted based on the maximum stress trend created. The study results show that steep slopes of valley walls, sudden changes in coefficient, and differences in compressibility properties of dam body materials have caused considerable stress transfer from core to adjacent valley walls, especially at its lower levels. This has resulted in the coefficient of confidence of the occurrence of hydraulic fracturing in each of these areas being close to one in each of the empirical equations used.

Keywords: arching, artificial neural network, FLAC3D, hydraulic fracturing, internal erosion, pore water pressure

Procedia PDF Downloads 173
15544 Presenting a Model Of Empowering New Knowledge-based Companies In Iran Insurance Industry

Authors: Pedram Saadati, Zahra Nazari

Abstract:

In the last decade, the role and importance of knowledge-based technological businesses in the insurance industry has greatly increased, and due to the weakness of previous studies in Iran, the current research deals with the design of the InsurTech empowerment model. In order to obtain the conceptual model of the research, a hybrid framework has been used. The statistical population of the research in the qualitative part were experts, and in the quantitative part, the InsurTech activists. The tools of data collection in the qualitative part were in-depth and semi-structured interviews and structured self-interaction matrix, and in the quantitative part, a researcher-made questionnaire. In the qualitative part, 55 indicators, 20 components and 8 concepts (dimensions) were obtained by the content analysis method, then the relationships of the concepts with each other and the levels of the components were investigated. In the quantitative part, the information was analyzed using the descriptive analytical method in the way of path analysis and confirmatory factor analysis. The proposed model consists of eight dimensions of supporter capability, supervisor of insurance innovation ecosystem, managerial, financial, technological, marketing, opportunity identification, innovative InsurTech capabilities. The results of statistical tests in identifying the relationships of the concepts with each other have been examined in detail and suggestions have been presented in the conclusion section.

Keywords: insurTech, knowledge-base, empowerment model, factor analysis, insurance

Procedia PDF Downloads 36
15543 Generation of Waste Streams in Small Model Reactors

Authors: Sara Mostofian

Abstract:

The nuclear industry is a technology that can fulfill future energy needs but requires special attention to ensure safety and reliability while minimizing any environmental impact. To meet these expectations, the nuclear industry is exploring different reactor technologies for power production. Several designs are under development and the technical viability of these new designs is the subject of many ongoing studies. One of these studies considers the radioactive emissions and radioactive waste generated during the life of a nuclear power production plant to allow a successful license process. For all the modern technologies, a good understanding of the radioactivity generated in the process systems of the plant is essential. Some of that understanding may be gleaned from the performance of some prototype reactors of similar design that operated decades ago. This paper presents how, with that understanding, a model can be developed to estimate the emissions as well as the radioactive waste during the normal operation of a nuclear power plant. The model would predict the radioactive material concentrations in different waste streams. Using this information, the radioactive emission and waste generated during the life of these new technologies can be estimated during the early stages of the design of the plant.

Keywords: SMRs, activity transport, model, radioactive waste

Procedia PDF Downloads 100
15542 Improving the Accuracy of Stress Intensity Factors Obtained by Scaled Boundary Finite Element Method on Hybrid Quadtree Meshes

Authors: Adrian W. Egger, Savvas P. Triantafyllou, Eleni N. Chatzi

Abstract:

The scaled boundary finite element method (SBFEM) is a semi-analytical numerical method, which introduces a scaling center in each element’s domain, thus transitioning from a Cartesian reference frame to one resembling polar coordinates. Consequently, an analytical solution is achieved in radial direction, implying that only the boundary need be discretized. The only limitation imposed on the resulting polygonal elements is that they remain star-convex. Further arbitrary p- or h-refinement may be applied locally in a mesh. The polygonal nature of SBFEM elements has been exploited in quadtree meshes to alleviate all issues conventionally associated with hanging nodes. Furthermore, since in 2D this results in only 16 possible cell configurations, these are precomputed in order to accelerate the forward analysis significantly. Any cells, which are clipped to accommodate the domain geometry, must be computed conventionally. However, since SBFEM permits polygonal elements, significantly coarser meshes at comparable accuracy levels are obtained when compared with conventional quadtree analysis, further increasing the computational efficiency of this scheme. The generalized stress intensity factors (gSIFs) are computed by exploiting the semi-analytical solution in radial direction. This is initiated by placing the scaling center of the element containing the crack at the crack tip. Taking an analytical limit of this element’s stress field as it approaches the crack tip, delivers an expression for the singular stress field. By applying the problem specific boundary conditions, the geometry correction factor is obtained, and the gSIFs are then evaluated based on their formal definition. Since the SBFEM solution is constructed as a power series, not unlike mode superposition in FEM, the two modes contributing to the singular response of the element can be easily identified in post-processing. Compared to the extended finite element method (XFEM) this approach is highly convenient, since neither enrichment terms nor a priori knowledge of the singularity is required. Computation of the gSIFs by SBFEM permits exceptional accuracy, however, when combined with hybrid quadtrees employing linear elements, this does not always hold. Nevertheless, it has been shown that crack propagation schemes are highly effective even given very coarse discretization since they only rely on the ratio of mode one to mode two gSIFs. The absolute values of the gSIFs may still be subject to large errors. Hence, we propose a post-processing scheme, which minimizes the error resulting from the approximation space of the cracked element, thus limiting the error in the gSIFs to the discretization error of the quadtree mesh. This is achieved by h- and/or p-refinement of the cracked element, which elevates the amount of modes present in the solution. The resulting numerical description of the element is highly accurate, with the main error source now stemming from its boundary displacement solution. Numerical examples show that this post-processing procedure can significantly improve the accuracy of the computed gSIFs with negligible computational cost even on coarse meshes resulting from hybrid quadtrees.

Keywords: linear elastic fracture mechanics, generalized stress intensity factors, scaled finite element method, hybrid quadtrees

Procedia PDF Downloads 137
15541 Imperfect Production Inventory Model with Inspection Errors and Fuzzy Demand and Deterioration Rates

Authors: Chayanika Rout, Debjani Chakraborty, Adrijit Goswami

Abstract:

Our work presents an inventory model which illustrates imperfect production and imperfect inspection processes for deteriorating items. A cost-minimizing model is studied considering two types of inspection errors, namely, Type I error of falsely screening out a proportion of non-defects, thereby passing them on for rework and Type II error of falsely not screening out a proportion of defects, thus selling those to customers which incurs a penalty cost. The screened items are reworked; however, no returns are entertained due to deteriorating nature of the items. In more practical situations, certain parameters such as the demand rate and the deterioration rate of inventory cannot be accurately determined, and therefore, they are assumed to be triangular fuzzy numbers in our model. We calculate the optimal lot size that must be produced in order to minimize the total inventory cost for both the crisp and the fuzzy models. A numerical example is also considered to exemplify the procedure which is followed by the analysis of sensitivity of various parameters on the decision variable and the objective function.

Keywords: deteriorating items, EPQ, imperfect quality, rework, type I and type II inspection errors

Procedia PDF Downloads 177
15540 A Mathematical Model of Blood Perfusion Dependent Temperature Distribution in Transient Case in Human Dermal Region

Authors: Yogesh Shukla

Abstract:

Many attempts have been made to study temperature distribution problem in human tissues under normal environmental and physiological conditions at constant arterial blood temperature. But very few attempts have been made to investigate temperature distribution in human tissues under different arterial blood temperature. In view of above, a finite element model has been developed to unsteady temperature distribution in dermal region in human body. The model has been developed for one dimension unsteady state case. The variation in parameters like thermal conductivity, blood mass flow and metabolic activity with respect to position and time has been incorporated in the model. Appropriate boundary conditions have been framed. The central difference approach has been used in space variable and trapezoidal rule has been employed a long time variable. Numerical results have been obtained to study relationship among temperature and time.

Keywords: rate of metabolism, blood mass flow rate, thermal conductivity, heat generation, finite element method

Procedia PDF Downloads 351
15539 Modeling Driving Distraction Considering Psychological-Physical Constraints

Authors: Yixin Zhu, Lishengsa Yue, Jian Sun, Lanyue Tang

Abstract:

Modeling driving distraction in microscopic traffic simulation is crucial for enhancing simulation accuracy. Current driving distraction models are mainly derived from physical motion constraints under distracted states, in which distraction-related error terms are added to existing microscopic driver models. However, the model accuracy is not very satisfying, due to a lack of modeling the cognitive mechanism underlying the distraction. This study models driving distraction based on the Queueing Network Human Processor model (QN-MHP). This study utilizes the queuing structure of the model to perform task invocation and switching for distracted operation and control of the vehicle under driver distraction. Based on the assumption of the QN-MHP model about the cognitive sub-network, server F is a structural bottleneck. The latter information must wait for the previous information to leave server F before it can be processed in server F. Therefore, the waiting time for task switching needs to be calculated. Since the QN-MHP model has different information processing paths for auditory information and visual information, this study divides driving distraction into two types: auditory distraction and visual distraction. For visual distraction, both the visual distraction task and the driving task need to go through the visual perception sub-network, and the stimuli of the two are asynchronous, which is called stimulus on asynchrony (SOA), so when calculating the waiting time for switching tasks, it is necessary to consider it. In the case of auditory distraction, the auditory distraction task and the driving task do not need to compete for the server resources of the perceptual sub-network, and their stimuli can be synchronized without considering the time difference in receiving the stimuli. According to the Theory of Planned Behavior for drivers (TPB), this study uses risk entropy as the decision criterion for driver task switching. A logistic regression model is used with risk entropy as the independent variable to determine whether the driver performs a distraction task, to explain the relationship between perceived risk and distraction. Furthermore, to model a driver’s perception characteristics, a neurophysiological model of visual distraction tasks is incorporated into the QN-MHP, and executes the classical Intelligent Driver Model. The proposed driving distraction model integrates the psychological cognitive process of a driver with the physical motion characteristics, resulting in both high accuracy and interpretability. This paper uses 773 segments of distracted car-following in Shanghai Naturalistic Driving Study data (SH-NDS) to classify the patterns of distracted behavior on different road facilities and obtains three types of distraction patterns: numbness, delay, and aggressiveness. The model was calibrated and verified by simulation. The results indicate that the model can effectively simulate the distracted car-following behavior of different patterns on various roadway facilities, and its performance is better than the traditional IDM model with distraction-related error terms. The proposed model overcomes the limitations of physical-constraints-based models in replicating dangerous driving behaviors, and internal characteristics of an individual. Moreover, the model is demonstrated to effectively generate more dangerous distracted driving scenarios, which can be used to construct high-value automated driving test scenarios.

Keywords: computational cognitive model, driving distraction, microscopic traffic simulation, psychological-physical constraints

Procedia PDF Downloads 82
15538 Electronic Patient Record (EPR) System in South Africa: Results of a Pilot Study

Authors: Temitope O. Tokosi, Visvanathan Naicker

Abstract:

Patient health records contain sensitive information for which an electronic patient record (EPR) system can safely secure and transmit amongst clinicians for use in improving health delivery. Clinician’s use of the behaviour of these systems is under scrutiny to assess their attributes towards health technology. South Africa (SA) clinicians responded to a pilot study survey to assess their understanding of EPR, what attributes are important towards technology use and more importantly streamlining the survey for a larger study. Descriptive statistics using mean scores was used because of the small sample size of 11 clinicians who completed the survey. Nine (9) constructs comprising 62 items were used and a Cronbach alpha score of 0.883 was obtained. Limitations and discussions conclude the study.

Keywords: EPR, clinicians, pilot study, South Africa

Procedia PDF Downloads 259
15537 A Hybrid Method for Determination of Effective Poles Using Clustering Dominant Pole Algorithm

Authors: Anuj Abraham, N. Pappa, Daniel Honc, Rahul Sharma

Abstract:

In this paper, an analysis of some model order reduction techniques is presented. A new hybrid algorithm for model order reduction of linear time invariant systems is compared with the conventional techniques namely Balanced Truncation, Hankel Norm reduction and Dominant Pole Algorithm (DPA). The proposed hybrid algorithm is known as Clustering Dominant Pole Algorithm (CDPA) is able to compute the full set of dominant poles and its cluster center efficiently. The dominant poles of a transfer function are specific eigenvalues of the state space matrix of the corresponding dynamical system. The effectiveness of this novel technique is shown through the simulation results.

Keywords: balanced truncation, clustering, dominant pole, Hankel norm, model reduction

Procedia PDF Downloads 593
15536 Physical Education Effect on Sports Science Analysis Technology

Authors: Peter Adly Hamdy Fahmy

Abstract:

The aim of the study was to examine the effects of a physical education program on student learning by combining the teaching of personal and social responsibility (TPSR) with a physical education model and TPSR with a traditional teaching model, these learning outcomes involving self-learning. -Study. Athletic performance, enthusiasm for sport, group cohesion, sense of responsibility and game performance. The participants were 3 secondary school physical education teachers and 6 physical education classes, 133 participants with students from the experimental group with 75 students and the control group with 58 students, and each teacher taught the experimental group and the control group for 16 weeks. The research methods used surveys, interviews and focus group meetings. Research instruments included the Personal and Social Responsibility Questionnaire, Sports Enthusiasm Scale, Group Cohesion Scale, Sports Self-Efficacy Scale, and Game Performance Assessment Tool. Multivariate analyzes of covariance and repeated measures ANOVA were used to examine differences in student learning outcomes between combining the TPSR with a physical education model and the TPSR with a traditional teaching model. The research findings are as follows: 1) The TPSR sports education model can improve students' learning outcomes, including sports self-efficacy, game performance, sports enthusiasm, team cohesion, group awareness and responsibility. 2) A traditional teaching model with TPSR could improve student learning outcomes, including sports self-efficacy, responsibility, and game performance. 3) The sports education model with TPSR could improve learning outcomes more than the traditional teaching model with TPSR, including sports self-efficacy, sports enthusiasm, responsibility and game performance. 4) Based on qualitative data on teachers' and students' learning experience, the physical education model with TPSR significantly improves learning motivation, group interaction and sense of play. The results suggest that physical education with TPSR could further improve learning outcomes in the physical education program. On the other hand, the hybrid model curriculum projects TPSR - Physical Education and TPSR - Traditional Education are good curriculum projects for moral character education that can be used in school physics.

Keywords: approach competencies, physical, education, teachers employment, graduate, physical education and sport sciences, SWOT analysis character education, sport season, game performance, sport competence

Procedia PDF Downloads 37
15535 Experimentally Validated Analytical Model for Thermal Analysis of Multi-Stage Depressed Collector

Authors: Vishant Gahlaut, A Mercy Latha, Sanjay Kumar Ghosh

Abstract:

Multi-stage depressed collectors (MDC) are used as an efficiency enhancement technique in traveling wave tubes the high-energy electron beam, after its interaction with the RF signal, gets velocity sorted and collected at various depressed electrodes of the MDC. The ultimate goal is to identify an optimum thermal management scheme (cooling mechanism) that could extract the heat efficiently from the electrodes. Careful thermal analysis, incorporating the cooling mechanism is required to ensure that the maximum temperature does not exceed the safe limits. A simple analytical model for quick prediction of the thermal has been developed. The model has been developed for the worst-case un-modulated DC condition, where all the thermal power is dissipated in the last electrode (typically, fourth electrode in the case of the four-stage depressed collector). It considers the thermal contact resistances at various braze joints accounting for the practical non-uniformities. Analytical results obtained from the model have been validated with simulated and experimental results.

Keywords: multi-stage depressed collector, TWTs, thermal contact resistance, thermal management

Procedia PDF Downloads 218
15534 Intentional Relationship Building: Stem Faculty Perceptions of Culturally Responsive Mentoring

Authors: Niesha Douglas, Lisa Merriweather, Cathy Howell, Anna Sancyzk

Abstract:

Many studies explain that mentoring in an academic setting contributes to student success and retention. However, in the United States, where the population is diverse and filled with multiple ethnic groups, mentoring has become too generalized and fails to offer a unique individualized experience for underrepresented minorities (URM). The purpose of this paper is to describe the findings of an ongoing qualitative study that investigates the relationships among STEM doctoral faculty and URM students. Several faculty from three different predominately white institutions (PWI) in the Southeastern region of the United States were interviewed and engaged in open dialogue about their experiences with mentoring. The data collection included semi-structured interviews that took place in the classroom (pre-COVID-19) as well as virtually. The theoretical framework draws on the idea of Critical Race Theory and how cultural, social constructs interfere with effective mentoring for URM Doctoral STEM students. The findings in this study suggest that though the faculty and several years of experience mentoring students, there were some gaps in understanding the needs of URM students and how mentoring is a unique relationship that should be specialized for each student and should not fit into one mold.

Keywords: culture, critical race theory, mentoring, STEM

Procedia PDF Downloads 191
15533 Hydro-Mechanical Forming of AZ31 Sheet

Authors: Yong-Nam Kwon

Abstract:

In the present study, we have designed the hydro-mechanical forming in which AZ31 sheet was drawn to a kind of preform step following gas blow forming for accurate geometry. In order to judge a formability enhancement of AZ31 sheet, model geometry came from a practical automotive part which had quite depth with complicated curvatures, which was proven that a single sheet forming could not gave a successful part. Experimentally, we succeeded to make the model part with accurate dimension. The optimum forming conditions for respective forming steps were considered most important technical features of this hydro-mechanical and would be discussed in details. Also, the effort to avoid detrimental abnormal grain growth was given and discussed for a practical application.

Keywords: hydro-mechanical forming, AZ31, abnormal grain growth, model geometry

Procedia PDF Downloads 508
15532 Online Learning for Modern Business Models: Theoretical Considerations and Algorithms

Authors: Marian Sorin Ionescu, Olivia Negoita, Cosmin Dobrin

Abstract:

This scientific communication reports and discusses learning models adaptable to modern business problems and models specific to digital concepts and paradigms. In the PAC (probably approximately correct) learning model approach, in which the learning process begins by receiving a batch of learning examples, the set of learning processes is used to acquire a hypothesis, and when the learning process is fully used, this hypothesis is used in the prediction of new operational examples. For complex business models, a lot of models should be introduced and evaluated to estimate the induced results so that the totality of the results are used to develop a predictive rule, which anticipates the choice of new models. In opposition, for online learning-type processes, there is no separation between the learning (training) and predictive phase. Every time a business model is approached, a test example is considered from the beginning until the prediction of the appearance of a model considered correct from the point of view of the business decision. After choosing choice a part of the business model, the label with the logical value "true" is known. Some of the business models are used as examples of learning (training), which helps to improve the prediction mechanisms for future business models.

Keywords: machine learning, business models, convex analysis, online learning

Procedia PDF Downloads 136
15531 Dynamics of the Coupled Fitzhugh-Rinzel Neurons

Authors: Sanjeev Kumar Sharma, Arnab Mondal, Ranjit Kumar Upadhyay

Abstract:

Excitable cells often produce different oscillatory activities that help us to understand the transmitting and processing of signals in the neural system. We consider a FitzHugh-Rinzel (FH-R) model and studied the different dynamics of the model by considering the parameter c as the predominant parameter. The model exhibits different types of neuronal responses such as regular spiking, mixed-mode bursting oscillations (MMBOs), elliptic bursting, etc. Based on the bifurcation diagram, we consider the three regimes (MMBOs, elliptic bursting, and quiescent state). An analytical treatment for the occurrence of the supercritical Hopf bifurcation is studied. Further, we extend our study to a network of a hundred neurons by considering the bi-directional synaptic coupling between them. In this article, we investigate the alternation of spiking propagation and bursting phenomena of an uncoupled and coupled FH-R neurons. We explore that the complete graph of heterogenous desynchronized neurons can exhibit different types of bursting oscillations for certain coupling strength. For higher coupling strength, all the neurons in the network show complete synchronization.

Keywords: excitable neuron model, spiking-bursting, stability and bifurcation, synchronization networks

Procedia PDF Downloads 118