Search results for: mathematical shape deformation model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19951

Search results for: mathematical shape deformation model

16501 Mathematical Model for Flow and Sediment Yield Estimation on Tel River Basin, India

Authors: Santosh Kumar Biswal, Ramakar Jha

Abstract:

Soil erosion is a slow and continuous process and one of the prominent problems across the world leading to many serious problems like loss of soil fertility, loss of soil structure, poor internal drainage, sedimentation deposits etc. In this paper remote sensing and GIS based methods have been applied for the determination of soil erosion and sediment yield. Tel River basin which is the second largest tributary of the river Mahanadi laying between latitude 19° 15' 32.4"N and, 20° 45' 0"N and longitude 82° 3' 36"E and 84° 18' 18"E chosen for the present study. The catchment was discretized into approximately homogeneous sub-areas (grid cells) to overcome the catchment heterogeneity. The gross soil erosion in each cell was computed using Universal Soil Loss Equation (USLE). Various parameters for USLE was determined as a function of land topography, soil texture, land use/land cover, rainfall, erosivity and crop management and practice in the watershed. The concept of transport limited accumulation was formulated and the transport capacity maps were generated. The gross soil erosion was routed to the catchment outlet. This study can help in recognizing critical erosion prone areas of the study basin so that suitable control measures can be implemented.

Keywords: Universal Soil Loss Equation (USLE), GIS, land use, sediment yield,

Procedia PDF Downloads 318
16500 Fuzzy Logic Based Fault Tolerant Model Predictive MLI Topology

Authors: Abhimanyu Kumar, Chirag Gupta

Abstract:

This work presents a comprehensive study on the employment of Model Predictive Control (MPC) for a three-phase voltage-source inverter to regulate the output voltage efficiently. The inverter is modeled via the Clarke Transformation, considering a scenario where the load is unknown. An LC filter model is developed, demonstrating its efficacy in Total Harmonic Distortion (THD) reduction. The system, when implemented with fault-tolerant multilevel inverter topologies, ensures reliable operation even under fault conditions, a requirement that is paramount with the increasing dependence on renewable energy sources. The research also integrates a Fuzzy Logic based fault tolerance system which identifies and manages faults, ensuring consistent inverter performance. The efficacy of the proposed methodology is substantiated through rigorous simulations and comparative results, shedding light on the voltage prediction efficiency and the robustness of the model even under fault conditions.

Keywords: total harmonic distortion, fuzzy logic, renewable energy sources, MLI

Procedia PDF Downloads 143
16499 Calibration and Validation of ArcSWAT Model for Estimation of Surface Runoff and Sediment Yield from Dhangaon Watershed

Authors: M. P. Tripathi, Priti Tiwari

Abstract:

Soil and Water Assessment Tool (SWAT) is a distributed parameter continuous time model and was tested on daily and fortnightly basis for a small agricultural watershed (Dhangaon) of Chhattisgarh state in India. The SWAT model recently interfaced with ArcGIS and called as ArcSWAT. The watershed and sub-watershed boundaries, drainage networks, slope and texture maps were generated in the environment of ArcGIS of ArcSWAT. Supervised classification method was used for land use/cover classification from satellite imageries of the years 2009 and 2012. Manning's roughness coefficient 'n' for overland flow and channel flow and Fraction of Field Capacity (FFC) were calibrated for monsoon season of the years 2009 and 2010. The model was validated on a daily basis for the years 2011 and 2012 by using the observed daily rainfall and temperature data. Calibration and validation results revealed that the model was predicting the daily surface runoff and sediment yield satisfactorily. Sensitivity analysis showed that the annual sediment yield was inversely proportional to the overland and channel 'n' values whereas; annual runoff and sediment yields were directly proportional to the FFC. The model was also tested (calibrated and validated) for the fortnightly runoff and sediment yield for the year 2009-10 and 2011-12, respectively. Simulated values of fortnightly runoff and sediment yield for the calibration and validation years compared well with their observed counterparts. The calibration and validation results revealed that the ArcSWAT model could be used for identification of critical sub-watershed and for developing management scenarios for the Dhangaon watershed. Further, the model should be tested for simulating the surface runoff and sediment yield using generated rainfall and temperature before applying it for developing the management scenario for the critical or priority sub-watersheds.

Keywords: watershed, hydrologic and water quality, ArcSWAT model, remote sensing, GIS, runoff and sediment yield

Procedia PDF Downloads 384
16498 Experimental Evaluation of UDP in Wireless LAN

Authors: Omar Imhemed Alramli

Abstract:

As Transmission Control Protocol (TCP), User Datagram Protocol (UDP) is transfer protocol in the transportation layer in Open Systems Interconnection model (OSI model) or in TCP/IP model of networks. The UDP aspects evaluation were not recognized by using the pcattcp tool on the windows operating system platform like TCP. The study has been carried out to find a tool which supports UDP aspects evolution. After the information collection about different tools, iperf tool was chosen and implemented on Cygwin tool which is installed on both Windows XP platform and also on Windows XP on virtual box machine on one computer only. Iperf is used to make experimental evaluation of UDP and to see what will happen during the sending the packets between the Host and Guest in wired and wireless networks. Many test scenarios have been done and the major UDP aspects such as jitter, packet losses, and throughput are evaluated.

Keywords: TCP, UDP, IPERF, wireless LAN

Procedia PDF Downloads 360
16497 Create a Dynamic Model in Project Control and Management

Authors: Hamed Saremi, Shahla Saremi

Abstract:

In this study, control and management of construction projects is evaluated through developing a dynamic model in which some means are used in order to evaluating planning assumptions and reviewing the effectiveness of some project control policies based on previous researches about time, cost, project schedule pressure management, source management, project control, adding elements and sub-systems from cost management such as estimating consumption budget from budget due to costs, budget shortage effects and etc. using sensitivity analysis, researcher has evaluated introduced model that during model simulation by VENSIM software and assuming optimistic times and adding information about doing job and changes rate and project is forecasted with 373 days (2 days sooner than forecasted) and final profit $ 1,960,670 (23% amount of contract) assuming 15% inflation rate in year and costs rate accordance with planned amounts and other input information and final profit.

Keywords: dynamic planning, cost, time, performance, project management

Procedia PDF Downloads 481
16496 Proposal of a Model Supporting Decision-Making Based on Multi-Objective Optimization Analysis on Information Security Risk Treatment

Authors: Ritsuko Kawasaki (Aiba), Takeshi Hiromatsu

Abstract:

Management is required to understand all information security risks within an organization, and to make decisions on which information security risks should be treated in what level by allocating how much amount of cost. However, such decision-making is not usually easy, because various measures for risk treatment must be selected with the suitable application levels. In addition, some measures may have objectives conflicting with each other. It also makes the selection difficult. Moreover, risks generally have trends and it also should be considered in risk treatment. Therefore, this paper provides the extension of the model proposed in the previous study. The original model supports the selection of measures by applying a combination of weighted average method and goal programming method for multi-objective analysis to find an optimal solution. The extended model includes the notion of weights to the risks, and the larger weight means the priority of the risk.

Keywords: information security risk treatment, selection of risk measures, risk acceptance, multi-objective optimization

Procedia PDF Downloads 466
16495 Sampled-Data Model Predictive Tracking Control for Mobile Robot

Authors: Wookyong Kwon, Sangmoon Lee

Abstract:

In this paper, a sampled-data model predictive tracking control method is presented for mobile robots which is modeled as constrained continuous-time linear parameter varying (LPV) systems. The presented sampled-data predictive controller is designed by linear matrix inequality approach. Based on the input delay approach, a controller design condition is derived by constructing a new Lyapunov function. Finally, a numerical example is given to demonstrate the effectiveness of the presented method.

Keywords: model predictive control, sampled-data control, linear parameter varying systems, LPV

Procedia PDF Downloads 316
16494 Applications of AFM in 4D to Optimize the Design of Genetic Nanoparticles

Authors: Hosam Abdelhady

Abstract:

Filming the behaviors of individual DNA molecules in their environment when they interact with individual medicinal nano-polymers in a molecular scale has opened the door to understand the effect of the molecular shape, size, and incubation time with nanocarriers on optimizing the design of robust genetic Nano molecules able to resist the enzymatic degradation, enter the cell, reach to the nucleus and kill individual cancer cells in their environment. To this end, we will show how we applied the 4D AFM as a guide to finetune the design of genetic nanoparticles and to film the effects of these nanoparticles on the nanomechanical and morphological profiles of individual cancer cells.

Keywords: AFM, dendrimers, nanoparticles, DNA, gene therapy, imaging

Procedia PDF Downloads 80
16493 Estimation of Maize Yield by Using a Process-Based Model and Remote Sensing Data in the Northeast China Plain

Authors: Jia Zhang, Fengmei Yao, Yanjing Tan

Abstract:

The accurate estimation of crop yield is of great importance for the food security. In this study, a process-based mechanism model was modified to estimate yield of C4 crop by modifying the carbon metabolic pathway in the photosynthesis sub-module of the RS-P-YEC (Remote-Sensing-Photosynthesis-Yield estimation for Crops) model. The yield was calculated by multiplying net primary productivity (NPP) and the harvest index (HI) derived from the ratio of grain to stalk yield. The modified RS-P-YEC model was used to simulate maize yield in the Northeast China Plain during the period 2002-2011. The statistical data of maize yield from study area was used to validate the simulated results at county-level. The results showed that the Pearson correlation coefficient (R) was 0.827 (P < 0.01) between the simulated yield and the statistical data, and the root mean square error (RMSE) was 712 kg/ha with a relative error (RE) of 9.3%. From 2002-2011, the yield of maize planting zone in the Northeast China Plain was increasing with smaller coefficient of variation (CV). The spatial pattern of simulated maize yield was consistent with the actual distribution in the Northeast China Plain, with an increasing trend from the northeast to the southwest. Hence the results demonstrated that the modified process-based model coupled with remote sensing data was suitable for yield prediction of maize in the Northeast China Plain at the spatial scale.

Keywords: process-based model, C4 crop, maize yield, remote sensing, Northeast China Plain

Procedia PDF Downloads 383
16492 Inverse Scattering of Two-Dimensional Objects Using an Enhancement Method

Authors: A.R. Eskandari, M.R. Eskandari

Abstract:

A 2D complete identification algorithm for dielectric and multiple objects immersed in air is presented. The employed technique consists of initially retrieving the shape and position of the scattering object using a linear sampling method and then determining the electric permittivity and conductivity of the scatterer using adjoint sensitivity analysis. This inversion algorithm results in high computational speed and efficiency, and it can be generalized for any scatterer structure. Also, this method is robust with respect to noise. The numerical results clearly show that this hybrid approach provides accurate reconstructions of various objects.

Keywords: inverse scattering, microwave imaging, two-dimensional objects, Linear Sampling Method (LSM)

Procedia PDF Downloads 390
16491 Eco-Design of Multifunctional System Based on a Shape Memory Polymer and ZnO Nanoparticles for Sportswear

Authors: Inês Boticas, Diana P. Ferreira, Ana Eusébio, Carlos Silva, Pedro Magalhães, Ricardo Silva, Raul Fangueiro

Abstract:

Since the beginning of the 20th century, sportswear has a major contribution to the impact of fashion on our lives. Nowadays, the embracing of sportswear fashion/looks is undoubtedly noticeable, as the modern consumer searches for high comfort and linear aesthetics for its clothes. This compromise lead to the arise of the athleisure trend. Athleisure surges as a new style area that combines both wearability and fashion sense, differentiated from the archetypal sportswear, usually associated to “gym clothes”. Additionally, the possibility to functionalize and implement new technologies have shifted and progressively empowers the connection between the concepts of physical activities practice and well-being, allowing clothing to be more interactive and responsive with its surroundings. In this study, a design inspired in retro and urban lifestyle was envisioned, engineering textile structures that can respond to external stimuli. These structures are enhanced to be responsive to heat, water vapor and humidity, integrating shape memory polymers (SMP) to improve the breathability and heat-responsive behavior of the textiles and zinc oxide nanoparticles (ZnO NPs) to heighten the surface hydrophobic properties. The best results for hydrophobic exhibited superhydrophobic behavior with water contact angle (WAC) of more than 150 degrees. For the breathability and heat-response properties, SMP-coated samples showed an increase in water vapour permeability values of about 50% when compared with non SMP-coated samples. These innovative technological approaches were endorsed to design innovative clothing, in line with circular economy and eco-design principles, by assigning a substantial degree of mutability and versatility to the clothing. The development of a coat and shirt, in which different parts can be purchased separately to create multiple products, aims to combine the technicality of both the fabrics used and the making of the garments. This concept translates itself into a real constructive mechanism through the symbiosis of high-tech functionalities and the timeless design that follows the athleisure aesthetics.

Keywords: breathability, sportswear and casual clothing, sustainable design, superhydrophobicity

Procedia PDF Downloads 140
16490 Facility Data Model as Integration and Interoperability Platform

Authors: Nikola Tomasevic, Marko Batic, Sanja Vranes

Abstract:

Emerging Semantic Web technologies can be seen as the next step in evolution of the intelligent facility management systems. Particularly, this considers increased usage of open source and/or standardized concepts for data classification and semantic interpretation. To deliver such facility management systems, providing the comprehensive integration and interoperability platform in from of the facility data model is a prerequisite. In this paper, one of the possible modelling approaches to provide such integrative facility data model which was based on the ontology modelling concept was presented. Complete ontology development process, starting from the input data acquisition, ontology concepts definition and finally ontology concepts population, was described. At the beginning, the core facility ontology was developed representing the generic facility infrastructure comprised of the common facility concepts relevant from the facility management perspective. To develop the data model of a specific facility infrastructure, first extension and then population of the core facility ontology was performed. For the development of the full-blown facility data models, Malpensa and Fiumicino airports in Italy, two major European air-traffic hubs, were chosen as a test-bed platform. Furthermore, the way how these ontology models supported the integration and interoperability of the overall airport energy management system was analyzed as well.

Keywords: airport ontology, energy management, facility data model, ontology modeling

Procedia PDF Downloads 453
16489 The Impacts of Local Decision Making on Customisation Process Speed across Distributed Boundaries

Authors: Abdulrahman M. Qahtani, Gary. B. Wills, Andy. M. Gravell

Abstract:

Communicating and managing customers’ requirements in software development projects play a vital role in the software development process. While it is difficult to do so locally, it is even more difficult to communicate these requirements over distributed boundaries and to convey them to multiple distribution customers. This paper discusses the communication of multiple distribution customers’ requirements in the context of customised software products. The main purpose is to understand the challenges of communicating and managing customisation requirements across distributed boundaries. We propose a model for Communicating Customisation Requirements of Multi-Clients in a Distributed Domain (CCRD). Thereafter, we evaluate that model by presenting the findings of a case study conducted with a company with customisation projects for 18 distributed customers. Then, we compare the outputs of the real case process and the outputs of the CCRD model using simulation methods. Our conjecture is that the CCRD model can reduce the challenge of communication requirements over distributed organisational boundaries, and the delay in decision making and in the entire customisation process time.

Keywords: customisation software products, global software engineering, local decision making, requirement engineering, simulation model

Procedia PDF Downloads 434
16488 Modeling of Surface Roughness in Hard Turning of DIN 1.2210 Cold Work Tool Steel with Ceramic Tools

Authors: Mehmet Erdi Korkmaz, Mustafa Günay

Abstract:

Nowadays, grinding is frequently replaced with hard turning for reducing set up time and higher accuracy. This paper focused on mathematical modeling of average surface roughness (Ra) in hard turning of AISI L2 grade (DIN 1.2210) cold work tool steel with ceramic tools. The steel was hardened to 60±1 HRC after the heat treatment process. Cutting speed, feed rate, depth of cut and tool nose radius was chosen as the cutting conditions. The uncoated ceramic cutting tools were used in the machining experiments. The machining experiments were performed according to Taguchi L27 orthogonal array on CNC lathe. Ra values were calculated by averaging three roughness values obtained from three different points of machined surface. The influences of cutting conditions on surface roughness were evaluated as statistical and experimental. The analysis of variance (ANOVA) with 95% confidence level was applied for statistical analysis of experimental results. Finally, mathematical models were developed using the artificial neural networks (ANN). ANOVA results show that feed rate is the dominant factor affecting surface roughness, followed by tool nose radius and cutting speed.

Keywords: ANN, hard turning, DIN 1.2210, surface roughness, Taguchi method

Procedia PDF Downloads 375
16487 Dynamical Relation of Poisson Spike Trains in Hodkin-Huxley Neural Ion Current Model and Formation of Non-Canonical Bases, Islands, and Analog Bases in DNA, mRNA, and RNA at or near the Transcription

Authors: Michael Fundator

Abstract:

Groundbreaking application of biomathematical and biochemical research in neural networks processes to formation of non-canonical bases, islands, and analog bases in DNA and mRNA at or near the transcription that contradicts the long anticipated statistical assumptions for the distribution of bases and analog bases compounds is implemented through statistical and stochastic methods apparatus with addition of quantum principles, where the usual transience of Poisson spike train becomes very instrumental tool for finding even almost periodical type of solutions to Fokker-Plank stochastic differential equation. Present article develops new multidimensional methods of finding solutions to stochastic differential equations based on more rigorous approach to mathematical apparatus through Kolmogorov-Chentsov continuity theorem that allows the stochastic processes with jumps under certain conditions to have γ-Holder continuous modification that is used as basis for finding analogous parallels in dynamics of neutral networks and formation of analog bases and transcription in DNA.

Keywords: Fokker-Plank stochastic differential equation, Kolmogorov-Chentsov continuity theorem, neural networks, translation and transcription

Procedia PDF Downloads 410
16486 Study of the Relationship between the Roughness Configuration of Channel Bottom and the Creation of Vortices at the Rough Area: Numerical Modelling

Authors: Youb Said, Fourar Ali

Abstract:

To describe the influence of bottom roughness on the free surface flows by numerical modeling, a two-dimensional model was developed. The equations of continuity and momentum (Naviers Stokes equations) are solved by the finite volume method. We considered a turbulent flow in an open channel with a bottom roughness. For our simulations, the K-ε model was used. After setting the initial and boundary conditions and solve the equations set, we were able to achieve the following results: vortex forming in the hollow causing substantial energy dissipation in the obstacle areas that form the bottom roughness. The comparison of our results with experimental ones shows a good agreement in terms of the results in the rough area. However, in other areas, differences were more or less important. These differences are in areas far from the bottom, especially the free surface area just after the bottom. These disagreements are probably due to experimental constants used by the k-ε model.

Keywords: modeling, free surface flow, turbulence, bottom roughness, finite volume, K-ε model, energy dissipation

Procedia PDF Downloads 382
16485 A Methodology Based on Image Processing and Deep Learning for Automatic Characterization of Graphene Oxide

Authors: Rafael do Amaral Teodoro, Leandro Augusto da Silva

Abstract:

Originated from graphite, graphene is a two-dimensional (2D) material that promises to revolutionize technology in many different areas, such as energy, telecommunications, civil construction, aviation, textile, and medicine. This is possible because its structure, formed by carbon bonds, provides desirable optical, thermal, and mechanical characteristics that are interesting to multiple areas of the market. Thus, several research and development centers are studying different manufacturing methods and material applications of graphene, which are often compromised by the scarcity of more agile and accurate methodologies to characterize the material – that is to determine its composition, shape, size, and the number of layers and crystals. To engage in this search, this study proposes a computational methodology that applies deep learning to identify graphene oxide crystals in order to characterize samples by crystal sizes. To achieve this, a fully convolutional neural network called U-net has been trained to segment SEM graphene oxide images. The segmentation generated by the U-net is fine-tuned with a standard deviation technique by classes, which allows crystals to be distinguished with different labels through an object delimitation algorithm. As a next step, the characteristics of the position, area, perimeter, and lateral measures of each detected crystal are extracted from the images. This information generates a database with the dimensions of the crystals that compose the samples. Finally, graphs are automatically created showing the frequency distributions by area size and perimeter of the crystals. This methodological process resulted in a high capacity of segmentation of graphene oxide crystals, presenting accuracy and F-score equal to 95% and 94%, respectively, over the test set. Such performance demonstrates a high generalization capacity of the method in crystal segmentation, since its performance considers significant changes in image extraction quality. The measurement of non-overlapping crystals presented an average error of 6% for the different measurement metrics, thus suggesting that the model provides a high-performance measurement for non-overlapping segmentations. For overlapping crystals, however, a limitation of the model was identified. To overcome this limitation, it is important to ensure that the samples to be analyzed are properly prepared. This will minimize crystal overlap in the SEM image acquisition and guarantee a lower error in the measurements without greater efforts for data handling. All in all, the method developed is a time optimizer with a high measurement value, considering that it is capable of measuring hundreds of graphene oxide crystals in seconds, saving weeks of manual work.

Keywords: characterization, graphene oxide, nanomaterials, U-net, deep learning

Procedia PDF Downloads 164
16484 Using of Particle Swarm Optimization for Loss Minimization of Vector-Controlled Induction Motors

Authors: V. Rashtchi, H. Bizhani, F. R. Tatari

Abstract:

This paper presents a new online loss minimization for an induction motor drive. Among the many loss minimization algorithms (LMAs) for an induction motor, a particle swarm optimization (PSO) has the advantages of fast response and high accuracy. However, the performance of the PSO and other optimization algorithms depend on the accuracy of the modeling of the motor drive and losses. In the development of the loss model, there is always a trade off between accuracy and complexity. This paper presents a new online optimization to determine an optimum flux level for the efficiency optimization of the vector-controlled induction motor drive. An induction motor (IM) model in d-q coordinates is referenced to the rotor magnetizing current. This transformation results in no leakage inductance on the rotor side, thus the decomposition into d-q components in the steady-state motor model can be utilized in deriving the motor loss model. The suggested algorithm is simple for implementation.

Keywords: induction machine, loss minimization, magnetizing current, particle swarm optimization

Procedia PDF Downloads 638
16483 A Convolutional Neural Network-Based Model for Lassa fever Virus Prediction Using Patient Blood Smear Image

Authors: A. M. John-Otumu, M. M. Rahman, M. C. Onuoha, E. P. Ojonugwa

Abstract:

A Convolutional Neural Network (CNN) model for predicting Lassa fever was built using Python 3.8.0 programming language, alongside Keras 2.2.4 and TensorFlow 2.6.1 libraries as the development environment in order to reduce the current high risk of Lassa fever in West Africa, particularly in Nigeria. The study was prompted by some major flaws in existing conventional laboratory equipment for diagnosing Lassa fever (RT-PCR), as well as flaws in AI-based techniques that have been used for probing and prognosis of Lassa fever based on literature. There were 15,679 blood smear microscopic image datasets collected in total. The proposed model was trained on 70% of the dataset and tested on 30% of the microscopic images in avoid overfitting. A 3x3x3 convolution filter was also used in the proposed system to extract features from microscopic images. The proposed CNN-based model had a recall value of 96%, a precision value of 93%, an F1 score of 95%, and an accuracy of 94% in predicting and accurately classifying the images into clean or infected samples. Based on empirical evidence from the results of the literature consulted, the proposed model outperformed other existing AI-based techniques evaluated. If properly deployed, the model will assist physicians, medical laboratory scientists, and patients in making accurate diagnoses for Lassa fever cases, allowing the mortality rate due to the Lassa fever virus to be reduced through sound decision-making.

Keywords: artificial intelligence, ANN, blood smear, CNN, deep learning, Lassa fever

Procedia PDF Downloads 125
16482 Residual Stresses and Crystallographic Texture of Magnesium AZ31-C Alloy Welded by Friction Stir Welding (FSW)

Authors: A. Kouadri-Henni, L. Barrallier

Abstract:

The objective of the study was to characterize the properties of a magnesium alloy welded by friction stir welding (FSW). The results led to a better understanding of the relationship between this process, the microstructure and anisotropic properties of alloy materials. Welding principally leads to a large reduction in grain size in welded zones due to the phenomenon of dynamic recrystallization. The most remarkable observation was that crystallographic textures changed from a base metal with one texture in two zones: the thermo-mechanically affected and stir welded zones. The latter zone has the peculiarity of possessing a marked texture with two components on the basal plane and the pyramidal plane. These characteristics disappeared in the TMAZ, which had only one component following the basal plane. These modifications have been explained by the nature of the plastic deformation in these zones, which occurs at a moderate temperature in the TMAZ and high temperature in the SWZ. In the same time, we compared this evolution with the nature and the level of the residual stresses obtained by X-ray diffraction.

Keywords: texture christallography, residual stresses, FSW process

Procedia PDF Downloads 370
16481 Developing Measurement Instruments for Enterprise Resources Planning (ERP) Post-Implementation Failure Model

Authors: Malihe Motiei, Nor Hidayati Zakaria, Davide Aloini

Abstract:

This study aims to present a method to develop the failure measurement model for ERP post-implementation. To achieve this outcome, the study firstly evaluates the suitability of Technology-Organization-Environment framework for the proposed conceptual model. This study explains how to discover the constructs and subsequently to design and evaluate the constructs as formative or reflective. Constructs are used including reflective and purely formative. Then, the risk dimensions are investigated to determine the instruments to examine the impact of risk on ERP failure after implementation. Two construct as formative constructs consist inadequate implementation and poor organizational decision making. Subsequently six construct as reflective construct include technical risks, operational risks, managerial risks, top management risks, lack of external risks, and user’s inefficiency risks. A survey was conducted among Iranian industries to collect data. 69 data were collected from manufacturing sectors and the data were analyzed by Smart PLS software. The results indicated that all measurements included 39 critical risk factors were acceptable for the ERP post-implementation failure model.

Keywords: critical risk factors (CRFs), ERP projects, ERP post-implementation, measurement instruments, ERP system failure measurement model

Procedia PDF Downloads 370
16480 The Methodology of System Modeling of Mechatronic Systems

Authors: Lakhoua Najeh

Abstract:

Aims of the work: After a presentation of the functionality of an example of a mechatronic system which is a paint mixer system, we present the concepts of modeling and safe operation. This paper briefly discusses how to model and protect the functioning of a mechatronic system relying mainly on functional analysis and safe operation techniques. Methods: For the study of an example of a mechatronic system, we use methods for external functional analysis that illustrate the relationships between a mechatronic system and its external environment. Thus, we present the Safe-Structured Analysis Design Technique method (Safe-SADT) which allows the representation of a mechatronic system. A model of operating safety and automation is proposed. This model enables us to use a functional analysis technique of the mechatronic system based on the GRAFCET (Graphe Fonctionnel de Commande des Etapes et Transitions: Step Transition Function Chart) method; study of the safe operation of the mechatronic system based on the Safe-SADT method; automation of the mechatronic system based on a software tool. Results: The expected results are to propose a model and safe operation of a mechatronic system. This methodology enables us to analyze the relevance of the different models based on Safe-SADT and GRAFCET in relation to the control and monitoring functions and to study the means allowing exploiting their synergy. Conclusion: In order to propose a general model of a mechatronic system, a model of analysis, safety operation and automation of a mechatronic system has been developed. This is how we propose to validate this methodology through a case study of a paint mixer system.

Keywords: mechatronic systems, system modeling, safe operation, Safe-SADT

Procedia PDF Downloads 247
16479 Sensitive Analysis of the ZF Model for ABC Multi Criteria Inventory Classification

Authors: Makram Ben Jeddou

Abstract:

The ABC classification is widely used by managers for inventory control. The classical ABC classification is based on the Pareto principle and according to the criterion of the annual use value only. Single criterion classification is often insufficient for a closely inventory control. Multi-criteria inventory classification models have been proposed by researchers in order to take into account other important criteria. From these models, we will consider the ZF model in order to make a sensitive analysis on the composite score calculated for each item. In fact, this score based on a normalized average between a good and a bad optimized index can affect the ABC items classification. We will then focus on the weights assigned to each index and propose a classification compromise.

Keywords: ABC classification, multi criteria inventory classification models, ZF-model

Procedia PDF Downloads 512
16478 An Economic Order Quantity Model for Deteriorating Items with Ramp Type Demand, Time Dependent Holding Cost and Price Discount Offered on Backorders

Authors: Arjun Paul, Adrijit Goswami

Abstract:

In our present work, an economic order quantity inventory model with shortages is developed where holding cost is expressed as linearly increasing function of time and demand rate is a ramp type function of time. The items considered in the model are deteriorating in nature so that a small fraction of the items is depleted with the passage of time. In order to consider a more realistic situation, the deterioration rate is assumed to follow a continuous uniform distribution with the parameters involved being triangular fuzzy numbers. The inventory manager offers his customer a discount in case he is willing to backorder his demand when there is a stock-out. The optimum ordering policy and the optimum discount offered for each backorder are determined by minimizing the total cost in a replenishment interval. For better illustration of our proposed model in both the crisp and fuzzy sense and for providing richer insights, a numerical example is cited to exemplify the policy and to analyze the sensitivity of the model parameters.

Keywords: fuzzy deterioration rate, price discount on backorder, ramp type demand, shortage, time varying holding cost

Procedia PDF Downloads 202
16477 Effect of Atmospheric Turbulence on Hybrid FSO/RF Link Availability under Qatar's Harsh Climate

Authors: Abir Touati, Syed Jawad Hussain, Farid Touati, Ammar Bouallegue

Abstract:

Although there has been a growing interest in the hybrid free-space optical link and radio frequency FSO/RF communication system, the current literature is limited to results obtained in moderate or cold environment. In this paper, using a soft switching approach, we investigate the effect of weather inhomogeneities on the strength of turbulence hence the channel refractive index under Qatar harsh environment and their influence on the hybrid FSO/RF availability. In this approach, either FSO/RF or simultaneous or none of them can be active. Based on soft switching approach and a finite state Markov Chain (FSMC) process, we model the channel fading for the two links and derive a mathematical expression for the outage probability of the hybrid system. Then, we evaluate the behavior of the hybrid FSO/RF under hazy and harsh weather. Results show that the FSO/RF soft switching renders the system outage probability less than that of each link individually. A soft switching algorithm is being implemented on FPGAs using Raptor code interfaced to the two terminals of a 1Gbps/100 Mbps FSO/RF hybrid system, the first being implemented in the region. Experimental results are compared to the above simulation results.

Keywords: atmospheric turbulence, haze, hybrid FSO/RF, outage probability, refractive index

Procedia PDF Downloads 423
16476 Enhancing Model Interoperability and Reuse by Designing and Developing a Unified Metamodel Standard

Authors: Arash Gharibi

Abstract:

Mankind has always used models to solve problems. Essentially, models are simplified versions of reality, whose need stems from having to deal with complexity; many processes or phenomena are too complex to be described completely. Thus a fundamental model requirement is that it contains the characteristic features that are essential in the context of the problem to be solved or described. Models are used in virtually every scientific domain to deal with various problems. During the recent decades, the number of models has increased exponentially. Publication of models as part of original research has traditionally been in in scientific periodicals, series, monographs, agency reports, national journals and laboratory reports. This makes it difficult for interested groups and communities to stay informed about the state-of-the-art. During the modeling process, many important decisions are made which impact the final form of the model. Without a record of these considerations, the final model remains ill-defined and open to varying interpretations. Unfortunately, the details of these considerations are often lost or in case there is any existing information about a model, it is likely to be written intuitively in different layouts and in different degrees of detail. In order to overcome these issues, different domains have attempted to implement their own approaches to preserve their models’ information in forms of model documentation. The most frequently cited model documentation approaches show that they are domain specific, not to applicable to the existing models and evolutionary flexibility and intrinsic corrections and improvements are not possible with the current approaches. These issues are all because of a lack of unified standards for model documentation. As a way forward, this research will propose a new standard for capturing and managing models’ information in a unified way so that interoperability and reusability of models become possible. This standard will also be evolutionary, meaning members of modeling realm could contribute to its ongoing developments and improvements. In this paper, the current 3 of the most common metamodels are reviewed and according to pros and cons of each, a new metamodel is proposed.

Keywords: metamodel, modeling, interoperability, reuse

Procedia PDF Downloads 204
16475 Optical and Double Folding Model Analysis for Alpha Particles Elastically Scattered from 9Be and 11B Nuclei at Different Energies

Authors: Ahmed H. Amer, A. Amar, Sh. Hamada, I. I. Bondouk, F. A. El-Hussiny

Abstract:

Elastic scattering of α-particles from 9Be and 11B nuclei at different alpha energies have been analyzed. Optical model parameters (OMPs) of α-particles elastic scattering by these nuclei at different energies have been obtained. In the present calculations, the real part of the optical potential are derived by folding of nucleon-nucleon (NN) interaction into nuclear matter density distribution of the projectile and target nuclei using computer code FRESCO. A density-dependent version of the M3Y interaction (CDM3Y6), which is based on the G-matrix elements of the Paris NN potential, has been used. Volumetric integrals of the real and imaginary potential depth (JR, JW) have been calculated and found to be energy dependent. Good agreement between the experimental data and the theoretical predictions in the whole angular range. In double folding (DF) calculations, the obtained normalization coefficient Nr is in the range 0.70–1.32.

Keywords: elastic scattering, optical model, double folding model, density distribution

Procedia PDF Downloads 295
16474 Predictive Modelling Approaches in Food Processing and Safety

Authors: Amandeep Sharma, Digvaijay Verma, Ruplal Choudhary

Abstract:

Food processing is an activity across the globe that help in better handling of agricultural produce, including dairy, meat, and fish. The operations carried out in the food industry includes raw material quality authenticity; sorting and grading; processing into various products using thermal treatments – heating, freezing, and chilling; packaging; and storage at the appropriate temperature to maximize the shelf life of the products. All this is done to safeguard the food products and to ensure the distribution up to the consumer. The approaches to develop predictive models based on mathematical or statistical tools or empirical models’ development has been reported for various milk processing activities, including plant maintenance and wastage. Recently AI is the key factor for the fourth industrial revolution. AI plays a vital role in the food industry, not only in quality and food security but also in different areas such as manufacturing, packaging, and cleaning. A new conceptual model was developed, which shows that smaller sample size as only spectra would be required to predict the other values hence leads to saving on raw materials and chemicals otherwise used for experimentation during the research and new product development activity. It would be a futuristic approach if these tools can be further clubbed with the mobile phones through some software development for their real time application in the field for quality check and traceability of the product.

Keywords: predictive modlleing, ann, ai, food

Procedia PDF Downloads 85
16473 Comparison of Solar Radiation Models

Authors: O. Behar, A. Khellaf, K. Mohammedi, S. Ait Kaci

Abstract:

Up to now, most validation studies have been based on the MBE and RMSE, and therefore, focused only on long and short terms performance to test and classify solar radiation models. This traditional analysis does not take into account the quality of modeling and linearity. In our analysis we have tested 22 solar radiation models that are capable to provide instantaneous direct and global radiation at any given location Worldwide. We introduce a new indicator, which we named Global Accuracy Indicator (GAI) to examine the linear relationship between the measured and predicted values and the quality of modeling in addition to long and short terms performance. Note that the quality of model has been represented by the T-Statistical test, the model linearity has been given by the correlation coefficient and the long and short term performance have been respectively known by the MBE and RMSE. An important founding of this research is that the use GAI allows avoiding default validation when using traditional methodology that might results in erroneous prediction of solar power conversion systems performances.

Keywords: solar radiation model, parametric model, performance analysis, Global Accuracy Indicator (GAI)

Procedia PDF Downloads 358
16472 Microwave Sintering and Its Application on Cemented Carbides

Authors: Rumman M. D. Raihanuzzaman, Lee Chang Chuan, Zonghan Xie, Reza Ghomashchi

Abstract:

Cemented carbides, owing to their excellent mechanical properties, have been of immense interest in the field of hard materials for the past few decades. A number of processing techniques have been developed to obtain high quality carbide tools, with a wide range of grain size depending on the application and requirements. Microwave sintering is one of the heating processes, which has been used on a wide range of materials including ceramics. The complete understanding of microwave sintering and its contribution towards control of grain growth and on deformation of the resulting carbide materials needs further studies and attention. In addition, the effect of binder materials and their behaviour as a function of microwave sintering is another area that requires clear understanding. This review aims to focus on microwave sintering, providing information of how the process works and what type of materials it is best suited for. In addition, a closer look at some microwave sintered Tungsten Carbide-Cobalt samples will be taken and discussed, addressing some of the key issues and challenges faced in the research.

Keywords: cemented carbides, consolidation, microwave sintering, mechanical properties

Procedia PDF Downloads 603