Search results for: low order model
26888 A Multi-Criteria Decision Making Approach for Disassembly-To-Order Systems under Uncertainty
Authors: Ammar Y. Alqahtani
Abstract:
In order to minimize the negative impact on the environment, it is essential to manage the waste that generated from the premature disposal of end-of-life (EOL) products properly. Consequently, government and international organizations introduced new policies and regulations to minimize the amount of waste being sent to landfills. Moreover, the consumers’ awareness regards environment has forced original equipment manufacturers to consider being more environmentally conscious. Therefore, manufacturers have thought of different ways to deal with waste generated from EOL products viz., remanufacturing, reusing, recycling, or disposing of EOL products. The rate of depletion of virgin natural resources and their dependency on the natural resources can be reduced by manufacturers when EOL products are treated as remanufactured, reused, or recycled, as well as this will cut on the amount of harmful waste sent to landfills. However, disposal of EOL products contributes to the problem and therefore is used as a last option. Number of EOL need to be estimated in order to fulfill the components demand. Then, disassembly process needs to be performed to extract individual components and subassemblies. Smart products, built with sensors embedded and network connectivity to enable the collection and exchange of data, utilize sensors that are implanted into products during production. These sensors are used for remanufacturers to predict an optimal warranty policy and time period that should be offered to customers who purchase remanufactured components and products. Sensor-provided data can help to evaluate the overall condition of a product, as well as the remaining lives of product components, prior to perform a disassembly process. In this paper, a multi-period disassembly-to-order (DTO) model is developed that takes into consideration the different system uncertainties. The DTO model is solved using Nonlinear Programming (NLP) in multiple periods. A DTO system is considered where a variety of EOL products are purchased for disassembly. The model’s main objective is to determine the best combination of EOL products to be purchased from every supplier in each period which maximized the total profit of the system while satisfying the demand. This paper also addressed the impact of sensor embedded products on the cost of warranties. Lastly, this paper presented and analyzed a case study involving various simulation conditions to illustrate the applicability of the model.Keywords: closed-loop supply chains, environmentally conscious manufacturing, product recovery, reverse logistics
Procedia PDF Downloads 13726887 Global Mittag-Leffler Stability of Fractional-Order Bidirectional Associative Memory Neural Network with Discrete and Distributed Transmission Delays
Authors: Swati Tyagi, Syed Abbas
Abstract:
Fractional-order Hopfield neural networks are generally used to model the information processing among the interacting neurons. To show the constancy of the processed information, it is required to analyze the stability of these systems. In this work, we perform Mittag-Leffler stability for the corresponding Caputo fractional-order bidirectional associative memory (BAM) neural networks with various time-delays. We derive sufficient conditions to ensure the existence and uniqueness of the equilibrium point by using the theory of topological degree theory. By applying the fractional Lyapunov method and Mittag-Leffler functions, we derive sufficient conditions for the global Mittag-Leffler stability, which further imply the global asymptotic stability of the network equilibrium. Finally, we present two suitable examples to show the effectiveness of the obtained results.Keywords: bidirectional associative memory neural network, existence and uniqueness, fractional-order, Lyapunov function, Mittag-Leffler stability
Procedia PDF Downloads 36326886 Geometric Contrast of a 3D Model Obtained by Means of Digital Photogrametry with a Quasimetric Camera on UAV Classical Methods
Authors: Julio Manuel de Luis Ruiz, Javier Sedano Cibrián, Rubén Pérez Álvarez, Raúl Pereda García, Cristina Diego Soroa
Abstract:
Nowadays, the use of drones has been extended to practically any human activity. One of the main applications is focused on the surveying field. In this regard, software programs that process the images captured by the sensor from the drone in an almost automatic way have been developed and commercialized, but they only allow contrasting the results through control points. This work proposes the contrast of a 3D model obtained from a flight developed by a drone and a non-metric camera (due to its low cost), with a second model that is obtained by means of the historically-endorsed classical methods. In addition to this, the contrast is developed over a certain territory with a significant unevenness, so as to test the model generated with photogrammetry, and considering that photogrammetry with drones finds more difficulties in terms of accuracy in this kind of situations. Distances, heights, surfaces and volumes are measured on the basis of the 3D models generated, and the results are contrasted. The differences are about 0.2% for the measurement of distances and heights, 0.3% for surfaces and 0.6% when measuring volumes. Although they are not important, they do not meet the order of magnitude that is presented by salespeople.Keywords: accuracy, classical topographic, model tridimensional, photogrammetry, Uav.
Procedia PDF Downloads 13326885 Measuring Financial Asset Return and Volatility Spillovers, with Application to Sovereign Bond, Equity, Foreign Exchange and Commodity Markets
Authors: Petra Palic, Maruska Vizek
Abstract:
We provide an in-depth analysis of interdependence of asset returns and volatilities in developed and developing countries. The analysis is split into three parts. In the first part, we use multivariate GARCH model in order to provide stylized facts on cross-market volatility spillovers. In the second part, we use a generalized vector autoregressive methodology developed by Diebold and Yilmaz (2009) in order to estimate separate measures of return spillovers and volatility spillovers among sovereign bond, equity, foreign exchange and commodity markets. In particular, our analysis is focused on cross-market return, and volatility spillovers in 19 developed and developing countries. In order to estimate named spillovers, we use daily data from 2008 to 2017. In the third part of the analysis, we use a generalized vector autoregressive framework in order to estimate total and directional volatility spillovers. We use the same daily data span for one developed and one developing country in order to characterize daily volatility spillovers across stock, bond, foreign exchange and commodities markets.Keywords: cross-market spillovers, sovereign bond markets, equity markets, value at risk (VAR)
Procedia PDF Downloads 26126884 Model Observability – A Monitoring Solution for Machine Learning Models
Authors: Amreth Chandrasehar
Abstract:
Machine Learning (ML) Models are developed and run in production to solve various use cases that help organizations to be more efficient and help drive the business. But this comes at a massive development cost and lost business opportunities. According to the Gartner report, 85% of data science projects fail, and one of the factors impacting this is not paying attention to Model Observability. Model Observability helps the developers and operators to pinpoint the model performance issues data drift and help identify root cause of issues. This paper focuses on providing insights into incorporating model observability in model development and operationalizing it in production.Keywords: model observability, monitoring, drift detection, ML observability platform
Procedia PDF Downloads 11226883 Create a Dynamic Model in Project Control and Management
Authors: Hamed Saremi, Shahla Saremi
Abstract:
In this study, control and management of construction projects is evaluated through developing a dynamic model in which some means are used in order to evaluating planning assumptions and reviewing the effectiveness of some project control policies based on previous researches about time, cost, project schedule pressure management, source management, project control, adding elements and sub-systems from cost management such as estimating consumption budget from budget due to costs, budget shortage effects and etc. using sensitivity analysis, researcher has evaluated introduced model that during model simulation by VENSIM software and assuming optimistic times and adding information about doing job and changes rate and project is forecasted with 373 days (2 days sooner than forecasted) and final profit $ 1,960,670 (23% amount of contract) assuming 15% inflation rate in year and costs rate accordance with planned amounts and other input information and final profit.Keywords: dynamic planning, cost, time, performance, project management
Procedia PDF Downloads 47826882 Methodology of Construction Equipment Optimization for Earthwork
Authors: Jaehyun Choi, Hyunjung Kim, Namho Kim
Abstract:
Earthwork is one of the critical civil construction operations that require large-quantities of resources due to its intensive dependency upon construction equipment. Therefore, efficient construction equipment management can highly contribute to productivity improvements and cost savings. Earthwork operation utilizes various combinations of construction equipment in order to meet project requirements such as time and cost. Identification of site condition and construction methods should be performed in advance in order to develop a proper execution plan. The factors to be considered include capacity of equipment assigned, the method of construction, the size of the site, and the surrounding condition. In addition, optimal combination of various construction equipment should be selected. However, in real world practice, equipment utilization plan is performed based on experience and intuition of management. The researchers evaluated the efficiency of various alternatives of construction equipment combinations by utilizing the process simulation model, validated the model from a case study project, and presented a methodology to find optimized plan among alternatives.Keywords: earthwork operation, construction equipment, process simulation, optimization
Procedia PDF Downloads 42626881 Reliability Assessment of Various Empirical Formulas for Prediction of Scour Hole Depth (Plunge Pool) Using a Comprehensive Physical Model
Authors: Majid Galoie, Khodadad Safavi, Abdolreza Karami Nejad, Reza Roshan
Abstract:
In this study, a comprehensive scouring model has been developed in order to evaluate the accuracy of various empirical relationships which were suggested for prediction of scour hole depth in plunge pools by Martins, Mason, Chian and Veronese. For this reason, scour hole depths caused by free falling jets from a flip bucket to a plunge pool were investigated. In this study various discharges, angles, scouring times, etc. have been considered. The final results demonstrated that the all mentioned empirical formulas, except Mason formula, were reasonably agreement with the experimental data.Keywords: scour hole depth, plunge pool, physical model, reliability assessment
Procedia PDF Downloads 53526880 Flow Characteristics around Rectangular Obstacles with the Varying Direction of Obstacles
Authors: Hee-Chang Lim
Abstract:
The study aims to understand the surface pressure distribution around the bodies such as the suction pressure in the leading edge on the top and side-face when the aspect ratio of bodies and the wind direction are changed, respectively. We carried out the wind tunnel measurement and numerical simulation around a series of rectangular bodies (40d×80w×80h, 80d×80w×80h, 160d×80w×80h, 80d×40w×80h and 80d×160w×80h in mm3) placed in a deep turbulent boundary layer. Based on a modern numerical platform, the Navier-Stokes equation with the typical 2-equation (k-ε model) and the DES (Detached Eddy Simulation) turbulence model has been calculated, and they are both compared with the measurement data. Regarding the turbulence model, the DES model makes a better prediction comparing with the k-ε model, especially when calculating the separated turbulent flow around a bluff body with sharp edged corner. In order to observe the effect of wind direction on the pressure variation around the cube (e.g., 80d×80w×80h in mm), it rotates at 0º, 10º, 20º, 30º, and 45º, which stands for the salient wind directions in the tunnel. The result shows that the surface pressure variation is highly dependent upon the approaching wind direction, especially on the top and the side-face of the cube. In addition, the transverse width has a substantial effect on the variation of surface pressure around the bodies, while the longitudinal length has little or no influence.Keywords: rectangular bodies, wind direction, aspect ratio, surface pressure distribution, wind-tunnel measurement, k-ε model, DES model, CFD
Procedia PDF Downloads 23526879 An Evaluation Model for Enhancing Flexibility in Production Systems through Additive Manufacturing
Authors: Angela Luft, Sebastian Bremen, Nicolae Balc
Abstract:
Additive manufacturing processes have entered large parts of the industry and their range of application have progressed and grown significantly in the course of time. A major advantage of additive manufacturing is the innate flexibility of the machines. This corelates with the ongoing demand of creating highly flexible production environments. However, the potential of additive manufacturing technologies to enhance the flexibility of production systems has not yet been truly considered and quantified in a systematic way. In order to determine the potential of additive manufacturing technologies with regards to the strategic flexibility design in production systems, an integrated evaluation model has been developed, that allows for the simultaneous consideration of both conventional as well as additive production resources. With the described model, an operational scope of action can be identified and quantified in terms of mix and volume flexibility, process complexity, and machine capacity that goes beyond the current cost-oriented approaches and offers a much broader and more holistic view on the potential of additive manufacturing. A respective evaluation model is presented this paper.Keywords: additive manufacturing, capacity planning, production systems, strategic production planning, flexibility enhancement
Procedia PDF Downloads 15726878 Inventory Decisions for Perishable Products with Age and Stock Dependent Demand Rate
Authors: Maher Agi, Hardik Soni
Abstract:
This paper presents a deterministic model for optimized control of the inventory of a perishable product subject to both physical deterioration and degradation of its freshness condition. The demand for the product depends on its current inventory level and freshness condition. Our model allows for any positive amount of end of cycle inventory. Some useful conditions that characterize the optimal solution of the model are derived and an algorithm is presented for finding the optimal values of the price, the inventory cycle, the end of cycle inventory level and the order quantity. Numerical examples are then given. Our work shows how the product freshness in conjunction with the inventory deterioration affects the inventory management decisions.Keywords: inventory management, lot sizing, perishable products, deteriorating inventory, age-dependent demand, stock-dependent demand
Procedia PDF Downloads 23426877 Degradation Model for UK Railway Drainage System
Authors: Yiqi Wu, Simon Tait, Andrew Nichols
Abstract:
Management of UK railway drainage assets is challenging due to the large amounts of historical assets with long asset life cycles. A major concern for asset managers is to maintain the required performance economically and efficiently while complying with the relevant regulation and legislation. As the majority of the drainage assets are buried underground and are often difficult or costly to examine, it is important for asset managers to understand and model the degradation process in order to foresee the upcoming reduction in asset performance and conduct proactive maintenance accordingly. In this research, a Markov chain approach is used to model the deterioration process of rail drainage assets. The study is based on historical condition scores and characteristics of drainage assets across the whole railway network in England, Scotland, and Wales. The model is used to examine the effect of various characteristics on the probabilities of degradation, for example, the regional difference in probabilities of degradation, and how material and shape can influence the deterioration process for chambers, channels, and pipes.Keywords: deterioration, degradation, markov models, probability, railway drainage
Procedia PDF Downloads 22126876 General Architecture for Automation of Machine Learning Practices
Authors: U. Borasi, Amit Kr. Jain, Rakesh, Piyush Jain
Abstract:
Data collection, data preparation, model training, model evaluation, and deployment are all processes in a typical machine learning workflow. Training data needs to be gathered and organised. This often entails collecting a sizable dataset and cleaning it to remove or correct any inaccurate or missing information. Preparing the data for use in the machine learning model requires pre-processing it after it has been acquired. This often entails actions like scaling or normalising the data, handling outliers, selecting appropriate features, reducing dimensionality, etc. This pre-processed data is then used to train a model on some machine learning algorithm. After the model has been trained, it needs to be assessed by determining metrics like accuracy, precision, and recall, utilising a test dataset. Every time a new model is built, both data pre-processing and model training—two crucial processes in the Machine learning (ML) workflow—must be carried out. Thus, there are various Machine Learning algorithms that can be employed for every single approach to data pre-processing, generating a large set of combinations to choose from. Example: for every method to handle missing values (dropping records, replacing with mean, etc.), for every scaling technique, and for every combination of features selected, a different algorithm can be used. As a result, in order to get the optimum outcomes, these tasks are frequently repeated in different combinations. This paper suggests a simple architecture for organizing this largely produced “combination set of pre-processing steps and algorithms” into an automated workflow which simplifies the task of carrying out all possibilities.Keywords: machine learning, automation, AUTOML, architecture, operator pool, configuration, scheduler
Procedia PDF Downloads 5726875 Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore
Authors: Ronal Muresano, Andrea Pagano
Abstract:
Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.Keywords: algorithm optimization, bank failures, OpenMP, parallel techniques, statistical tool
Procedia PDF Downloads 36926874 Estimation of the Parameters of Muskingum Methods for the Prediction of the Flood Depth in the Moudjar River Catchment
Authors: Fares Laouacheria, Said Kechida, Moncef Chabi
Abstract:
The objective of the study was based on the hydrological routing modelling for the continuous monitoring of the hydrological situation in the Moudjar river catchment, especially during floods with Hydrologic Engineering Center–Hydrologic Modelling Systems (HEC-HMS). The HEC-GeoHMS was used to transform data from geographic information system (GIS) to HEC-HMS for delineating and modelling the catchment river in order to estimate the runoff volume, which is used as inputs to the hydrological routing model. Two hydrological routing models were used, namely Muskingum and Muskingum routing models, for conducting this study. In this study, a comparison between the parameters of the Muskingum and Muskingum-Cunge routing models in HEC-HMS was used for modelling flood routing in the Moudjar river catchment and determining the relationship between these parameters and the physical characteristics of the river. The results indicate that the effects of input parameters such as the weighting factor "X" and travel time "K" on the output results are more significant, where the Muskingum routing model was more sensitive to input parameters than the Muskingum-Cunge routing model. This study can contribute to understand and improve the knowledge of the mechanisms of river floods, especially in ungauged river catchments.Keywords: HEC-HMS, hydrological modelling, Muskingum routing model, Muskingum-Cunge routing model
Procedia PDF Downloads 27626873 Derivation of Bathymetry from High-Resolution Satellite Images: Comparison of Empirical Methods through Geographical Error Analysis
Authors: Anusha P. Wijesundara, Dulap I. Rathnayake, Nihal D. Perera
Abstract:
Bathymetric information is fundamental importance to coastal and marine planning and management, nautical navigation, and scientific studies of marine environments. Satellite-derived bathymetry data provide detailed information in areas where conventional sounding data is lacking and conventional surveys are inaccessible. The two empirical approaches of log-linear bathymetric inversion model and non-linear bathymetric inversion model are applied for deriving bathymetry from high-resolution multispectral satellite imagery. This study compares these two approaches by means of geographical error analysis for the site Kankesanturai using WorldView-2 satellite imagery. Based on the Levenberg-Marquardt method calibrated the parameters of non-linear inversion model and the multiple-linear regression model was applied to calibrate the log-linear inversion model. In order to calibrate both models, Single Beam Echo Sounding (SBES) data in this study area were used as reference points. Residuals were calculated as the difference between the derived depth values and the validation echo sounder bathymetry data and the geographical distribution of model residuals was mapped. The spatial autocorrelation was calculated by comparing the performance of the bathymetric models and the results showing the geographic errors for both models. A spatial error model was constructed from the initial bathymetry estimates and the estimates of autocorrelation. This spatial error model is used to generate more reliable estimates of bathymetry by quantifying autocorrelation of model error and incorporating this into an improved regression model. Log-linear model (R²=0.846) performs better than the non- linear model (R²=0.692). Finally, the spatial error models improved bathymetric estimates derived from linear and non-linear models up to R²=0.854 and R²=0.704 respectively. The Root Mean Square Error (RMSE) was calculated for all reference points in various depth ranges. The magnitude of the prediction error increases with depth for both the log-linear and the non-linear inversion models. Overall RMSE for log-linear and the non-linear inversion models were ±1.532 m and ±2.089 m, respectively.Keywords: log-linear model, multi spectral, residuals, spatial error model
Procedia PDF Downloads 29726872 Modeling the Time-Dependent Rheological Behavior of Clays Used in Fabrication of Ceramic
Authors: Larbi Hammadi, N. Boudjenane, N. Benhallou, R. Houjedje, R. Reffis, M. Belhadri
Abstract:
Many of clays exhibited the thixotropic behavior in which, the apparent viscosity of material decreases with time of shearing at constant shear rate. The structural kinetic model (SKM) was used to characterize the thixotropic behavior of two different kinds of clays used in fabrication of ceramic. Clays selected for analysis represent the fluid and semisolid clays materials. The SKM postulates that the change in the rheological behavior is associated with shear-induced breakdown of the internal structure of the clays. This model for the structure decay with time at constant shear rate assumes nth order kinetics for the decay of the material structure with a rate constant.Keywords: ceramic, clays, structural kinetic model, thixotropy, viscosity
Procedia PDF Downloads 40926871 Design and Analysis of Formula One Car Halo
Authors: Indira priyadarshini, B. Tulja Lal, K. Anusha, P. Sai Varun
Abstract:
Formula One cars are the fastest road course racing cars in the world, owing to very high cornering speeds achieved through the generation of large amounts of aerodynamic downforce. The main intentions and goals of this paper are to reduce the accidents and improving the safety without affecting the visibility of the driver by redesigning Halo that was developed by Mercedes in conjunction with the FIA to deflect flying debris, such as a loose wheel, away from a driver’s head while the hinged locking mechanism can quickly be removed for easy access. Halo design has been modified in order to reduce the weight without affecting the aerodynamics of the car. CFD simulation is carried out to observe the flow over the Halo. The velocity profile and pressure contours were analyzed. Halo is designed using SOLIDWORKS Furthermore, using the software ANSYS FLUENT 3D simulation of the airflow contour around the Halo in order to make changes in the geometry to improve the design by reducing air resistance and improving aerodynamics. According to our assumption, new 3D Halo model has better aerodynamic properties in order to analyse possible improvements compared to the initial design. Structural analysis is also done by using ANSYS by making an F1 tire colliding with Halo at 225 kmph in order to know the deflections in the structure.Keywords: aerodynamics, Halo, safety, visibility
Procedia PDF Downloads 37226870 Seismic Performance of Reinforced Concrete Frame Structure Based on Plastic Rotation
Authors: Kahil Amar, Meziani Faroudja, Khelil Nacim
Abstract:
The principal objective of this study is the evaluation of the seismic performance of reinforced concrete frame structures, taking into account of the behavior laws, reflecting the real behavior of materials, using CASTEM2000 software. A finite element model used is based in modified Takeda model with Timoshenko elements for columns and beams. This model is validated on a Vecchio experimental reinforced concrete (RC) frame model. Then, a study focused on the behavior of a RC frame with three-level and three-story in order to visualize the positioning the plastic hinge (plastic rotation), determined from the curvature distribution along the elements. The results obtained show that the beams of the 1st and 2nd level developed a very large plastic rotations, or these rotations exceed the values corresponding to CP (Collapse prevention with cp qCP = 0.02 rad), against those developed at the 3rd level, are between IO and LS (Immediate occupancy and life Safety with qIO = 0.005 rad and rad qLS = 0.01 respectively), so the beams of first and second levels submit a very significant damage.Keywords: seismic performance, performance level, pushover analysis, plastic rotation, plastic hinge
Procedia PDF Downloads 13026869 Numerical Investigation of Geotextile Application in Clay Reinforcement in ABAQUS Software
Authors: Seyed Abolhasan Naeini, Eisa Aliagahei
Abstract:
Today, the use of geosynthetic materials in geotechnical activities is increasing significantly. One of the main uses of these materials is to increase the compressive strength of clay reinforced by geotextile layers. In the present study, the effect of clay reinforcement by geotextile layers in increasing the compressive strength of clay has been investigated using modeling in ABAQUS 6.11.3 software. For this purpose, the modified Drager Prager model has been chosen to simulate the stress-strain behavior of soil layers and the linear elastic model for the geotextile layer. Unreinforced samples and reinforced samples are modeled by geotextile layers (1, 2 and 3 geotextile layers) by software. In order to validate the results, an article in the same field was used and the numerical modeling results were calibrated with the laboratory results. Based on the obtained results, the software has a suitable capability for modeling and the results of the numerical model overlap with the laboratory results to a very acceptable extent, by increasing the number of geotextile layers, the error between the results of the laboratory sample and the software model increases. The highest amount of error is related to the sample reinforced with three layers of geotextile and is 7.3%.Keywords: Abaqus, cap model, clay, geotextile layer, reinforced soil
Procedia PDF Downloads 8826868 Seismic Fragility for Sliding Failure of Weir Structure Considering the Process of Concrete Aging
Authors: HoYoung Son, Ki Young Kim, Woo Young Jung
Abstract:
This study investigated the change of weir structure performances when durability of concrete, which is the main material of weir structure, decreased due to their aging by mean of seismic fragility analysis. In the analysis, it was assumed that the elastic modulus of concrete was reduced by 10% in order to account for their aged deterioration. Additionally, the analysis of seismic fragility was based on Monte Carlo Simulation method combined with a 2D nonlinear finite element in ABAQUS platform with the consideration of deterioration of concrete. Finally, the comparison of seismic fragility of model pre- and post-deterioration was made to study the performance of weir. Results show that the probability of failure in moderate damage for deteriorated model was found to be larger than pre-deterioration model when peak ground acceleration (PGA) passed 0.4 g.Keywords: weir, FEM, concrete, fragility, aging
Procedia PDF Downloads 42426867 A Novel Model for Saturation Velocity Region of Graphene Nanoribbon Transistor
Authors: Mohsen Khaledian, Razali Ismail, Mehdi Saeidmanesh, Mahdiar Hosseinghadiry
Abstract:
A semi-analytical model for impact ionization coefficient of graphene nanoribbon (GNR) is presented. The model is derived by calculating probability of electrons reaching ionization threshold energy Et and the distance traveled by electron gaining Et. In addition, ionization threshold energy is semi-analytically modeled for GNR. We justify our assumptions using analytic modeling and comparison with simulation results. Gaussian simulator together with analytical modeling is used in order to calculate ionization threshold energy and Kinetic Monte Carlo is employed to calculate ionization coefficient and verify the analytical results. Finally, the profile of ionization is presented using the proposed models and simulation and the results are compared with that of silicon.Keywords: nanostructures, electronic transport, semiconductor modeling, systems engineering
Procedia PDF Downloads 47126866 Multiscale Modelling of Citrus Black Spot Transmission Dynamics along the Pre-Harvest Supply Chain
Authors: Muleya Nqobile, Winston Garira
Abstract:
We presented a compartmental deterministic multi-scale model which encompass internal plant defensive mechanism and pathogen interaction, then we consider nesting the model into the epidemiological model. The objective was to improve our understanding of the transmission dynamics of within host and between host of Guignardia citricapa Kiely. The inflow of infected class was scaled down to individual level while the outflow was scaled up to average population level. Conceptual model and mathematical model were constructed to display a theoretical framework which can be used for predicting or identify disease pattern.Keywords: epidemiological model, mathematical modelling, multi-scale modelling, immunological model
Procedia PDF Downloads 45826865 The Future of Insurance: P2P Innovation versus Traditional Business Model
Authors: Ivan Sosa Gomez
Abstract:
Digitalization has impacted the entire insurance value chain, and the growing movement towards P2P platforms and the collaborative economy is also beginning to have a significant impact. P2P insurance is defined as innovation, enabling policyholders to pool their capital, self-organize, and self-manage their own insurance. In this context, new InsurTech start-ups are emerging as peer-to-peer (P2P) providers, based on a model that differs from traditional insurance. As a result, although P2P platforms do not change the fundamental basis of insurance, they do enable potentially more efficient business models to be established in terms of ensuring the coverage of risk. It is therefore relevant to determine whether p2p innovation can have substantial effects on the future of the insurance sector. For this purpose, it is considered necessary to develop P2P innovation from a business perspective, as well as to build a comparison between a traditional model and a P2P model from an actuarial perspective. Objectives: The objectives are (1) to represent P2P innovation in the business model compared to the traditional insurance model and (2) to establish a comparison between a traditional model and a P2P model from an actuarial perspective. Methodology: The research design is defined as action research in terms of understanding and solving the problems of a collectivity linked to an environment, applying theory and best practices according to the approach. For this purpose, the study is carried out through the participatory variant, which involves the collaboration of the participants, given that in this design, participants are considered experts. For this purpose, prolonged immersion in the field is carried out as the main instrument for data collection. Finally, an actuarial model is developed relating to the calculation of premiums that allows for the establishment of projections of future scenarios and the generation of conclusions between the two models. Main Contributions: From an actuarial and business perspective, we aim to contribute by developing a comparison of the two models in the coverage of risk in order to determine whether P2P innovation can have substantial effects on the future of the insurance sector.Keywords: Insurtech, innovation, business model, P2P, insurance
Procedia PDF Downloads 9226864 Companies’ Internationalization: Multi-Criteria-Based Prioritization Using Fuzzy Logic
Authors: Jorge Anibal Restrepo Morales, Sonia Martín Gómez
Abstract:
A model based on a logical framework was developed to quantify SMEs' internationalization capacity. To do so, linguistic variables, such as human talent, infrastructure, innovation strategies, FTAs, marketing strategies, finance, etc. were integrated. It is argued that a company’s management of international markets depends on internal factors, especially capabilities and resources available. This study considers internal factors as the biggest business challenge because they force companies to develop an adequate set of capabilities. At this stage, importance and strategic relevance have to be defined in order to build competitive advantages. A fuzzy inference system is proposed to model the resources, skills, and capabilities that determine the success of internationalization. Data: 157 linguistic variables were used. These variables were defined by international trade entrepreneurs, experts, consultants, and researchers. Using expert judgment, the variables were condensed into18 factors that explain SMEs’ export capacity. The proposed model is applied by means of a case study of the textile and clothing cluster in Medellin, Colombia. In the model implementation, a general index of 28.2 was obtained for internationalization capabilities. The result confirms that the sector’s current capabilities and resources are not sufficient for a successful integration into the international market. The model specifies the factors and variables, which need to be worked on in order to improve export capability. In the case of textile companies, the lack of a continuous recording of information stands out. Likewise, there are very few studies directed towards developing long-term plans, and., there is little consistency in exports criteria. This method emerges as an innovative management tool linked to internal organizational spheres and their different abilities.Keywords: business strategy, exports, internationalization, fuzzy set methods
Procedia PDF Downloads 29426863 A Location Routing Model for the Logistic System in the Mining Collection Centers of the Northern Region of Boyacá-Colombia
Authors: Erika Ruíz, Luis Amaya, Diego Carreño
Abstract:
The main objective of this study is to design a mathematical model for the logistics of mining collection centers in the northern region of the department of Boyacá (Colombia), determining the structure that facilitates the flow of products along the supply chain. In order to achieve this, it is necessary to define a suitable design of the distribution network, taking into account the products, customer’s characteristics and the availability of information. Likewise, some other aspects must be defined, such as number and capacity of collection centers to establish, routes that must be taken to deliver products to the customers, among others. This research will use one of the operation research problems, which is used in the design of distribution networks known as Location Routing Problem (LRP).Keywords: location routing problem, logistic, mining collection, model
Procedia PDF Downloads 21726862 Test of Capital Account Monetary Model of Floating Exchange Rate Determination: Further Evidence from Selected African Countries
Authors: Oloyede John Adebayo
Abstract:
This paper tested a variant of the monetary model of exchange rate determination, called Frankel’s Capital Account Monetary Model (CAAM) based on Real Interest Rate Differential, on the floating exchange rate experiences of three developing countries of Africa; viz: Ghana, Nigeria and the Gambia. The study adopted the Auto regressive Instrumental Package (AIV) and Almon Polynomial Lag Procedure of regression analysis based on the assumption that the coefficients follow a third-order Polynomial with zero-end constraint. The results found some support for the CAAM hypothesis that exchange rate responds proportionately to changes in money supply, inversely to income and positively to interest rates and expected inflation differentials. On this basis, the study points the attention of monetary authorities and researchers to the relevance and usefulness of CAAM as appropriate tool and useful benchmark for analyzing the exchange rate behaviour of most developing countries.Keywords: exchange rate, monetary model, interest differentials, capital account
Procedia PDF Downloads 41226861 Assessing Readiness Model for Business Intelligence Implementation in Organization
Authors: Abdul Razak Rahmat, Azizah Ahmad, Azman Ta’aa
Abstract:
The deployment of Business Intelligence (BI) for organization at the beginning phase is very crucial. Results from the previous studies found that more than half of the BI project fails to meet the objective even though a lot money are spent. Based on that problem, the readiness level of BI for the organization is important to identify in order to reduce the risk before the actual BI project is implemented. In this paper, rigorous literature review on the aspect success factors such as Critical Success Factors (CSFs), Readiness Factors (RFs), Success Factors (SFs), are discussed by different authors. The paper also adopted a few models from previous study as a guide for the assessment of BI readiness. The expected finding from this research is the Business Intelligent Readiness Model (BiRM) as a guild before implement the BI system.Keywords: business intelligence readiness model, business intelligence for higher learning, BI readiness factors, BI critical success factors(CSF)
Procedia PDF Downloads 37126860 Domain-Specific Deep Neural Network Model for Classification of Abnormalities on Chest Radiographs
Authors: Nkechinyere Joy Olawuyi, Babajide Samuel Afolabi, Bola Ibitoye
Abstract:
This study collected a preprocessed dataset of chest radiographs and formulated a deep neural network model for detecting abnormalities. It also evaluated the performance of the formulated model and implemented a prototype of the formulated model. This was with the view to developing a deep neural network model to automatically classify abnormalities in chest radiographs. In order to achieve the overall purpose of this research, a large set of chest x-ray images were sourced for and collected from the CheXpert dataset, which is an online repository of annotated chest radiographs compiled by the Machine Learning Research Group, Stanford University. The chest radiographs were preprocessed into a format that can be fed into a deep neural network. The preprocessing techniques used were standardization and normalization. The classification problem was formulated as a multi-label binary classification model, which used convolutional neural network architecture to make a decision on whether an abnormality was present or not in the chest radiographs. The classification model was evaluated using specificity, sensitivity, and Area Under Curve (AUC) score as the parameter. A prototype of the classification model was implemented using Keras Open source deep learning framework in Python Programming Language. The AUC ROC curve of the model was able to classify Atelestasis, Support devices, Pleural effusion, Pneumonia, A normal CXR (no finding), Pneumothorax, and Consolidation. However, Lung opacity and Cardiomegaly had a probability of less than 0.5 and thus were classified as absent. Precision, recall, and F1 score values were 0.78; this implies that the number of False Positive and False Negative is the same, revealing some measure of label imbalance in the dataset. The study concluded that the developed model is sufficient to classify abnormalities present in chest radiographs into present or absent.Keywords: transfer learning, convolutional neural network, radiograph, classification, multi-label
Procedia PDF Downloads 12726859 Software Quality Measurement System for Telecommunication Industry in Malaysia
Authors: Nor Fazlina Iryani Abdul Hamid, Mohamad Khatim Hasan
Abstract:
Evolution of software quality measurement has been started since McCall introduced his quality model in year 1977. Starting from there, several software quality models and software quality measurement methods had emerged but none of them focused on telecommunication industry. In this paper, the implementation of software quality measurement system for telecommunication industry was compulsory to accommodate the rapid growth of telecommunication industry. The quality value of the telecommunication related software could be calculated using this system by entering the required parameters. The system would calculate the quality value of the measured system based on predefined quality metrics and aggregated by referring to the quality model. It would classify the quality level of the software based on Net Satisfaction Index (NSI). Thus, software quality measurement system was important to both developers and users in order to produce high quality software product for telecommunication industry.Keywords: software quality, quality measurement, quality model, quality metric, net satisfaction index
Procedia PDF Downloads 592