Search results for: Leland option pricing models.
1944 The Relationship between Business-model Innovation and Firm Value: A Dynamic Perspective
Authors: Yung C. Ho, Hui C. Fang, Ming J. Hsieh
Abstract:
When consistently innovative business-models can give companies a competitive advantage, longitudinal empirical research, which can reflect dynamic business-model changes, has yet to prove a definitive connection. This study consequently employs a dynamic perspective in conjunction with innovation theory to examine the relationship between the types of business-model innovation and firm value. This study tries to examine various types of business-model innovation in high-end and low-end technology industries such as HTC and the 7-Eleven chain stores with research periods of 14 years and 32 years, respectively. The empirical results suggest that adopting radical business-model innovation in addition to expanding new target markets can successfully lead to a competitive advantage. Sustained advanced technological competences and service/product innovation are the key successful factors in high-end and low-end technology industry business-models respectively. In sum up, the business-model innovation can yield a higher market value and financial value in high-end technology industries than low-end ones.Keywords: Business-model, Dynamic Perspective, Firm Value, Innovation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27431943 Material Parameter Identification of Modified AbdelKarim-Ohno Model
Authors: M. Cermak, T. Karasek, J. Rojicek
Abstract:
The key role in phenomenological modelling of cyclic plasticity is good understanding of stress-strain behaviour of given material. There are many models describing behaviour of materials using numerous parameters and constants. Combination of individual parameters in those material models significantly determines whether observed and predicted results are in compliance. Parameter identification techniques such as random gradient, genetic algorithm and sensitivity analysis are used for identification of parameters using numerical modelling and simulation. In this paper genetic algorithm and sensitivity analysis are used to study effect of 4 parameters of modified AbdelKarim-Ohno cyclic plasticity model. Results predicted by Finite Element (FE) simulation are compared with experimental data from biaxial ratcheting test with semi-elliptical loading path.
Keywords: Genetic algorithm, sensitivity analysis, inverse approach, finite element method, cyclic plasticity, ratcheting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23721942 Optimization of Thermal and Discretization Parameters in Laser Welding Simulation Nd:YAG Applied for Shin Plate Transparent Mode Of DP600
Authors: Chansopheak Seang, Afia David Kouadri, Eric Ragneau
Abstract:
Three dimensional analysis of thermal model in laser full penetration welding, Nd:YAG, by transparent mode DP600 alloy steel 1.25mm of thickness and gap of 0.1mm. Three models studied the influence of thermal dependent temperature properties, thermal independent temperature and the effect of peak value of specific heat at phase transformation temperature, AC1, on the transient temperature. Another seven models studied the influence of discretization, meshes on the temperature distribution in weld plate. It is shown that for the effects of thermal properties, the errors less 4% of maximum temperature in FZ and HAZ have identified. The minimum value of discretization are at least one third increment per radius for temporal discretization and the spatial discretization requires two elements per radius and four elements through thickness of the assembled plate, which therefore represent the minimum requirements of modeling for the laser welding in order to get minimum errors less than 5% compared to the fine mesh.Keywords: FEA, welding, discretization, ABAQUS user subroutine DFLUX
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18171941 Routing Capability and Blocking Analysis of Dynamic ROADM Optical Networks (Category - II) for Dynamic Traffic
Authors: Indumathi T. S., T. Srinivas, B. Siva Kumar
Abstract:
Reconfigurable optical add/drop multiplexers (ROADMs) can be classified into three categories based on their underlying switching technologies. Category I consists of a single large optical switch; category II is composed of a number of small optical switches aligned in parallel; and category III has a single optical switch and only one wavelength being added/dropped. In this paper, to evaluate the wavelength-routing capability of ROADMs of category-II in dynamic optical networks,the dynamic traffic models are designed based on Bernoulli, Poisson distributions for smooth and regular types of traffic. Through Analytical and Simulation results, the routing power of cat-II of ROADM networks for two traffic models are determined.Keywords: Fully-Reconfigurable Optical Add-Drop Multiplexers (FROADMs), Limited Tunability in Reconfigurable Optical Add-Drop multiplexers (LROADM), Multiplexer/De- Multiplexer (MUX/DEMUX), Reconfigurable Optical Add-Drop Multiplexers (ROADMs), Wavelength Division Multiplexing (WDM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15271940 Identification of Aircraft Gas Turbine Engines Temperature Condition
Authors: Pashayev A., Askerov D., C. Ardil, Sadiqov R., Abdullayev P.
Abstract:
Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.
Keywords: Identification of a technical condition, aviation gasturbine engine, fuzzy logic and neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16581939 A Comparison of Artificial Neural Networks for Prediction of Suspended Sediment Discharge in River- A Case Study in Malaysia
Authors: M.R. Mustafa, M.H. Isa, R.B. Rezaur
Abstract:
Prediction of highly non linear behavior of suspended sediment flow in rivers has prime importance in the field of water resources engineering. In this study the predictive performance of two Artificial Neural Networks (ANNs) namely, the Radial Basis Function (RBF) Network and the Multi Layer Feed Forward (MLFF) Network have been compared. Time series data of daily suspended sediment discharge and water discharge at Pari River was used for training and testing the networks. A number of statistical parameters i.e. root mean square error (RMSE), mean absolute error (MAE), coefficient of efficiency (CE) and coefficient of determination (R2) were used for performance evaluation of the models. Both the models produced satisfactory results and showed a good agreement between the predicted and observed data. The RBF network model provided slightly better results than the MLFF network model in predicting suspended sediment discharge.Keywords: ANN, discharge, modeling, prediction, suspendedsediment,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17231938 Predictive Modelling Techniques in Sediment Yield and Hydrological Modelling
Authors: Adesoji T. Jaiyeola, Josiah Adeyemo
Abstract:
This paper presents an extensive review of literature relevant to the modelling techniques adopted in sediment yield and hydrological modelling. Several studies relating to sediment yield are discussed. Many research areas of sedimentation in rivers, runoff and reservoirs are presented. Different types of hydrological models, different methods employed in selecting appropriate models for different case studies are analysed. Applications of evolutionary algorithms and artificial intelligence techniques are discussed and compared especially in water resources management and modelling. This review concentrates on Genetic Programming (GP) and fully discusses its theories and applications. The successful applications of GP as a soft computing technique were reviewed in sediment modelling. Some fundamental issues such as benchmark, generalization ability, bloat, over-fitting and other open issues relating to the working principles of GP are highlighted. This paper concludes with the identification of some research gaps in hydrological modelling and sediment yield.Keywords: Artificial intelligence, evolutionary algorithm, genetic programming, sediment yield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18601937 Data Privacy and Safety with Large Language Models
Authors: Ashly Joseph, Jithu Paulose
Abstract:
Large language models (LLMs) have revolutionized natural language processing capabilities, enabling applications such as chatbots, dialogue agents, image, and video generators. Nevertheless, their trainings on extensive datasets comprising personal information poses notable privacy and safety hazards. This study examines methods for addressing these challenges, specifically focusing on approaches to enhance the security of LLM outputs, safeguard user privacy, and adhere to data protection rules. We explore several methods including post-processing detection algorithms, content filtering, reinforcement learning from human and AI inputs, and the difficulties in maintaining a balance between model safety and performance. The study also emphasizes the dangers of unintentional data leakage, privacy issues related to user prompts, and the possibility of data breaches. We highlight the significance of corporate data governance rules and optimal methods for engaging with chatbots. In addition, we analyze the development of data protection frameworks, evaluate the adherence of LLMs to General Data Protection Regulation (GDPR), and examine privacy legislation in academic and business policies. We demonstrate the difficulties and remedies involved in preserving data privacy and security in the age of sophisticated artificial intelligence by employing case studies and real-life instances. This article seeks to educate stakeholders on practical strategies for improving the security and privacy of LLMs, while also assuring their responsible and ethical implementation.
Keywords: Data privacy, large language models, artificial intelligence, machine learning, cybersecurity, general data protection regulation, data safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1011936 Valorization of Residues from Forest Industry for the Generation of Energy
Authors: M. A. Amezcua-Allieri, E. Torres, J. A. Zermeño Eguía-Lis, M. Magdaleno, L. A. Melgarejo, E. Palmerín, A. Rosas, D. López, J. Aburto
Abstract:
The use of biomass to produce renewable energy is one of the forms that can be used to reduce the impact of energy production. Like any other energy resource, there are limitations for biomass use, and it must compete not only with fossil fuels but also with other renewable energy sources such as solar or wind energy. Combustion is currently the most efficient and widely used waste-to-energy process, in the areas where direct use of biomass is possible, without the need to make large transfers of raw material. Many industrial facilities can use agricultural or forestry waste, straw, chips, bagasse, etc. in their thermal systems without making major transformations or adjustments in the feeding to the ovens, making this waste an attractive and cost-effective option in terms of availability, access, and costs. In spite of the facilities and benefits, the environmental reasons (emission of gases and particulate material) are decisive for its use for energy purpose. This paper describes a valorization of residues from forest industry to generate energy, using a case study.
Keywords: Bioenergy, forest waste, life-cycle assessment, waste-to-energy, electricity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8121935 Identification of Aircraft Gas Turbine Engine's Temperature Condition
Authors: Pashayev A., Askerov D., C. Ardil, Sadiqov R., Abdullayev P.
Abstract:
Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.
Keywords: Identification of a technical condition, aviation gasturbine engine, fuzzy logic and neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16701934 Electroencephalography-Based Intention Recognition and Consensus Assessment during Emergency Response
Abstract:
After natural and man-made disasters, robots can bypass the danger, expedite the search, and acquire unprecedented situational awareness to design rescue plans. Brain-computer interface is a promising option to overcome the limitations of tedious manual control and operation of robots in the urgent search-and-rescue tasks. This study aims to test the feasibility of using electroencephalography (EEG) signals to decode human intentions and detect the level of consensus on robot-provided information. EEG signals were classified using machine-learning and deep-learning methods to discriminate search intentions and agreement perceptions. The results show that the average classification accuracy for intention recognition and consensus assessment is 67% and 72%, respectively, proving the potential of incorporating recognizable users’ bioelectrical responses into advanced robot-assisted systems for emergency response.
Keywords: Consensus assessment, electroencephalogram, EEG, emergency response, human-robot collaboration, intention recognition, search and rescue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3431933 Development of a Non-invasive System to Measure the Thickness of the Subcutaneous Adipose Tissue Layer for Human
Authors: Hyuck Ki Hong, Young Chang Jo, Yeon Shik Choi, Beom Joon Kim, Hyo Derk Park
Abstract:
To measure the thickness of the subcutaneous adipose tissue layer, a non-invasive optical measurement system (λ=1300 nm) is introduced. Animal and human subjects are used for the experiments. The results of human subjects are compared with the data of ultrasound device measurements, and a high correlation (r=0.94 for n=11) is observed. There are two modes in the corresponding signals measured by the optical system, which can be explained by two-layered and three-layered tissue models. If the target tissue is thinner than the critical thickness, detected data using diffuse reflectance method follow the three-layered tissue model, so the data increase as the thickness increases. On the other hand, if the target tissue is thicker than the critical thickness, the data follow the two-layered tissue model, so they decrease as the thickness increases.Keywords: Subcutaneous adipose tissue layer, non-invasive measurement system, two-layered and three-layered tissue models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18451932 Energy Communities from Municipality Level to Province Level: A Comparison Using Autoregressive Integrated Moving Average Model
Authors: Amro Issam Hamed Attia Ramadan, Marco Zappatore, Pasquale Balena, Antonella Longo
Abstract:
Considering the energy crisis that is hitting Europe, it becomes increasingly necessary to change energy policies to depend less on fossil fuels and replace them with energy from renewable sources. This has triggered the urge to use clean energy, not only to satisfy energy needs and fulfill the required consumption, but also to decrease the danger of climatic changes due to harmful emissions. Many countries have already started creating energy communities based on renewable energy sources. The first step to understanding energy needs in any place is to perfectly know the consumption. In this work, we aim to estimate electricity consumption for a municipality that makes up part of a rural area located in southern Italy using forecast models that allow for the estimation of electricity consumption for the next 10 years, and we then apply the same model to the province where the municipality is located and estimate the future consumption for the same period to examine whether it is possible to start from the municipality level to reach the province level when creating energy communities.
Keywords: ARIMA, electricity consumption, forecasting models, time series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2821931 Relative Mapping Errors of Linear Time Invariant Systems Caused By Particle Swarm Optimized Reduced Order Model
Authors: G. Parmar, S. Mukherjee, R. Prasad
Abstract:
The authors present an optimization algorithm for order reduction and its application for the determination of the relative mapping errors of linear time invariant dynamic systems by the simplified models. These relative mapping errors are expressed by means of the relative integral square error criterion, which are determined for both unit step and impulse inputs. The reduction algorithm is based on minimization of the integral square error by particle swarm optimization technique pertaining to a unit step input. The algorithm is simple and computer oriented. It is shown that the algorithm has several advantages, e.g. the reduced order models retain the steady-state value and stability of the original system. Two numerical examples are solved to illustrate the superiority of the algorithm over some existing methods.Keywords: Order reduction, Particle swarm optimization, Relative mapping error, Stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15731930 IoT Device Cost Effective Storage Architecture and Real-Time Data Analysis/Data Privacy Framework
Authors: Femi Elegbeleye, Seani Rananga
Abstract:
This paper focused on cost effective storage architecture using fog and cloud data storage gateway, and presented the design of the framework for the data privacy model and data analytics framework on a real-time analysis when using machine learning method. The paper began with the system analysis, system architecture and its component design, as well as the overall system operations. Several results obtained from this study on data privacy models show that when two or more data privacy models are integrated via a fog storage gateway, we often have more secure data. Our main focus in the study is to design a framework for the data privacy model, data storage, and real-time analytics. This paper also shows the major system components and their framework specification. And lastly, the overall research system architecture was shown, including its structure, and its interrelationships.
Keywords: IoT, fog storage, cloud storage, data analysis, data privacy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2411929 Regret, Choice, and Outcome
Authors: Olimpia Matarazzo, Lucia Abbamonte
Abstract:
In two studies we challenged the well consolidated position in regret literature according to which the necessary condition for the emergence of regret is a bad outcome ensuing from free decisions. Without free choice, and, consequently, personal responsibility, other emotions, such as disappointment, but not regret, are supposed to be elicited. In our opinion, a main source of regret is being obliged by circumstance out of our control to chose an undesired option. We tested the hypothesis that regret resulting from a forced choice is more intense than regret derived from a free choice and that the outcome affects the latter, not the former. Besides, we investigated whether two other variables – the perception of the level of freedom of the choice and the choice justifiability – mediated the relationships between choice and regret, as well as the other four emotions we examined: satisfaction, anger toward oneself, disappointment, anger towards circumstances. The two studies were based on the scenario methodology and implied a 2 x 2 (choice x outcome) between design. In the first study the foreseen short-term effects of the choice were assessed; in the second study the experienced long-term effects of the choice were assessed. In each study 160 students of the Second University of Naples participated. Results largely corroborated our hypotheses. They were discussed in the light of the main theories on regret and decision making.Keywords: Choice, outcome, regret.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16211928 CFD Analysis of Incompressible Turbulent Swirling Flow through Circle Grids Space Filling Plate
Authors: B. Manshoor, M. Jaat, Amir Khalid
Abstract:
Circle grid space filling plate is a flow conditioner with a fractal pattern and used to eliminate turbulence originating from pipe fittings in experimental fluid flow applications. In this paper, steady state, incompressible, swirling turbulent flow through circle grid space filling plate has been studied. The solution and the analysis were carried out using finite volume CFD solver FLUENT 6.2. Three turbulence models were used in the numerical investigation and their results were compared with the pressure drop correlation of BS EN ISO 5167-2:2003. The turbulence models investigated here are the standard k-ε, realizable k-ε, and the Reynolds Stress Model (RSM). The results showed that the RSM model gave the best agreement with the ISO pressure drop correlation. The effects of circle grids space filling plate thickness and Reynolds number on the flow characteristics have been investigated as well.
Keywords: Flow conditioning, turbulent flow, turbulent modeling, CFD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20751927 Reduced Order Modelling of Linear Dynamic Systems using Particle Swarm Optimized Eigen Spectrum Analysis
Authors: G. Parmar, S. Mukherjee, R. Prasad
Abstract:
The authors present an algorithm for order reduction of linear time invariant dynamic systems using the combined advantages of the eigen spectrum analysis and the error minimization by particle swarm optimization technique. Pole centroid and system stiffness of both original and reduced order systems remain same in this method to determine the poles, whereas zeros are synthesized by minimizing the integral square error in between the transient responses of original and reduced order models using particle swarm optimization technique, pertaining to a unit step input. It is shown that the algorithm has several advantages, e.g. the reduced order models retain the steady-state value and stability of the original system. The algorithm is illustrated with the help of two numerical examples and the results are compared with the other existing techniques.Keywords: Eigen spectrum, Integral square error, Orderreduction, Particle swarm optimization, Stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16611926 Competitors’ Influence Analysis of a Retailer by Using Customer Value and Huff’s Gravity Model
Authors: Yepeng Cheng, Yasuhiko Morimoto
Abstract:
Customer relationship analysis is vital for retail stores, especially for supermarkets. The point of sale (POS) systems make it possible to record the daily purchasing behaviors of customers as an identification point of sale (ID-POS) database, which can be used to analyze customer behaviors of a supermarket. The customer value is an indicator based on ID-POS database for detecting the customer loyalty of a store. In general, there are many supermarkets in a city, and other nearby competitor supermarkets significantly affect the customer value of customers of a supermarket. However, it is impossible to get detailed ID-POS databases of competitor supermarkets. This study firstly focused on the customer value and distance between a customer's home and supermarkets in a city, and then constructed the models based on logistic regression analysis to analyze correlations between distance and purchasing behaviors only from a POS database of a supermarket chain. During the modeling process, there are three primary problems existed, including the incomparable problem of customer values, the multicollinearity problem among customer value and distance data, and the number of valid partial regression coefficients. The improved customer value, Huff’s gravity model, and inverse attractiveness frequency are considered to solve these problems. This paper presents three types of models based on these three methods for loyal customer classification and competitors’ influence analysis. In numerical experiments, all types of models are useful for loyal customer classification. The type of model, including all three methods, is the most superior one for evaluating the influence of the other nearby supermarkets on customers' purchasing of a supermarket chain from the viewpoint of valid partial regression coefficients and accuracy.Keywords: Customer value, Huff's Gravity Model, POS, retailer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6111925 Design Optimization of Cutting Parameters when Turning Inconel 718 with Cermet Inserts
Authors: M. Aruna, V. Dhanalaksmi
Abstract:
Inconel 718, a nickel based super-alloy is an extensively used alloy, accounting for about 50% by weight of materials used in an aerospace engine, mainly in the gas turbine compartment. This is owing to their outstanding strength and oxidation resistance at elevated temperatures in excess of 5500 C. Machining is a requisite operation in the aircraft industries for the manufacture of the components especially for gas turbines. This paper is concerned with optimization of the surface roughness when turning Inconel 718 with cermet inserts. Optimization of turning operation is very useful to reduce cost and time for machining. The approach is based on Response Surface Method (RSM). In this work, second-order quadratic models are developed for surface roughness, considering the cutting speed, feed rate and depth of cut as the cutting parameters, using central composite design. The developed models are used to determine the optimum machining parameters. These optimized machining parameters are validated experimentally, and it is observed that the response values are in reasonable agreement with the predicted values.Keywords: Inconel 718, Optimization, Response Surface Methodology (RSM), Surface roughness
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28371924 Travel Time Evaluation of an Innovative U-Turn Facility on Urban Arterial Roadways
Authors: Ali Pirdavani, Tom Brijs, Tom Bellemans, Geert Wets, Koen Vanhoof
Abstract:
Signalized intersections on high-volume arterials are often congested during peak hours, causing a decrease in through movement efficiency on the arterial. Much of the vehicle delay incurred at conventional intersections is caused by high left-turn demand. Unconventional intersection designs attempt to reduce intersection delay and travel time by rerouting left-turns away from the main intersection and replacing it with right-turn followed by Uturn. The proposed new type of U-turn intersection is geometrically designed with a raised island which provides a protected U-turn movement. In this study several scenarios based on different distances between U-turn and main intersection, traffic volume of major/minor approaches and percentage of left-turn volumes were simulated by use of AIMSUN, a type of traffic microsimulation software. Subsequently some models are proposed in order to compute travel time of each movement. Eventually by correlating these equations to some in-field collected data of some implemented U-turn facilities, the reliability of the proposed models are approved. With these models it would be possible to calculate travel time of each movement under any kind of geometric and traffic condition. By comparing travel time of a conventional signalized intersection with U-turn intersection travel time, it would be possible to decide on converting signalized intersections into this new kind of U-turn facility or not. However comparison of travel time is not part of the scope of this research. In this paper only travel time of this innovative U-turn facility would be predicted. According to some before and after study about the traffic performance of some executed U-turn facilities, it is found that commonly, this new type of U-turn facility produces lower travel time. Thus, evaluation of using this type of unconventional intersection should be seriously considered.Keywords: Innovative U-turn facility, Microsimulation, Traveltime, Unconventional intersection design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13401923 Effect of Friction Models on Stress Distribution of Sheet Materials during V-Bending Process
Authors: Maziar Ramezani, Zaidi Mohd Ripin
Abstract:
In a metal forming process, the friction between the material and the tools influences the process by modifying the stress distribution of the workpiece. This frictional behaviour is often taken into account by using a constant coefficient of friction in the finite element simulations of sheet metal forming processes. However, friction coefficient varies in time and space with many parameters. The Stribeck friction model is investigated in this study to predict springback behaviour of AA6061-T4 sheets during V-bending process. The coefficient of friction in Stribeck curve depends on sliding velocity and contact pressure. The plane-strain bending process is simulated in ABAQUS/Standard. We compared the computed punch load-stroke curves and springback related to the constant coefficient of friction with the defined friction model. The results clearly showed that the new friction model provides better agreement between experiments and results of numerical simulations. The influence of friction models on stress distribution in the workpiece is also studied numericallyKeywords: Friction model, Stress distribution, V-bending.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27381922 Proffering a Brand New Methodology to Resource Discovery in Grid based on Economic Criteria Using Learning Automata
Authors: Ali Sarhadi, Mohammad Reza Meybodi, Ali Yousefi
Abstract:
Resource discovery is one of the chief services of a grid. A new approach to discover the provenances in grid through learning automata has been propounded in this article. The objective of the aforementioned resource-discovery service is to select the resource based upon the user-s applications and the mercantile yardsticks that is to say opting for an originator which can accomplish the user-s tasks in the most economic manner. This novel service is submitted in two phases. We proffered an applicationbased categorization by means of an intelligent nerve-prone plexus. The user in question sets his or her application as the input vector of the nerve-prone nexus. The output vector of the aforesaid network limns the appropriateness of any one of the resource for the presented executive procedure. The most scrimping option out of those put forward in the previous stage which can be coped with to fulfill the task in question is picked out. Te resource choice is carried out by means of the presented algorithm based upon the learning automata.
Keywords: Resource discovery, learning automata, neural network, economic policy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14521921 Estimation of the Drought Index Based on the Climatic Projections of Precipitation of the Uruguay River Basin
Authors: José Leandro Melgar Néris, Claudinéia Brazil, Luciane Teresa Salvi, Isabel Cristina Damin
Abstract:
The impact the climate change is not recent, the main variable in the hydrological cycle is the sequence and shortage of a drought, which has a significant impact on the socioeconomic, agricultural and environmental spheres. This study aims to characterize and quantify, based on precipitation climatic projections, the rainy and dry events in the region of the Uruguay River Basin, through the Standardized Precipitation Index (SPI). The database is the image that is part of the Intercomparison of Model Models, Phase 5 (CMIP5), which provides condition prediction models, organized according to the Representative Routes of Concentration (CPR). Compared to the normal set of climates in the Uruguay River Watershed through precipitation projections, seasonal precipitation increases for all proposed scenarios, with a low climate trend. From the data of this research, the idea is that this article can be used to support research and the responsible bodies can use it as a subsidy for mitigation measures in other hydrographic basins.
Keywords: Drought index, climatic projections, precipitation of the Uruguay River Basin, Standardized Precipitation Index.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5951920 Using Radial Basis Function Neural Networks to Calibrate Water Quality Model
Authors: Lihui Ma, Kunlun Xin, Suiqing Liu
Abstract:
Modern managements of water distribution system (WDS) need water quality models that are able to accurately predict the dynamics of water quality variations within the distribution system environment. Before water quality models can be applied to solve system problems, they should be calibrated. Although former researchers use GA solver to calibrate relative parameters, it is difficult to apply on the large-scale or medium-scale real system for long computational time. In this paper a new method is designed which combines both macro and detailed model to optimize the water quality parameters. This new combinational algorithm uses radial basis function (RBF) metamodeling as a surrogate to be optimized for the purpose of decreasing the times of time-consuming water quality simulation and can realize rapidly the calibration of pipe wall reaction coefficients of chlorine model of large-scaled WDS. After two cases study this method is testified to be more efficient and promising, and deserve to generalize in the future.Keywords: Metamodeling, model calibration, radial basisfunction, water distribution system, water quality model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20201919 Improved Rare Species Identification Using Focal Loss Based Deep Learning Models
Authors: Chad Goldsworthy, B. Rajeswari Matam
Abstract:
The use of deep learning for species identification in camera trap images has revolutionised our ability to study, conserve and monitor species in a highly efficient and unobtrusive manner, with state-of-the-art models achieving accuracies surpassing the accuracy of manual human classification. The high imbalance of camera trap datasets, however, results in poor accuracies for minority (rare or endangered) species due to their relative insignificance to the overall model accuracy. This paper investigates the use of Focal Loss, in comparison to the traditional Cross Entropy Loss function, to improve the identification of minority species in the “255 Bird Species” dataset from Kaggle. The results show that, although Focal Loss slightly decreased the accuracy of the majority species, it was able to increase the F1-score by 0.06 and improve the identification of the bottom two, five and ten (minority) species by 37.5%, 15.7% and 10.8%, respectively, as well as resulting in an improved overall accuracy of 2.96%.
Keywords: Convolutional neural networks, data imbalance, deep learning, focal loss, species classification, wildlife conservation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14181918 Improvement of Passengers Ride Comfort in Rail Vehicles Equipped with Air Springs
Authors: H. Sayyaadi, N. Shokouhi
Abstract:
In rail vehicles, air springs are very important isolating component, which guarantee good ride comfort for passengers during their trip. In the most new rail–vehicle models, developed by researchers, the thermo–dynamical effects of air springs are ignored and secondary suspension is modeled by simple springs and dampers. As the performance of suspension components have significant effects on rail–vehicle dynamics and ride comfort of passengers, a complete nonlinear thermo–dynamical air spring model, which is a combination of two different models, is introduced. Result from field test shows remarkable agreement between proposed model and experimental data. Effects of air suspension parameters on the system performances are investigated here and then these parameters are tuned to minimize Sperling ride comfort index during the trip. Results showed that by modification of air suspension parameters, passengers comfort is improved and ride comfort index is reduced about 10%.
Keywords: Air spring, Ride comfort improvement, Thermo– dynamical effects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31221917 Logistics Model for Improving Quality in Railway Transport
Authors: Eva Nedeliakova, Juraj Camaj, Jaroslav Masek
Abstract:
This contribution is focused on the methodology for identifying levels of quality and improving quality through new logistics model in railway transport. It is oriented on the application of dynamic quality models, which represent an innovative method of evaluation quality services. Through this conception, time factor, expected, and perceived quality in each moment of the transportation process within logistics chain can be taken into account. Various models describe the improvement of the quality which emphases the time factor throughout the whole transportation logistics chain. Quality of services in railway transport can be determined by the existing level of service quality, by detecting the causes of dissatisfaction employees but also customers, to uncover strengths and weaknesses. This new logistics model is able to recognize critical processes in logistic chain. It includes service quality rating that must respect its specific properties, which are unrepeatability, impalpability, their use right at the time they are provided and particularly changeability, which is significant factor in the conditions of rail transport as well. These peculiarities influence the quality of service regarding the constantly increasing requirements and that result in new ways of finding progressive attitudes towards the service quality rating.Keywords: Logistics model, quality, railway transport.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18271916 Numerical Simulation of Axially Loaded to Failure Large Diameter Bored Pile
Authors: M. Ezzat, Y. Zaghloul, T. Sorour, A. Hefny, M. Eid
Abstract:
Ultimate capacity of large diameter bored piles is usually determined from pile loading tests as recommended by several international codes and foundation design standards. However, loading of this type of piles till achieving apparent failure is practically seldom. In this paper, numerical analyses are carried out to simulate load test of a large diameter bored pile performed at the location of Alzey highway bridge project (Germany). Test results of pile load settlement relationship till failure as well as results of the base and shaft resistances are available. Apparent failure was indicated in this test by the significant increase of the induced settlement during the last load increment applied on the pile head. Measurements of this pile load test are used to assess the quality of the numerical models investigated. Three different material soil models are implemented in the analyses: Mohr coulomb (MC), Soft soil (SS), and Modified Mohr coulomb (MMC). Very good agreement is obtained between the field measured settlement and the calculated settlement using the MMC model. Results of analysis showed also that the MMC constitutive model is superior to MC, and SS models in predicting the ultimate base and shaft resistances of the large diameter bored pile. After calibrating the numerical model, behavior of large diameter bored piles under axial loads is discussed and the formation of the plastic zone around the pile is explored. Results obtained showed that the plastic zone below the base of the pile at failure extended laterally to about four times the pile diameter and vertically to about three times the pile diameter.
Keywords: Ultimate capacity, large diameter bored piles, plastic zone, failure, pile load test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9181915 When Explanations “Cause“ Error: A Look at Representations and Compressions
Authors: Michael Lissack
Abstract:
We depend upon explanation in order to “make sense" out of our world. And, making sense is all the more important when dealing with change. But, what happens if our explanations are wrong? This question is examined with respect to two types of explanatory model. Models based on labels and categories we shall refer to as “representations." More complex models involving stories, multiple algorithms, rules of thumb, questions, ambiguity we shall refer to as “compressions." Both compressions and representations are reductions. But representations are far more reductive than compressions. Representations can be treated as a set of defined meanings – coherence with regard to a representation is the degree of fidelity between the item in question and the definition of the representation, of the label. By contrast, compressions contain enough degrees of freedom and ambiguity to allow us to make internal predictions so that we may determine our potential actions in the possibility space. Compressions are explanatory via mechanism. Representations are explanatory via category. Managers are often confusing their evocation of a representation (category inclusion) as the creation of a context of compression (description of mechanism). When this type of explanatory error occurs, more errors follow. In the drive for efficiency such substitutions are all too often proclaimed – at the manager-s peril..Keywords: Coherence, Emergence, Reduction, Model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1238