Search results for: Process models
19794 Investigations of Flow Field with Different Turbulence Models on NREL Phase VI Blade
Authors: T. Y. Liu, C. H. Lin, Y. M. Ferng
Abstract:
Wind energy is one of the clean renewable energy. However, the low frequency (20-200HZ) noise generated from the wind turbine blades, which bothers the residents, becomes the major problem to be developed. It is useful for predicting the aerodynamic noise by flow field and pressure distribution analysis on the wind turbine blades. Therefore, the main objective of this study is to use different turbulence models to analyse the flow field and pressure distributions of the wing blades. Three-dimensional Computation Fluid Dynamics (CFD) simulation of the flow field was used to calculate the flow phenomena for the National Renewable Energy Laboratory (NREL) Phase VI horizontal axis wind turbine rotor. Two different flow cases with different wind speeds were investigated: 7m/s with 72rpm and 15m/s with 72rpm. Four kinds of RANS-based turbulence models, Standard k-ε, Realizable k-ε, SST k-ω, and v2f, were used to predict and analyse the results in the present work. The results show that the predictions on pressure distributions with SST k-ω and v2f turbulence models have good agreements with experimental data.Keywords: horizontal axis wind turbine, turbulence model, noise, fluid dynamics
Procedia PDF Downloads 26519793 Review of the Model-Based Supply Chain Management Research in the Construction Industry
Authors: Aspasia Koutsokosta, Stefanos Katsavounis
Abstract:
This paper reviews the model-based qualitative and quantitative Operations Management research in the context of Construction Supply Chain Management (CSCM). Construction industry has been traditionally blamed for low productivity, cost and time overruns, waste, high fragmentation and adversarial relationships. The construction industry has been slower than other industries to employ the Supply Chain Management (SCM) concept and develop models that support the decision-making and planning. However the last decade there is a distinct shift from a project-based to a supply-based approach of construction management. CSCM comes up as a new promising management tool of construction operations and improves the performance of construction projects in terms of cost, time and quality. Modeling the Construction Supply Chain (CSC) offers the means to reap the benefits of SCM, make informed decisions and gain competitive advantage. Different modeling approaches and methodologies have been applied in the multi-disciplinary and heterogeneous research field of CSCM. The literature review reveals that a considerable percentage of CSC modeling accommodates conceptual or process models which discuss general management frameworks and do not relate to acknowledged soft OR methods. We particularly focus on the model-based quantitative research and categorize the CSCM models depending on their scope, mathematical formulation, structure, objectives, solution approach, software used and decision level. Although over the last few years there has been clearly an increase of research papers on quantitative CSC models, we identify that the relevant literature is very fragmented with limited applications of simulation, mathematical programming and simulation-based optimization. Most applications are project-specific or study only parts of the supply system. Thus, some complex interdependencies within construction are neglected and the implementation of the integrated supply chain management is hindered. We conclude this paper by giving future research directions and emphasizing the need to develop robust mathematical optimization models for the CSC. We stress that CSC modeling needs a multi-dimensional, system-wide and long-term perspective. Finally, prior applications of SCM to other industries have to be taken into account in order to model CSCs, but not without the consequential reform of generic concepts to match the unique characteristics of the construction industry.Keywords: construction supply chain management, modeling, operations research, optimization, simulation
Procedia PDF Downloads 50319792 Climate Change Effects on Agriculture
Authors: Abdellatif Chebboub
Abstract:
Agricultural production is sensitive to weather and thus directly affected by climate change. Plausible estimates of these climate change impacts require combined use of climate, crop, and economic models. Results from previous studies vary substantially due to differences in models, scenarios, and data. This paper is part of a collective effort to systematically integrate these three types of models. We focus on the economic component of the assessment, investigating how nine global economic models of agriculture represent endogenous responses to seven standardized climate change scenarios produced by two climate and five crop models. These responses include adjustments in yields, area, consumption, and international trade. We apply biophysical shocks derived from the Intergovernmental Panel on Climate Change’s representative concentration pathway with end-of-century radiative forcing of 8.5 W/m2. The mean biophysical yield effect with no incremental CO2 fertilization is a 17% reduction globally by 2050 relative to a scenario with unchanging climate. Endogenous economic responses reduce yield loss to 11%, increase area of major crops by 11%, and reduce consumption by 3%. Agricultural production, cropland area, trade, and prices show the greatest degree of variability in response to climate change, and consumption the lowest. The sources of these differences include model structure and specification; in particular, model assumptions about ease of land use conversion, intensification, and trade. This study identifies where models disagree on the relative responses to climate shocks and highlights research activities needed to improve the representation of agricultural adaptation responses to climate change.Keywords: climate change, agriculture, weather change, danger of climate change
Procedia PDF Downloads 31619791 Proposing a Strategic Management Maturity Model for Continues Innovation
Authors: Ferhat Demir
Abstract:
Even if strategic management is highly critical for all types of organizations, only a few maturity models have been proposed in business literature for the area of strategic management activities. This paper updates previous studies and presents a new conceptual model for assessing the maturity of strategic management in any organization. Strategic management maturity model (S-3M) is basically composed of 6 maturity levels with 7 dimensions. The biggest contribution of S-3M is to put innovation into agenda of strategic management. The main objective of this study is to propose a model to align innovation with business strategies. This paper suggests that innovation (breakthrough new products/services and business models) is the only way of creating sustainable growth and strategy studies cannot ignore this aspect. Maturity models should embrace innovation to respond dynamic business environment and rapidly changing customer behaviours.Keywords: strategic management, innovation, business model, maturity model
Procedia PDF Downloads 19419790 Machine Learning Data Architecture
Authors: Neerav Kumar, Naumaan Nayyar, Sharath Kashyap
Abstract:
Most companies see an increase in the adoption of machine learning (ML) applications across internal and external-facing use cases. ML applications vend output either in batch or real-time patterns. A complete batch ML pipeline architecture comprises data sourcing, feature engineering, model training, model deployment, model output vending into a data store for downstream application. Due to unclear role expectations, we have observed that scientists specializing in building and optimizing models are investing significant efforts into building the other components of the architecture, which we do not believe is the best use of scientists’ bandwidth. We propose a system architecture created using AWS services that bring industry best practices to managing the workflow and simplifies the process of model deployment and end-to-end data integration for an ML application. This narrows down the scope of scientists’ work to model building and refinement while specialized data engineers take over the deployment, pipeline orchestration, data quality, data permission system, etc. The pipeline infrastructure is built and deployed as code (using terraform, cdk, cloudformation, etc.) which makes it easy to replicate and/or extend the architecture to other models that are used in an organization.Keywords: data pipeline, machine learning, AWS, architecture, batch machine learning
Procedia PDF Downloads 6419789 DISGAN: Efficient Generative Adversarial Network-Based Method for Cyber-Intrusion Detection
Authors: Hongyu Chen, Li Jiang
Abstract:
Ubiquitous anomalies endanger the security of our system con- stantly. They may bring irreversible damages to the system and cause leakage of privacy. Thus, it is of vital importance to promptly detect these anomalies. Traditional supervised methods such as Decision Trees and Support Vector Machine (SVM) are used to classify normality and abnormality. However, in some case, the abnormal status are largely rarer than normal status, which leads to decision bias of these methods. Generative adversarial network (GAN) has been proposed to handle the case. With its strong generative ability, it only needs to learn the distribution of normal status, and identify the abnormal status through the gap between it and the learned distribution. Nevertheless, existing GAN-based models are not suitable to process data with discrete values, leading to immense degradation of detection performance. To cope with the discrete features, in this paper, we propose an efficient GAN-based model with specifically-designed loss function. Experiment results show that our model outperforms state-of-the-art models on discrete dataset and remarkably reduce the overhead.Keywords: GAN, discrete feature, Wasserstein distance, multiple intermediate layers
Procedia PDF Downloads 12919788 Correlation between Speech Emotion Recognition Deep Learning Models and Noises
Authors: Leah Lee
Abstract:
This paper examines the correlation between deep learning models and emotions with noises to see whether or not noises mask emotions. The deep learning models used are plain convolutional neural networks (CNN), auto-encoder, long short-term memory (LSTM), and Visual Geometry Group-16 (VGG-16). Emotion datasets used are Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), Crowd-sourced Emotional Multimodal Actors Dataset (CREMA-D), Toronto Emotional Speech Set (TESS), and Surrey Audio-Visual Expressed Emotion (SAVEE). To make it four times bigger, audio set files, stretch, and pitch augmentations are utilized. From the augmented datasets, five different features are extracted for inputs of the models. There are eight different emotions to be classified. Noise variations are white noise, dog barking, and cough sounds. The variation in the signal-to-noise ratio (SNR) is 0, 20, and 40. In summation, per a deep learning model, nine different sets with noise and SNR variations and just augmented audio files without any noises will be used in the experiment. To compare the results of the deep learning models, the accuracy and receiver operating characteristic (ROC) are checked.Keywords: auto-encoder, convolutional neural networks, long short-term memory, speech emotion recognition, visual geometry group-16
Procedia PDF Downloads 7519787 An Agent-Based Model of Innovation Diffusion Using Heterogeneous Social Interaction and Preference
Authors: Jang kyun Cho, Jeong-dong Lee
Abstract:
The advent of the Internet, mobile communications, and social network services has stimulated social interactions among consumers, allowing people to affect one another’s innovation adoptions by exchanging information more frequently and more quickly. Previous diffusion models, such as the Bass model, however, face limitations in reflecting such recent phenomena in society. These models are weak in their ability to model interactions between agents; they model aggregated-level behaviors only. The agent based model, which is an alternative to the aggregate model, is good for individual modeling, but it is still not based on an economic perspective of social interactions so far. This study assumes the presence of social utility from other consumers in the adoption of innovation and investigates the effect of individual interactions on innovation diffusion by developing a new model called the interaction-based diffusion model. By comparing this model with previous diffusion models, the study also examines how the proposed model explains innovation diffusion from the perspective of economics. In addition, the study recommends the use of a small-world network topology instead of cellular automata to describe innovation diffusion. This study develops a model based on individual preference and heterogeneous social interactions using utility specification, which is expandable and, thus, able to encompass various issues in diffusion research, such as reservation price. Furthermore, the study proposes a new framework to forecast aggregated-level market demand from individual level modeling. The model also exhibits a good fit to real market data. It is expected that the study will contribute to our understanding of the innovation diffusion process through its microeconomic theoretical approach.Keywords: innovation diffusion, agent based model, small-world network, demand forecasting
Procedia PDF Downloads 34119786 Operating System Based Virtualization Models in Cloud Computing
Authors: Dev Ras Pandey, Bharat Mishra, S. K. Tripathi
Abstract:
Cloud computing is ready to transform the structure of businesses and learning through supplying the real-time applications and provide an immediate help for small to medium sized businesses. The ability to run a hypervisor inside a virtual machine is important feature of virtualization and it is called nested virtualization. In today’s growing field of information technology, many of the virtualization models are available, that provide a convenient approach to implement, but decision for a single model selection is difficult. This paper explains the applications of operating system based virtualization in cloud computing with an appropriate/suitable model with their different specifications and user’s requirements. In the present paper, most popular models are selected, and the selection was based on container and hypervisor based virtualization. Selected models were compared with a wide range of user’s requirements as number of CPUs, memory size, nested virtualization supports, live migration and commercial supports, etc. and we identified a most suitable model of virtualization.Keywords: virtualization, OS based virtualization, container based virtualization, hypervisor based virtualization
Procedia PDF Downloads 32919785 Current of Drain for Various Values of Mobility in the Gaas Mesfet
Authors: S. Belhour, A. K. Ferouani, C. Azizi
Abstract:
In recent years, a considerable effort (experience, numerical simulation, and theoretical prediction models) has characterised by high efficiency and low cost. Then an improved physics analytical model for simulating is proposed. The performance of GaAs MESFETs has been developed for use in device design for high frequency. This model is based on mathematical analysis, and a new approach for the standard model is proposed, this approach allowed to conceive applicable model for MESFET’s operating in the turn-one or pinch-off region and valid for the short-channel and the long channel MESFET’s in which the two dimensional potential distribution contributed by the depletion layer under the gate is obtained by conventional approximation. More ever, comparisons between the analytical models with different values of mobility are proposed, and a good agreement is obtained.Keywords: analytical, gallium arsenide, MESFET, mobility, models
Procedia PDF Downloads 7419784 Investigation of Heat Transfer Mechanism Inside Shell and Tube Latent Heat Thermal Energy Storage Systems
Authors: Saeid Seddegh, Xiaolin Wang, Alan D. Henderson, Dong Chen, Oliver Oims
Abstract:
The main objective of this research is to study the heat transfer processes and phase change behaviour of a phase change material (PCM) in shell and tube latent heat thermal energy storage (LHTES) systems. The thermal behaviour in a vertical and horizontal shell-and-tube heat energy storage system using a pure thermal conduction model and a combined conduction-convection heat transfer model is compared in this paper. The model is first validated using published experimental data available in literature and then used to study the temperature variation, solid-liquid interface, phase distribution, total melting and solidification time during melting and solidification processes of PCMs. The simulated results show that the combined convection and conduction model can better describe the energy transfer in PCMs during melting process. In contrast, heat transfer by conduction is more significant during the solidification process since the two models show little difference. Also, it was concluded that during the charging process for the horizontal orientation, convective heat transfer has a strong effect on melting of the upper part of the solid PCM and is less significant during melting of the lower half of the solid PCM. However, in the vertical orientation, convective heat transfer is the same active during the entire charging process. In the solidification process, the thermal behavior does not show any difference between horizontal and vertical systems.Keywords: latent heat thermal energy storage, phase change material, natural convection, melting, shell and tube heat exchanger, melting, solidification
Procedia PDF Downloads 55419783 Software Tool Design for Heavy Oil Upgrading by Hydrogen Donor Addition in a Hydrodynamic Cavitation Process
Authors: Munoz A. Tatiana, Solano R. Brandon, Montes C. Juan, Cierco G. Javier
Abstract:
The hydrodynamic cavitation is a process in which the energy that the fluids have in the phase changes is used. From this energy, local temperatures greater than 5000 °C are obtained where thermal cracking of the fluid molecules takes place. The process applied to heavy oil affects variables such as viscosity, density, and composition, which constitutes an important improvement in the quality of crude oil. In this study, the need to design a software through mathematical integration models of mixing, cavitation, kinetics, and reactor, allows modeling changes in density, viscosity, and composition of a heavy oil crude, when the fluid passes through a hydrodynamic cavitation reactor. In order to evaluate the viability of this technique in the industry, a heavy oil of 18° API gravity, was simulated using naphtha as a hydrogen donor at concentrations of 1, 2 and 5% vol, where the simulation results showed an API gravity increase to 0.77, 1.21 and 1.93° respectively and a reduction viscosity by 9.9, 12.9 and 15.8%. The obtained results allow to have a favorable panorama on this technological development, an appropriate visualization on the generation of innovative knowledge of this technique and the technical-economic opportunity that benefits the development of the hydrocarbon sector related to heavy crude oil that includes the largest world oil production.Keywords: hydrodynamic cavitation, thermal cracking, hydrogen donor, heavy oil upgrading, simulator
Procedia PDF Downloads 15019782 Landslide Hazard Assessment Using Physically Based Mathematical Models in Agricultural Terraces at Douro Valley in North of Portugal
Authors: C. Bateira, J. Fernandes, A. Costa
Abstract:
The Douro Demarked Region (DDR) is a production Porto wine region. On the NE of Portugal, the strong incision of the Douro valley developed very steep slopes, organized with agriculture terraces, have experienced an intense and deep transformation in order to implement the mechanization of the work. The old terrace system, based on stone vertical wall support structure, replaced by terraces with earth embankments experienced a huge terrace instability. This terrace instability has important economic and financial consequences on the agriculture enterprises. This paper presents and develops cartographic tools to access the embankment instability and identify the area prone to instability. The priority on this evaluation is related to the use of physically based mathematical models and develop a validation process based on an inventory of the past embankment instability. We used the shallow landslide stability model (SHALSTAB) based on physical parameters such us cohesion (c’), friction angle(ф), hydraulic conductivity, soil depth, soil specific weight (ϱ), slope angle (α) and contributing areas by Multiple Flow Direction Method (MFD). A terraced area can be analysed by this models unless we have very detailed information representative of the terrain morphology. The slope angle and the contributing areas depend on that. We can achieve that propose using digital elevation models (DEM) with great resolution (pixel with 40cm side), resulting from a set of photographs taken by a flight at 100m high with pixel resolution of 12cm. The slope angle results from this DEM. In the other hand, the MFD contributing area models the internal flow and is an important element to define the spatial variation of the soil saturation. That internal flow is based on the DEM. That is supported by the statement that the interflow, although not coincident with the superficial flow, have important similitude with it. Electrical resistivity monitoring values which related with the MFD contributing areas build from a DEM of 1m resolution and revealed a consistent correlation. That analysis, performed on the area, showed a good correlation with R2 of 0,72 and 0,76 at 1,5m and 2m depth, respectively. Considering that, a DEM with 1m resolution was the base to model the real internal flow. Thus, we assumed that the contributing area of 1m resolution modelled by MFD is representative of the internal flow of the area. In order to solve this problem we used a set of generalized DEMs to build the contributing areas used in the SHALSTAB. Those DEMs, with several resolutions (1m and 5m), were built from a set of photographs with 50cm resolution taken by a flight with 5km high. Using this maps combination, we modelled several final maps of terrace instability and performed a validation process with the contingency matrix. The best final instability map resembles the slope map from a DEM of 40cm resolution and a MFD map from a DEM of 1m resolution with a True Positive Rate (TPR) of 0,97, a False Positive Rate of 0,47, Accuracy (ACC) of 0,53, Precision (PVC) of 0,0004 and a TPR/FPR ratio of 2,06.Keywords: agricultural terraces, cartography, landslides, SHALSTAB, vineyards
Procedia PDF Downloads 17719781 Functionalized Mesoporous Silica: Absorbents for Water Purification
Authors: Saima Nasreen, Uzaira Rafique, Shery Ehrman, Muhammad Aqeel Ashraf
Abstract:
The release of heavy metals into the environment is a potential threat to water and soil quality as well as to plant, animal and human health. In current research work, organically functionalized mesoporous silicates (MSU-H) were prepared by the co-condensation between sodium silicate and oregano alkoxysilanes in the presence of the nonionic surfactant triblock copolymer P104. The surfactant was used as a template for improving the porosity of the hybrid gels. Synthesized materials were characterized by TEM, FT-IR, SEM/EDX, TG, surface area analysis. The surface morphology and textural properties of such materials varied with various kinds of groups in the channels. In this study, removal of some heavy metals ions from aqueous solution by adsorption process was investigated. Batch adsorption studies show that the adsorption capacity of metal ions on the functionalized silicates is more than that on pure MSU-H. Data shows adsorption on synthesized materials is a time efficient process, suggesting adsorption on external surface as well as the mesoporous process. Adsorption models of Langmuir, Freundlich, and Temkin depicted equal goodness for all adsorbents, whereas pseudo 2nd order kinetics is in best agreement with experimental data.Keywords: heavy metals, mesoporous silica, hybrid, adsorption, freundlich, langmuir, temkin
Procedia PDF Downloads 26819780 Mixed Model Sequencing in Painting Production Line
Authors: Unchalee Inkampa, Tuanjai Somboonwiwat
Abstract:
Painting process of automobiles and automobile parts, which is a continuous process based on EDP (Electrode position paint, EDP). Through EDP, all work pieces will be continuously sent to the painting process. Work process can be divided into 2 groups based on the running time: Painting Room 1 and Painting Room 2. This leads to continuous operation. The problem that arises is waiting for workloads onto Painting Room. The grading process EDP to Painting Room is a major problem. Therefore, this paper aim to develop production sequencing method by applying EDP to painting process. It also applied fixed rate launching for painting room and earliest due date (EDD) for EDP process and swap pairwise interchange for waiting time to a minimum of machine. The result found that the developed method could improve painting reduced waiting time, on time delivery, meeting customers wants and improved productivity of painting unit.Keywords: sequencing, mixed model lines, painting process, electrode position paint
Procedia PDF Downloads 42019779 Analytical Investigation on Seismic Behavior of Infilled Reinforced Concrete Frames Strengthened with Precast Diagonal Concrete Panels
Authors: Ceyhun Aksoylu, Rifat Sezer
Abstract:
In this study, a strengthening method applicable without any evacuation process was investigated. In this analytical study, the pushover analysis results carry out by using the software of SAP2000. For this purpose, 1/3 scaled, 1-bay and 2-story R/C seven frames having usual deficiencies faults produced, one of which were not strengthened, but having brick-infill wall and the other 3 frames with infill walls strengthened with various shaped of high strength-precast diagonal concrete panels. The prepared analytical models investigated under reversed-cyclic loading that resembles the seismic effect. As a result of the analytical study, the properties of the reinforced concrete frames, such as strength, rigidity, energy dissipation capacity, etc. were determined and the strengthened models were compared with the unstrengthened one having the same properties. As a result of this study, the contributions of precast diagonal concrete applied on the infill walls of the existing frame systems against seismic effects were introduced with its advantages and disadvantages.Keywords: RC frame, seismic effect, infill wall, strengthening, precast diagonal concrete panel, pushover analysis
Procedia PDF Downloads 34719778 Trace Logo: A Notation for Representing Control-Flow of Operational Process
Authors: M. V. Manoj Kumar, Likewin Thomas, Annappa
Abstract:
Process mining research discipline bridges the gap between data mining and business process modeling and analysis, it offers the process-centric and end-to-end methods/techniques for analyzing information of real-world process detailed in operational event-logs. In this paper, we have proposed a notation called trace logo for graphically representing control-flow perspective (order of execution of activities) of process. A trace logo consists of a stack of activity names at each position, sizes of the activity name indicates their frequency in the traces and the total height of the activity depicts the information content of the position. A trace logo created from a set of aligned traces generated using Multiple Trace Alignment technique.Keywords: consensus trace, process mining, multiple trace alignment, trace logo
Procedia PDF Downloads 34919777 Rethinking Urban Green Space Quality and Planning Models from Users and Experts’ Perspective for Sustainable Development: The Case of Debre Berhan and Debre Markos Cities, Ethiopia
Authors: Alemaw Kefale, Aramde Fetene, Hayal Desta
Abstract:
This study analyzed the users' and experts' views on the green space quality and planning models in Debre Berhan (DB) and Debre Markos (DM) cities in Ethiopia. A questionnaire survey was conducted on 350 park users (148 from DB and 202 from DM) to rate the accessibility, size, shape, vegetation cover, social and cultural context, conservation and heritage, community participation, attractiveness, comfort, safety, inclusiveness, and maintenance of green spaces using a Likert scale. A key informant interview was held with 13 experts in DB and 12 in DM. Descriptive statistics and tests of independence of variables using the chi-square test were done. A statistically significant association existed between the perception of green space quality attributes and users' occupation (χ² (160, N = 350) = 224.463, p < 0.001), age (χ² (128, N = 350) = 212.812, p < 0.001), gender (χ² (32, N = 350) = 68.443, p < 0.001), and education level (χ² (192, N = 350) = 293.396, p < 0.001). 61.7 % of park users were unsatisfied with the quality of urban green spaces. The users perceived dense vegetation cover as "good," with a mean value of 3.41, while the remaining were perceived as "medium with a mean value of 2.62 – 3.32". Only quantitative space standards are practiced as a green space planning model, while other models are unfamiliar and never used in either city. Therefore, experts need to be aware of and practice urban green models during urban planning to ensure that new developments include green spaces to accommodate the community's and the environment's needs.Keywords: urban green space, quality, users and experts, green space planning models, Ethiopia
Procedia PDF Downloads 5819776 Axiomatic Design of Laser Beam Machining Process
Authors: Nikhil Deshpande, Rahul Mahajan
Abstract:
Laser Beam Machining (LBM) is a non-traditional machining process that has inherent problems like dross, striation, and Heat Affected Zone (HAZ) which reduce the quality of machining. In the present day scenario, these problems are controlled only by iteratively adjusting a large number of process parameters. This paper applies Axiomatic Design principles to design LBM process so as to eliminate the problem of dross and striation and minimize the effect of HAZ. Process parameters and their ranges are proposed to set-up the LBM process, execute the cut and finish the workpiece so as to obtain the best quality cut.Keywords: laser beam machining, dross, striation, heat affected zone, axiomatic design
Procedia PDF Downloads 37019775 Size, Shape, and Compositional Effects on the Order-Disorder Phase Transitions in Au-Cu and Pt-M (M = Fe, Co, and Ni) Nanocluster Alloys
Authors: Forrest Kaatz, Adhemar Bultheel
Abstract:
Au-Cu and Pt-M (M = Fe, Co, and Ni) nanocluster alloys are currently being investigated worldwide by many researchers for their interesting catalytic and nanophase properties. The low-temperature behavior of the phase diagrams is not well understood for alloys with nanometer sizes and shapes. These systems have similar bulk phase diagrams with the L12 (Au3Cu, Pt3M, AuCu3, and PtM3) structurally ordered intermetallics and the L10 structure for the AuCu and PtM intermetallics. We consider three models for low temperature ordering in the phase diagrams of Au–Cu and Pt–M nanocluster alloys. These models are valid for sizes ~ 5 nm and approach bulk values for sizes ~ 20 nm. We study the phase transition in nanoclusters with cubic, octahedral, and cuboctahedral shapes, covering the compositions of interest. These models are based on studying the melting temperatures in nanoclusters using the regular solution, mixing model for alloys. Experimentally, it is extremely challenging to determine thermodynamic data on nano–sized alloys. Reasonable agreement is found between these models and recent experimental data on nanometer clusters in the Au–Cu and Pt–M nanophase systems. From our data, experiments on nanocubes about 5 nm in size, of stoichiometric AuCu and PtM composition, could help differentiate between the models. Some available evidence indicates that ordered intermetallic nanoclusters have better catalytic properties than disordered ones. We conclude with a discussion of physical mechanisms whereby ordering could improve the catalytic properties of nanocluster alloys.Keywords: catalytic reactions, gold nanoalloys, phase transitions, platinum nanoalloys
Procedia PDF Downloads 17619774 Process Modeling of Electric Discharge Machining of Inconel 825 Using Artificial Neural Network
Authors: Himanshu Payal, Sachin Maheshwari, Pushpendra S. Bharti
Abstract:
Electrical discharge machining (EDM), a non-conventional machining process, finds wide applications for shaping difficult-to-cut alloys. Process modeling of EDM is required to exploit the process to the fullest. Process modeling of EDM is a challenging task owing to involvement of so many electrical and non-electrical parameters. This work is an attempt to model the EDM process using artificial neural network (ANN). Experiments were carried out on die-sinking EDM taking Inconel 825 as work material. ANN modeling has been performed using experimental data. The prediction ability of trained network has been verified experimentally. Results indicate that ANN can predict the values of performance measures of EDM satisfactorily.Keywords: artificial neural network, EDM, metal removal rate, modeling, surface roughness
Procedia PDF Downloads 41219773 Elastic and Plastic Collision Comparison Using Finite Element Method
Authors: Gustavo Rodrigues, Hans Weber, Larissa Driemeier
Abstract:
The prevision of post-impact conditions and the behavior of the bodies during the impact have been object of several collision models. The formulation from Hertz’s theory is generally used dated from the 19th century. These models consider the repulsive force as proportional to the deformation of the bodies under contact and may consider it proportional to the rate of deformation. The objective of the present work is to analyze the behavior of the bodies during impact using the Finite Element Method (FEM) with elastic and plastic material models. The main parameters to evaluate are, the contact force, the time of contact and the deformation of the bodies. An advantage of using the FEM approach is the possibility to apply a plastic deformation to the model according to the material definition: there will be used Johnson–Cook plasticity model whose parameters are obtained through empirical tests of real materials. This model allows analyzing the permanent deformation caused by impact, phenomenon observed in real world depending on the forces applied to the body. These results are compared between them and with the model-based Hertz theory.Keywords: collision, impact models, finite element method, Hertz Theory
Procedia PDF Downloads 17519772 Internal Methane Dry Reforming Kinetic Models in Solid Oxide Fuel Cells
Authors: Saeed Moarrefi, Shou-Han Zhou, Liyuan Fan
Abstract:
Coupling with solid oxide fuel cells, methane dry reforming is a promising pathway for energy production while mitigating carbon emissions. However, the influence of carbon dioxide and electrochemical reactions on the internal dry reforming reaction within the fuel cells remains debatable, requiring accurate kinetic models to describe the internal reforming behaviors. We employed the Power-Law and Langmuir Hinshelwood–Hougen Watson models in an electrolyte-supported solid oxide fuel cell with a NiO-GDC-YSZ anode. The current density used in this study ranges from 0 to 1000 A/m2 at 973 K to 1173 K to estimate various kinetic parameters. The influence of the electrochemical reactions on the adsorption terms, the equilibrium of the reactions, the activation energy, the pre-exponential factor of the rate constant, and the adsorption equilibrium constant were studied. This study provides essential parameters for future simulations and highlights the need for a more detailed examination of reforming kinetic models.Keywords: dry reforming kinetics, Langmuir Hinshelwood–Hougen Watson, power-law, SOFC
Procedia PDF Downloads 2219771 Web-Based Cognitive Writing Instruction (WeCWI): A Theoretical-and-Pedagogical e-Framework for Language Development
Authors: Boon Yih Mah
Abstract:
Web-based Cognitive Writing Instruction (WeCWI)’s contribution towards language development can be divided into linguistic and non-linguistic perspectives. In linguistic perspective, WeCWI focuses on the literacy and language discoveries, while the cognitive and psychological discoveries are the hubs in non-linguistic perspective. In linguistic perspective, WeCWI draws attention to free reading and enterprises, which are supported by the language acquisition theories. Besides, the adoption of process genre approach as a hybrid guided writing approach fosters literacy development. Literacy and language developments are interconnected in the communication process; hence, WeCWI encourages meaningful discussion based on the interactionist theory that involves input, negotiation, output, and interactional feedback. Rooted in the e-learning interaction-based model, WeCWI promotes online discussion via synchronous and asynchronous communications, which allows interactions happened among the learners, instructor, and digital content. In non-linguistic perspective, WeCWI highlights on the contribution of reading, discussion, and writing towards cognitive development. Based on the inquiry models, learners’ critical thinking is fostered during information exploration process through interaction and questioning. Lastly, to lower writing anxiety, WeCWI develops the instructional tool with supportive features to facilitate the writing process. To bring a positive user experience to the learner, WeCWI aims to create the instructional tool with different interface designs based on two different types of perceptual learning style.Keywords: WeCWI, literacy discovery, language discovery, cognitive discovery, psychological discovery
Procedia PDF Downloads 56119770 Kou Jump Diffusion Model: An Application to the SP 500; Nasdaq 100 and Russell 2000 Index Options
Authors: Wajih Abbassi, Zouhaier Ben Khelifa
Abstract:
The present research points towards the empirical validation of three options valuation models, the ad-hoc Black-Scholes model as proposed by Berkowitz (2001), the constant elasticity of variance model of Cox and Ross (1976) and the Kou jump-diffusion model (2002). Our empirical analysis has been conducted on a sample of 26,974 options written on three indexes, the S&P 500, Nasdaq 100 and the Russell 2000 that were negotiated during the year 2007 just before the sub-prime crisis. We start by presenting the theoretical foundations of the models of interest. Then we use the technique of trust-region-reflective algorithm to estimate the structural parameters of these models from cross-section of option prices. The empirical analysis shows the superiority of the Kou jump-diffusion model. This superiority arises from the ability of this model to portray the behavior of market participants and to be closest to the true distribution that characterizes the evolution of these indices. Indeed the double-exponential distribution covers three interesting properties that are: the leptokurtic feature, the memory less property and the psychological aspect of market participants. Numerous empirical studies have shown that markets tend to have both overreaction and under reaction over good and bad news respectively. Despite of these advantages there are not many empirical studies based on this model partly because probability distribution and option valuation formula are rather complicated. This paper is the first to have used the technique of nonlinear curve-fitting through the trust-region-reflective algorithm and cross-section options to estimate the structural parameters of the Kou jump-diffusion model.Keywords: jump-diffusion process, Kou model, Leptokurtic feature, trust-region-reflective algorithm, US index options
Procedia PDF Downloads 42919769 A Study on Automotive Attack Database and Data Flow Diagram for Concretization of HEAVENS: A Car Security Model
Authors: Se-Han Lee, Kwang-Woo Go, Gwang-Hyun Ahn, Hee-Sung Park, Cheol-Kyu Han, Jun-Bo Shim, Geun-Chul Kang, Hyun-Jung Lee
Abstract:
In recent years, with the advent of smart cars and the expansion of the market, the announcement of 'Adventures in Automotive Networks and Control Units' at the DEFCON21 conference in 2013 revealed that cars are not safe from hacking. As a result, the HEAVENS model considering not only the functional safety of the vehicle but also the security has been suggested. However, the HEAVENS model only presents a simple process, and there are no detailed procedures and activities for each process, making it difficult to apply it to the actual vehicle security vulnerability check. In this paper, we propose an automated attack database that systematically summarizes attack vectors, attack types, and vulnerable vehicle models to prepare for various car hacking attacks, and data flow diagrams that can detect various vulnerabilities and suggest a way to materialize the HEAVENS model.Keywords: automotive security, HEAVENS, car hacking, security model, information security
Procedia PDF Downloads 36219768 Programming Systems in Implementation of Process Safety at Chemical Process Industry
Authors: Maryam Shayan
Abstract:
Programming frameworks have been utilized as a part of chemical industry process safety operation and configuration to enhance its effectiveness. This paper gives a brief survey and investigation of the best in class and effects of programming frameworks in process security. A study was completed by talking staff accountable for procedure wellbeing practices in the Iranian chemical process industry and diving into writing of innovation for procedure security. This article investigates the useful and operational attributes of programming frameworks for security and endeavors to sort the product as indicated by its level of effect in the administration chain of importance. The study adds to better comprehension of the parts of Information Communication Technology in procedure security, the future patterns and conceivable gaps for innovative work.Keywords: programming frameworks, chemical industry process, process security, administration chain, information communication technology
Procedia PDF Downloads 37319767 Flexible Capacitive Sensors Based on Paper Sheets
Authors: Mojtaba Farzaneh, Majid Baghaei Nejad
Abstract:
This article proposes a new Flexible Capacitive Tactile Sensors based on paper sheets. This method combines the parameters of sensor's material and dielectric, and forms a new model of flexible capacitive sensors. The present article tries to present a practical explanation of this method's application and advantages. With the use of this new method, it is possible to make a more flexibility and accurate sensor in comparison with the current models. To assess the performance of this model, the common capacitive sensor is simulated and the proposed model of this article and one of the existing models are assessed. The results of this article indicate that the proposed model of this article can enhance the speed and accuracy of tactile sensor and has less error in comparison with the current models. Based on the results of this study, it can be claimed that in comparison with the current models, the proposed model of this article is capable of representing more flexibility and more accurate output parameters for touching the sensor, especially in abnormal situations and uneven surfaces, and increases accuracy and practicality.Keywords: capacitive sensor, paper sheets, flexible, tactile, uneven
Procedia PDF Downloads 35319766 Resume Ranking Using Custom Word2vec and Rule-Based Natural Language Processing Techniques
Authors: Subodh Chandra Shakya, Rajendra Sapkota, Aakash Tamang, Shushant Pudasaini, Sujan Adhikari, Sajjan Adhikari
Abstract:
Lots of efforts have been made in order to measure the semantic similarity between the text corpora in the documents. Techniques have been evolved to measure the similarity of two documents. One such state-of-art technique in the field of Natural Language Processing (NLP) is word to vector models, which converts the words into their word-embedding and measures the similarity between the vectors. We found this to be quite useful for the task of resume ranking. So, this research paper is the implementation of the word2vec model along with other Natural Language Processing techniques in order to rank the resumes for the particular job description so as to automate the process of hiring. The research paper proposes the system and the findings that were made during the process of building the system.Keywords: chunking, document similarity, information extraction, natural language processing, word2vec, word embedding
Procedia PDF Downloads 15819765 Deep Learning Approach to Trademark Design Code Identification
Authors: Girish J. Showkatramani, Arthi M. Krishna, Sashi Nareddi, Naresh Nula, Aaron Pepe, Glen Brown, Greg Gabel, Chris Doninger
Abstract:
Trademark examination and approval is a complex process that involves analysis and review of the design components of the marks such as the visual representation as well as the textual data associated with marks such as marks' description. Currently, the process of identifying marks with similar visual representation is done manually in United States Patent and Trademark Office (USPTO) and takes a considerable amount of time. Moreover, the accuracy of these searches depends heavily on the experts determining the trademark design codes used to catalog the visual design codes in the mark. In this study, we explore several methods to automate trademark design code classification. Based on recent successes of convolutional neural networks in image classification, we have used several different convolutional neural networks such as Google’s Inception v3, Inception-ResNet-v2, and Xception net. The study also looks into other techniques to augment the results from CNNs such as using Open Source Computer Vision Library (OpenCV) to pre-process the images. This paper reports the results of the various models trained on year of annotated trademark images.Keywords: trademark design code, convolutional neural networks, trademark image classification, trademark image search, Inception-ResNet-v2
Procedia PDF Downloads 232