Search results for: model based engineering MBE
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 38965

Search results for: model based engineering MBE

32065 Alloy Design of Single Crystal Ni-base Superalloys by Combined Method of Neural Network and CALPHAD

Authors: Mehdi Montakhabrazlighi, Ercan Balikci

Abstract:

The neural network (NN) method is applied to alloy development of single crystal Ni-base Superalloys with low density and improved mechanical strength. A set of 1200 dataset which includes chemical composition of the alloys, applied stress and temperature as inputs and density and time to rupture as outputs is used for training and testing the network. Thermodynamic phase diagram modeling of the screened alloys is performed with Thermocalc software to model the equilibrium phases and also microsegregation in solidification processing. The model is first trained by 80% of the data and the 20% rest is used to test it. Comparing the predicted values and the experimental ones showed that a well-trained network is capable of accurately predicting the density and time to rupture strength of the Ni-base superalloys. Modeling results is used to determine the effect of alloying elements, stress, temperature and gamma-prime phase volume fraction on rupture strength of the Ni-base superalloys. This approach is in line with the materials genome initiative and integrated computed materials engineering approaches promoted recently with the aim of reducing the cost and time for development of new alloys for critical aerospace components. This work has been funded by TUBITAK under grant number 112M783.

Keywords: neural network, rupture strength, superalloy, thermocalc

Procedia PDF Downloads 310
32064 Calibration of the Radical Installation Limit Error of the Accelerometer in the Gravity Gradient Instrument

Authors: Danni Cong, Meiping Wu, Xiaofeng He, Junxiang Lian, Juliang Cao, Shaokuncai, Hao Qin

Abstract:

Gravity gradient instrument (GGI) is the core of the gravity gradiometer, so the structural error of the sensor has a great impact on the measurement results. In order not to affect the aimed measurement accuracy, limit error is required in the installation of the accelerometer. In this paper, based on the established measuring principle model, the radial installation limit error is calibrated, which is taken as an example to provide a method to calculate the other limit error of the installation under the premise of ensuring the accuracy of the measurement result. This method provides the idea for deriving the limit error of the geometry structure of the sensor, laying the foundation for the mechanical precision design and physical design.

Keywords: gravity gradient sensor, radial installation limit error, accelerometer, uniaxial rotational modulation

Procedia PDF Downloads 417
32063 Cryptanalysis of ID-Based Deniable Authentication Protocol Based On Diffie-Hellman Problem on Elliptic Curve

Authors: Eun-Jun Yoon

Abstract:

Deniable authentication protocol is a new security authentication mechanism which can enable a receiver to identify the true source of a given message, but not to prove the identity of the sender to a third party. In 2013, Kar proposed a secure ID-based deniable authentication protocol whose security is based on computational infeasibility of solving Elliptic Curve Diffie-Hellman Problem (ECDHP). Kar claimed that the proposed protocol achieves properties of deniable authentication, mutual authentication, and message confidentiality. However, this paper points out that Kar's protocol still suffers from sender spoofing attack and message modification attack unlike its claims.

Keywords: deniable authentication, elliptic curve cryptography, Diffie-Hellman problem, cryptanalysis

Procedia PDF Downloads 327
32062 Project Progress Prediction in Software Devlopment Integrating Time Prediction Algorithms and Large Language Modeling

Authors: Dong Wu, Michael Grenn

Abstract:

Managing software projects effectively is crucial for meeting deadlines, ensuring quality, and managing resources well. Traditional methods often struggle with predicting project timelines accurately due to uncertain schedules and complex data. This study addresses these challenges by combining time prediction algorithms with Large Language Models (LLMs). It makes use of real-world software project data to construct and validate a model. The model takes detailed project progress data such as task completion dynamic, team Interaction and development metrics as its input and outputs predictions of project timelines. To evaluate the effectiveness of this model, a comprehensive methodology is employed, involving simulations and practical applications in a variety of real-world software project scenarios. This multifaceted evaluation strategy is designed to validate the model's significant role in enhancing forecast accuracy and elevating overall management efficiency, particularly in complex software project environments. The results indicate that the integration of time prediction algorithms with LLMs has the potential to optimize software project progress management. These quantitative results suggest the effectiveness of the method in practical applications. In conclusion, this study demonstrates that integrating time prediction algorithms with LLMs can significantly improve the predictive accuracy and efficiency of software project management. This offers an advanced project management tool for the industry, with the potential to improve operational efficiency, optimize resource allocation, and ensure timely project completion.

Keywords: software project management, time prediction algorithms, large language models (LLMS), forecast accuracy, project progress prediction

Procedia PDF Downloads 71
32061 Impact of Tablet Based Learning on Continuous Assessment (ESPRIT Smart School Framework)

Authors: Mehdi Attia, Sana Ben Fadhel, Lamjed Bettaieb

Abstract:

Mobile technology has become a part of our daily lives and assist learners (despite their level and age) in their leaning process using various apparatus and mobile devices (laptop, tablets, etc.). This paper presents a new learning framework based on tablets. This solution has been developed and tested in ESPRIT “Ecole Supérieure Privée d’Igénieurie et de Technologies”, a Tunisian school of engineering. This application is named ESSF: Esprit Smart School Framework. In this work, the main features of the proposed solution are listed, particularly its impact on the learners’ evaluation process. Learner’s assessment has always been a critical component of the learning process as it measures students’ knowledge. However, traditional evaluation methods in which the learner is evaluated once or twice each year cannot reflect his real level. This is why a continuous assessment (CA) process becomes necessary. In this context we have proved that ESSF offers many important features that enhance and facilitate the implementation of the CA process.

Keywords: continuous assessment, mobile learning, tablet based learning, smart school, ESSF

Procedia PDF Downloads 332
32060 Investigating the English Speech Processing System of EFL Japanese Older Children

Authors: Hiromi Kawai

Abstract:

This study investigates the nature of EFL older children’s L2 perceptive and productive abilities using classroom data, in order to find a pedagogical solution to the teaching of L2 sounds at an early stage of learning in a formal school setting. It is still inconclusive whether older children with only EFL formal school instruction at the initial stage of L2 learning are able to attain native-like perception and production in English within the very limited amount of exposure to the target language available. Based on the notion of the lack of study of EFL Japanese children’s acquisition of English segments, the researcher uses a model of L1 speech processing which was developed for investigating L1 English children’s speech and literacy difficulties using a psycholinguistic framework. The model is composed of input channel, output channel, and lexical representation, and examines how a child receives information from spoken or written language, remembers and stores it within the lexical representations and how the child selects and produces spoken or written words. Concerning language universality and language specificity in the language acquisitional process, the aim of finding any sound errors in L1 English children seemed to conform to the author’s intention to find abilities of English sounds in older Japanese children at the novice level of English in an EFL setting. 104 students in Grade 5 (between the ages of 10 and 11 years old) of an elementary school in Tokyo participated in this study. Four tests to measure their perceptive ability and three oral repetition tests to measure their productive ability were conducted with/without reference to lexical representation. All the test items were analyzed to calculate item facility (IF) indices, and correlational analyses and Structural Equation Modeling (SEM) were conducted to examine the relationship between the receptive ability and the productive ability. IF analysis showed that (1) the participants were better at perceiving a segment than producing a segment, (2) they had difficulty in auditory discrimination of paired consonants when one of them does not exist in the Japanese inventory, (3) they had difficulty in both perceiving and producing English vowels, and (4) their L1 loan word knowledge had an influence on their ability to perceive and produce L2 sounds. The result of the Multiple Regression Modeling showed that the two production tests could predict the participants’ auditory ability of real words in English. The result of SEM showed that the hypothesis that perceptive ability affects productive ability was supported. Based on these findings, the author discusses the possible explicit method of teaching English segments to EFL older children in a formal school setting.

Keywords: EFL older children, english segments, perception, production, speech processing system

Procedia PDF Downloads 240
32059 The Creation of a Yeast Model for 5-oxoproline Accumulation

Authors: Pratiksha Dubey, Praveen Singh, Shantanu Sen Gupta, Anand K. Bachhawat

Abstract:

5-oxoproline (pyroglutamic acid) is a cyclic lactam of glutamic acid. In the cell, it can be produced by several different pathways and is metabolized into glutamate with the help of the 5-oxoprolinase enzyme (OPLAH or OXP1). The inhibition of 5-oxoprolinase enzyme in mammals was found to result in heart failure and is thought to be a consequence of oxidative stress [1]. To analyze the consequences of 5-oxoproline accumulation more clearly, we are generating models for 5-oxoproline accumulation in yeast. The 5-oxoproline accumulation model in yeast is being developed by two different strategies. The first one is by overexpression of the mouse  -glutamylcyclotransferase enzyme. It degrades -glu-met dipeptide into 5-oxoproline and methionine taken by the cell from the medium. The second strategy is by providing high concentration of 5-oxoproline externally to the yeast cells. The intracellular 5-oxoproline levels in both models are being evaluated. In addition, the metabolic and cellular consequences are being investigated.

Keywords: 5-oxoproline, pyroglutamic acid, yeast, genetics

Procedia PDF Downloads 80
32058 Mask-Prompt-Rerank: An Unsupervised Method for Text Sentiment Transfer

Authors: Yufen Qin

Abstract:

Text sentiment transfer is an important branch of text style transfer. The goal is to generate text with another sentiment attribute based on a text with a specific sentiment attribute while maintaining the content and semantic information unrelated to sentiment unchanged in the process. There are currently two main challenges in this field: no parallel corpus and text attribute entanglement. In response to the above problems, this paper proposed a novel solution: Mask-Prompt-Rerank. Use the method of masking the sentiment words and then using prompt regeneration to transfer the sentence sentiment. Experiments on two sentiment benchmark datasets and one formality transfer benchmark dataset show that this approach makes the performance of small pre-trained language models comparable to that of the most advanced large models, while consuming two orders of magnitude less computing and memory.

Keywords: language model, natural language processing, prompt, text sentiment transfer

Procedia PDF Downloads 77
32057 Numerical Simulation of Flexural Strength of Steel Fiber Reinforced High Volume Fly Ash Concrete by Finite Element Analysis

Authors: Mahzabin Afroz, Indubhushan Patnaikuni, Srikanth Venkatesan

Abstract:

It is well-known that fly ash can be used in high volume as a partial replacement of cement to get beneficial effects on concrete. High volume fly ash (HVFA) concrete is currently emerging as a popular option to strengthen by fiber. Although studies have supported the use of fibers with fly ash, a unified model along with the incorporation into finite element software package to estimate the maximum flexural loads need to be developed. In this study, nonlinear finite element analysis of steel fiber reinforced high strength HVFA concrete beam under static loadings was conducted to investigate their failure modes in terms of ultimate load. First of all, the experimental investigation of mechanical properties of high strength HVFA concrete was done and validates with developed numerical model with the appropriate modeling of element size and mesh by ANSYS 16.2. To model the fiber within the concrete, three-dimensional random fiber distribution was simulated by spherical coordinate system. Three types of high strength HVFA concrete beams were analyzed reinforced with 0.5, 1 and 1.5% volume fractions of steel fibers with specific mechanical and physical properties. The result reveals that the use of nonlinear finite element analysis technique and three-dimensional random fiber orientation exhibited fairly good agreement with the experimental results of flexural strength, load deflection and crack propagation mechanism. By utilizing this improved model, it is possible to determine the flexural behavior of different types and proportions of steel fiber reinforced HVFA concrete beam under static load. So, this paper has the originality to predict the flexural properties of steel fiber reinforced high strength HVFA concrete by numerical simulations.

Keywords: finite element analysis, high volume fly ash, steel fibers, spherical coordinate system

Procedia PDF Downloads 130
32056 Numerical Study of Wettability on the Triangular Micro-pillared Surfaces Using Lattice Boltzmann Method

Authors: Ganesh Meshram, Gloria Biswal

Abstract:

In this study, we present the numerical investigation of surface wettability on triangular micropillar surfaces by using a two-dimensional (2D) pseudo-potential multiphase lattice Boltzmann method with a D2Q9 model for various interaction parameters of the range varies from -1.40 to -2.50. Initially, simulation of the equilibrium state of a water droplet on a flat surface is considered for various interaction parameters to examine the accuracy of the present numerical model. We then imposed the microscale pillars on the bottom wall of the surface with different heights of the pillars to form the hydrophobic and superhydrophobic surfaces which enable the higher contact angle. The wettability of surfaces is simulated with water droplets of radius 100 lattice units in the domain of 800x800 lattice units. The present study shows that increasing the interaction parameter of the pillared hydrophobic surfaces dramatically reduces the contact area between water droplets and solid walls due to the momentum redirection phenomenon. Contact angles for different values of interaction strength have been validated qualitatively with the analytical results.

Keywords: contact angle, lattice boltzmann method, d2q9 model, pseudo-potential multiphase method, hydrophobic surfaces, wenzel state, cassie-baxter state, wettability

Procedia PDF Downloads 64
32055 Factors for Entry Timing Choices Using Principal Axis Factorial Analysis and Logistic Regression Model

Authors: C. M. Mat Isa, H. Mohd Saman, S. R. Mohd Nasir, A. Jaapar

Abstract:

International market expansion involves a strategic process of market entry decision through which a firm expands its operation from domestic to the international domain. Hence, entry timing choices require the needs to balance the early entry risks and the problems in losing opportunities as a result of late entry into a new market. Questionnaire surveys administered to 115 Malaysian construction firms operating in 51 countries worldwide have resulted in 39.1 percent response rate. Factor analysis was used to determine the most significant factors affecting entry timing choices of the firms to penetrate the international market. A logistic regression analysis used to examine the firms’ entry timing choices, indicates that the model has correctly classified 89.5 per cent of cases as late movers. The findings reveal that the most significant factor influencing the construction firms’ choices as late movers was the firm factor related to the firm’s international experience, resources, competencies and financing capacity. The study also offers valuable information to construction firms with intention to internationalize their businesses.

Keywords: factors, early movers, entry timing choices, late movers, logistic regression model, principal axis factorial analysis, Malaysian construction firms

Procedia PDF Downloads 373
32054 Application of Human Biomonitoring and Physiologically-Based Pharmacokinetic Modelling to Quantify Exposure to Selected Toxic Elements in Soil

Authors: Eric Dede, Marcus Tindall, John W. Cherrie, Steve Hankin, Christopher Collins

Abstract:

Current exposure models used in contaminated land risk assessment are highly conservative. Use of these models may lead to over-estimation of actual exposures, possibly resulting in negative financial implications due to un-necessary remediation. Thus, we are carrying out a study seeking to improve our understanding of human exposure to selected toxic elements in soil: arsenic (As), cadmium (Cd), chromium (Cr), nickel (Ni), and lead (Pb) resulting from allotment land-use. The study employs biomonitoring and physiologically-based pharmacokinetic (PBPK) modelling to quantify human exposure to these elements. We recruited 37 allotment users (adults > 18 years old) in Scotland, UK, to participate in the study. Concentrations of the elements (and their bioaccessibility) were measured in allotment samples (soil and allotment produce). Amount of produce consumed by the participants and participants’ biological samples (urine and blood) were collected for up to 12 consecutive months. Ethical approval was granted by the University of Reading Research Ethics Committee. PBPK models (coded in MATLAB) were used to estimate the distribution and accumulation of the elements in key body compartments, thus indicating the internal body burden. Simulating low element intake (based on estimated ‘doses’ from produce consumption records), predictive models suggested that detection of these elements in urine and blood was possible within a given period of time following exposure. This information was used in planning biomonitoring, and is currently being used in the interpretation of test results from biological samples. Evaluation of the models is being carried out using biomonitoring data, by comparing model predicted concentrations and measured biomarker concentrations. The PBPK models will be used to generate bioavailability values, which could be incorporated in contaminated land exposure models. Thus, the findings from this study will promote a more sustainable approach to contaminated land management.

Keywords: biomonitoring, exposure, PBPK modelling, toxic elements

Procedia PDF Downloads 318
32053 Prediction of Unsteady Heat Transfer over Square Cylinder in the Presence of Nanofluid by Using ANN

Authors: Ajoy Kumar Das, Prasenjit Dey

Abstract:

Heat transfer due to forced convection of copper water based nanofluid has been predicted by Artificial Neural network (ANN). The present nanofluid is formed by mixing copper nano particles in water and the volume fractions are considered here are 0% to 15% and the Reynolds number are kept constant at 100. The back propagation algorithm is used to train the network. The present ANN is trained by the input and output data which has been obtained from the numerical simulation, performed in finite volume based Computational Fluid Dynamics (CFD) commercial software Ansys Fluent. The numerical simulation based results are compared with the back propagation based ANN results. It is found that the forced convection heat transfer of water based nanofluid can be predicted correctly by ANN. It is also observed that the back propagation ANN can predict the heat transfer characteristics of nanofluid very quickly compared to standard CFD method.

Keywords: forced convection, square cylinder, nanofluid, neural network

Procedia PDF Downloads 316
32052 The Role of Demographics and Service Quality in the Adoption and Diffusion of E-Government Services: A Study in India

Authors: Sayantan Khanra, Rojers P. Joseph

Abstract:

Background and Significance: This study is aimed at analyzing the role of demographic and service quality variables in the adoption and diffusion of e-government services among the users in India. The study proposes to examine the users' perception about e-Government services and investigate the key variables that are most salient to the Indian populace. Description of the Basic Methodologies: The methodology to be adopted in this study is Hierarchical Regression Analysis, which will help in exploring the impact of the demographic variables and the quality dimensions on the willingness to use e-government services in two steps. First, the impact of demographic variables on the willingness to use e-government services is to be examined. In the second step, quality dimensions would be used as inputs to the model for explaining variance in excess of prior contribution by the demographic variables. Present Status: Our study is in the data collection stage in collaboration with a highly reliable, authentic and adequate source of user data. Assuming that the population of the study comprises all the Internet users in India, a massive sample size of more than 10,000 random respondents is being approached. Data is being collected using an online survey questionnaire. A pilot survey has already been carried out to refine the questionnaire with inputs from an expert in management information systems and a small group of users of e-government services in India. The first three questions in the survey pertain to the Internet usage pattern of a respondent and probe whether the person has used e-government services. If the respondent confirms that he/she has used e-government services, then an aggregate of 15 indicators are used to measure the quality dimensions under consideration and the willingness of the respondent to use e-government services, on a five-point Likert scale. If the respondent reports that he/she has not used e-government services, then a few optional questions are asked to understand the reason(s) behind the same. Last four questions in the survey are dedicated to collect data related to the demographic variables. An indication of the Major Findings: Based on the extensive literature review carried out to develop several propositions; a research model is prescribed to start with. A major outcome expected at the completion of the study is the development of a research model that would help to understand the relationship involving the demographic variables and service quality dimensions, and the willingness to adopt e-government services, particularly in an emerging economy like India. Concluding Statement: Governments of emerging economies and other relevant agencies can use the findings from the study in designing, updating, and promoting e-government services to enhance public participation, which in turn, would help to improve efficiency, convenience, engagement, and transparency in implementing these services.

Keywords: adoption and diffusion of e-government services, demographic variables, hierarchical regression analysis, service quality dimensions

Procedia PDF Downloads 263
32051 Evaluating the Effect of Climate Change and Land Use/Cover Change on Catchment Hydrology of Gumara Watershed, Upper Blue Nile Basin, Ethiopia

Authors: Gashaw Gismu Chakilu

Abstract:

Climate and land cover change are very important issues in terms of global context and their responses to environmental and socio-economic drivers. The dynamic of these two factors is currently affecting the environment in unbalanced way including watershed hydrology. In this paper individual and combined impacts of climate change and land use land cover change on hydrological processes were evaluated through applying the model Soil and Water Assessment Tool (SWAT) in Gumara watershed, Upper Blue Nile basin Ethiopia. The regional climate; temperature and rainfall data of the past 40 years in the study area were prepared and changes were detected by using trend analysis applying Mann-Kendall trend test. The land use land cover data were obtained from land sat image and processed by ERDAS IMAGIN 2010 software. Three land use land cover data; 1973, 1986, and 2013 were prepared and these data were used for base line, model calibration and change study respectively. The effects of these changes on high flow and low flow of the catchment have also been evaluated separately. The high flow of the catchment for these two decades was analyzed by using Annual Maximum (AM) model and the low flow was evaluated by seven day sustained low flow model. Both temperature and rainfall showed increasing trend; and then the extent of changes were evaluated in terms of monthly bases by using two decadal time periods; 1973-1982 was taken as baseline and 2004-2013 was used as change study. The efficiency of the model was determined by Nash-Sutcliffe (NS) and Relative Volume error (RVe) and their values were 0.65 and 0.032 for calibration and 0.62 and 0.0051 for validation respectively. The impact of climate change was higher than that of land use land cover change on stream flow of the catchment; the flow has been increasing by 16.86% and 7.25% due to climate and LULC change respectively, and the combined change effect accounted 22.13% flow increment. The overall results of the study indicated that Climate change is more responsible for high flow than low flow; and reversely the land use land cover change showed more significant effect on low flow than high flow of the catchment. From the result we conclude that the hydrology of the catchment has been altered because of changes of climate and land cover of the study area.

Keywords: climate, LULC, SWAT, Ethiopia

Procedia PDF Downloads 373
32050 Reflections on Lyotard's Reading of the Kantian Sublime and Its Political Import

Authors: Tugba Ayas Onol

Abstract:

The paper revisits Jean-François Lyotard’s interpretation of the Kantian Sublime as a tool for understanding politics after modernity. In 1985 Lyotard announces the end of rational politics based on consensus and claims that new strategies are urged to recognize the political imperatives of marginalized groups. The charm of the sublime as a reflective judgment is grounded on the fact that the judgment of sublime is free from any notion of consensus or common sense in particular. Lyotard interprets this feature of the sublime as a respect for heterogeneity and for him aesthetic judgments can be a model for understanding justice in postmodern times, in which it seems hard to follow a single universal law among different phrase regimes. More importantly, the Kantian sublime speaks to what Lyotard addresses as the incommensurability of phase genres. The present paper shall try to evaluate Lyotard’s employment of the Kantian notion of the sublime in relation to its possible political import.

Keywords: Kant, Lyotard, sublime, politics

Procedia PDF Downloads 376
32049 Wildfire-Related Debris-Flow and Flooding Using 2-D Hydrologic Model

Authors: Cheong Hyeon Oh, Dongho Nam, Byungsik Kim

Abstract:

Due to the recent climate change, flood damage caused by local floods and typhoons has frequently occurred, the incidence rate and intensity of wildfires are greatly increased due to increased temperatures and changes in precipitation patterns. Wildfires cause primary damage, such as loss of forest resources, as well as secondary disasters, such as landslides, floods, and debris flow. In many countries around the world, damage and economic losses from secondary damage are occurring as well as the direct effects of forest fires. Therefore, in this study, the Rainfall-Runoff model(S-RAT) was used for the wildfire affected areas in Gangneung and Goseong, which occurred on April 2019, when the stability of vegetation and soil were destroyed by wildfires. Rainfall data from Typhoon Rusa were used in the S-RAT model, and flood discharge was calculated according to changes in land cover before and after wildfire damage. The results of the calculation showed that flood discharge increased significantly due to changes in land cover, as the increase in flood discharge increases the possibility of the occurrence of the debris flow and the extent of the damage, the debris flow height and range were calculated before and after forest fire using RAMMS. The analysis results showed that the height and extent of damage increased after wildfire, but the result value was underestimated due to the characteristics that using DEM and maximum flood discharge of the RAMMS model. This research was supported by a grant(2017-MOIS31-004) from Fundamental Technology Development Program for Extreme Disaster Response funded by Korean Ministry of Interior and Safety (MOIS). This paper work (or document) was financially supported by Ministry of the Interior and Safety as 'Human resoure development Project in Disaster management'.

Keywords: wildfire, debris flow, land cover, rainfall-runoff meodel S-RAT, RAMMS, height

Procedia PDF Downloads 117
32048 Comparative Study od Three Artificial Intelligence Techniques for Rain Domain in Precipitation Forecast

Authors: Nabilah Filzah Mohd Radzuan, Andi Putra, Zalinda Othman, Azuraliza Abu Bakar, Abdul Razak Hamdan

Abstract:

Precipitation forecast is important to avoid natural disaster incident which can cause losses in the involved area. This paper reviews three techniques logistic regression, decision tree, and random forest which are used in making precipitation forecast. These combination techniques through the vector auto-regression (VAR) model help in finding the advantages and strengths of each technique in the forecast process. The data-set contains variables of the rain’s domain. Adaptation of artificial intelligence techniques involved in rain domain enables the forecast process to be easier and systematic for precipitation forecast.

Keywords: logistic regression, decisions tree, random forest, VAR model

Procedia PDF Downloads 443
32047 A Review of the Run to Run (R to R) Control in the Manufacturing Processes

Authors: Khalil Aghapouramin, Mostafa Ranjbar

Abstract:

Run- to- Run (R2 R) control was developed in order to monitor and control different semiconductor manufacturing processes based upon the fundamental engineering frameworks. This technology allows rectification in the optimum direction. This control always had a significant potency in which was appeared in a variety of processes. The term run to run refers to the case where the act of control would take with the aim of getting batches of silicon wafers which produced in a manufacturing process. In the present work, a brief review about run-to-run control investigated which mainly is effective in the manufacturing process.

Keywords: Run-to-Run (R2R) control, manufacturing, process in engineering, manufacturing controls

Procedia PDF Downloads 484
32046 The Impact of Land Cover Change on Stream Discharges and Water Resources in Luvuvhu River Catchment, Vhembe District, Limpopo Province, South Africa

Authors: P. M. Kundu, L. R. Singo, J. O. Odiyo

Abstract:

Luvuvhu River catchment in South Africa experiences floods resulting from heavy rainfall of intensities exceeding 15 mm per hour associated with the Inter-tropical Convergence Zone (ITCZ). The generation of runoff is triggered by the rainfall intensity and soil moisture status. In this study, remote sensing and GIS techniques were used to analyze the hydrologic response to land cover changes. Runoff was calculated as a product of the net precipitation and a curve number coefficient. It was then routed using the Muskingum-Cunge method using a diffusive wave transfer model that enabled the calculation of response functions between start and end point. Flood frequency analysis was determined using theoretical probability distributions. Spatial data on land cover was obtained from multi-temporal Landsat images while data on rainfall, soil type, runoff and stream discharges was obtained by direct measurements in the field and from the Department of Water. A digital elevation model was generated from contour maps available at http://www.ngi.gov.za. The results showed that land cover changes had impacted negatively to the hydrology of the catchment. Peak discharges in the whole catchment were noted to have increased by at least 17% over the period while flood volumes were noted to have increased by at least 11% over the same period. The flood time to peak indicated a decreasing trend, in the range of 0.5 to 1 hour within the years. The synergism between remotely sensed digital data and GIS for land surface analysis and modeling was realized, and it was therefore concluded that hydrologic modeling has potential for determining the influence of changes in land cover on the hydrologic response of the catchment.

Keywords: catchment, digital elevation model, hydrological model, routing, runoff

Procedia PDF Downloads 564
32045 Cache Analysis and Software Optimizations for Faster on-Chip Network Simulations

Authors: Khyamling Parane, B. M. Prabhu Prasad, Basavaraj Talawar

Abstract:

Fast simulations are critical in reducing time to market in CMPs and SoCs. Several simulators have been used to evaluate the performance and power consumed by Network-on-Chips. Researchers and designers rely upon these simulators for design space exploration of NoC architectures. Our experiments show that simulating large NoC topologies take hours to several days for completion. To speed up the simulations, it is necessary to investigate and optimize the hotspots in simulator source code. Among several simulators available, we choose Booksim2.0, as it is being extensively used in the NoC community. In this paper, we analyze the cache and memory system behaviour of Booksim2.0 to accurately monitor input dependent performance bottlenecks. Our measurements show that cache and memory usage patterns vary widely based on the input parameters given to Booksim2.0. Based on these measurements, the cache configuration having least misses has been identified. To further reduce the cache misses, we use software optimization techniques such as removal of unused functions, loop interchanging and replacing post-increment operator with pre-increment operator for non-primitive data types. The cache misses were reduced by 18.52%, 5.34% and 3.91% by employing above technology respectively. We also employ thread parallelization and vectorization to improve the overall performance of Booksim2.0. The OpenMP programming model and SIMD are used for parallelizing and vectorizing the more time-consuming portions of Booksim2.0. Speedups of 2.93x and 3.97x were observed for the Mesh topology with 30 × 30 network size by employing thread parallelization and vectorization respectively.

Keywords: cache behaviour, network-on-chip, performance profiling, vectorization

Procedia PDF Downloads 192
32044 Numerical Verification of a Backfill-Rectangular Tank-Fluid System

Authors: Ramazan Livaoğlu, Tufan Çakır

Abstract:

The performance of rectangular tanks during earthquakes has been observed to depend significantly on the existence of water in the container and the presence of the backfill acting on tank wall. Therefore, in design of rectangular tanks, the topics of fluid-structure-backfill interactions and determination of modal characteristics of the interaction system have traditionally been one of the great theoretical and practical controversy. Although finite element method has been and will continue to be used to a significant extent in treating the response of the system, experimental verification of numerical models remains prerequisite for their adoption and reliable application in practice. Thus, in this study, the numerical and experimental investigations were performed on the backfill-exterior wall-fluid interaction system. Firstly, three dimensional finite element model (3D-FEM) was developed to acquire modal frequencies and mode shapes of the system by means of ANSYS. Secondly, a series of in-situ tests were fulfilled to define modal characteristics of same system to determine the applicability of the FEM to a real physical situation under field conditions. Finally, comparing the theoretical predictions from the model to results from experimental measurement, a close agreement was found between theory and experiment. Thus, it can be easily stated that experimental verification provides strong support for the use of proposed model in further investigations.

Keywords: fluid-structure interaction, modal analysis, rectangular tank, soil structure interaction

Procedia PDF Downloads 390
32043 Modelling Optimal Control of Diabetes in the Workplace

Authors: Eunice Christabel Chukwu

Abstract:

Introduction: Diabetes is a chronic medical condition which is characterized by high levels of glucose in the blood and urine; it is usually diagnosed by means of a glucose tolerance test (GTT). Diabetes can cause a range of health problems if left unmanaged, as it can lead to serious complications. It is essential to manage the condition effectively, particularly in the workplace where the impact on work productivity can be significant. This paper discusses the modelling of optimal control of diabetes in the workplace using a control theory approach. Background: Diabetes mellitus is a condition caused by too much glucose in the blood. Insulin, a hormone produced by the pancreas, controls the blood sugar level by regulating the production and storage of glucose. In diabetes, there may be a decrease in the body’s ability to respond to insulin or a decrease in insulin produced by the pancreas which will lead to abnormalities in the metabolism of carbohydrates, proteins, and fats. In addition to the health implications, the condition can also have a significant impact on work productivity, as employees with uncontrolled diabetes are at risk of absenteeism, reduced performance, and increased healthcare costs. While several interventions are available to manage diabetes, the most effective approach is to control blood glucose levels through a combination of lifestyle modifications and medication. Methodology: The control theory approach involves modelling the dynamics of the system and designing a controller that can regulate the system to achieve optimal performance. In the case of diabetes, the system dynamics can be modelled using a mathematical model that describes the relationship between insulin, glucose, and other variables. The controller can then be designed to regulate the glucose levels to maintain them within a healthy range. Results: The modelling of optimal control of diabetes in the workplace using a control theory approach has shown promising results. The model has been able to predict the optimal dose of insulin required to maintain glucose levels within a healthy range, taking into account the individual’s lifestyle, medication regimen, and other relevant factors. The approach has also been used to design interventions that can improve diabetes management in the workplace, such as regular glucose monitoring and education programs. Conclusion: The modelling of optimal control of diabetes in the workplace using a control theory approach has significant potential to improve diabetes management and work productivity. By using a mathematical model and a controller to regulate glucose levels, the approach can help individuals with diabetes to achieve optimal health outcomes while minimizing the impact of the condition on their work performance. Further research is needed to validate the model and develop interventions that can be implemented in the workplace.

Keywords: mathematical model, blood, insulin, pancreas, model, glucose

Procedia PDF Downloads 57
32042 INRAM-3DCNN: Multi-Scale Convolutional Neural Network Based on Residual and Attention Module Combined with Multilayer Perceptron for Hyperspectral Image Classification

Authors: Jianhong Xiang, Rui Sun, Linyu Wang

Abstract:

In recent years, due to the continuous improvement of deep learning theory, Convolutional Neural Network (CNN) has played a great superior performance in the research of Hyperspectral Image (HSI) classification. Since HSI has rich spatial-spectral information, only utilizing a single dimensional or single size convolutional kernel will limit the detailed feature information received by CNN, which limits the classification accuracy of HSI. In this paper, we design a multi-scale CNN with MLP based on residual and attention modules (INRAM-3DCNN) for the HSI classification task. We propose to use multiple 3D convolutional kernels to extract the packet feature information and fully learn the spatial-spectral features of HSI while designing residual 3D convolutional branches to avoid the decline of classification accuracy due to network degradation. Secondly, we also design the 2D Inception module with a joint channel attention mechanism to quickly extract key spatial feature information at different scales of HSI and reduce the complexity of the 3D model. Due to the high parallel processing capability and nonlinear global action of the Multilayer Perceptron (MLP), we use it in combination with the previous CNN structure for the final classification process. The experimental results on two HSI datasets show that the proposed INRAM-3DCNN method has superior classification performance and can perform the classification task excellently.

Keywords: INRAM-3DCNN, residual, channel attention, hyperspectral image classification

Procedia PDF Downloads 74
32041 Learning from Dendrites: Improving the Point Neuron Model

Authors: Alexander Vandesompele, Joni Dambre

Abstract:

The diversity in dendritic arborization, as first illustrated by Santiago Ramon y Cajal, has always suggested a role for dendrites in the functionality of neurons. In the past decades, thanks to new recording techniques and optical stimulation methods, it has become clear that dendrites are not merely passive electrical components. They are observed to integrate inputs in a non-linear fashion and actively participate in computations. Regardless, in simulations of neural networks dendritic structure and functionality are often overlooked. Especially in a machine learning context, when designing artificial neural networks, point neuron models such as the leaky-integrate-and-fire (LIF) model are dominant. These models mimic the integration of inputs at the neuron soma, and ignore the existence of dendrites. In this work, the LIF point neuron model is extended with a simple form of dendritic computation. This gives the LIF neuron increased capacity to discriminate spatiotemporal input sequences, a dendritic functionality as observed in another study. Simulations of the spiking neurons are performed using the Bindsnet framework. In the common LIF model, incoming synapses are independent. Here, we introduce a dependency between incoming synapses such that the post-synaptic impact of a spike is not only determined by the weight of the synapse, but also by the activity of other synapses. This is a form of short term plasticity where synapses are potentiated or depressed by the preceding activity of neighbouring synapses. This is a straightforward way to prevent inputs from simply summing linearly at the soma. To implement this, each pair of synapses on a neuron is assigned a variable,representing the synaptic relation. This variable determines the magnitude ofthe short term plasticity. These variables can be chosen randomly or, more interestingly, can be learned using a form of Hebbian learning. We use Spike-Time-Dependent-Plasticity (STDP), commonly used to learn synaptic strength magnitudes. If all neurons in a layer receive the same input, they tend to learn the same through STDP. Adding inhibitory connections between the neurons creates a winner-take-all (WTA) network. This causes the different neurons to learn different input sequences. To illustrate the impact of the proposed dendritic mechanism, even without learning, we attach five input neurons to two output neurons. One output neuron isa regular LIF neuron, the other output neuron is a LIF neuron with dendritic relationships. Then, the five input neurons are allowed to fire in a particular order. The membrane potentials are reset and subsequently the five input neurons are fired in the reversed order. As the regular LIF neuron linearly integrates its inputs at the soma, the membrane potential response to both sequences is similar in magnitude. In the other output neuron, due to the dendritic mechanism, the membrane potential response is different for both sequences. Hence, the dendritic mechanism improves the neuron’s capacity for discriminating spa-tiotemporal sequences. Dendritic computations improve LIF neurons even if the relationships between synapses are established randomly. Ideally however, a learning rule is used to improve the dendritic relationships based on input data. It is possible to learn synaptic strength with STDP, to make a neuron more sensitive to its input. Similarly, it is possible to learn dendritic relationships with STDP, to make the neuron more sensitive to spatiotemporal input sequences. Feeding structured data to a WTA network with dendritic computation leads to a significantly higher number of discriminated input patterns. Without the dendritic computation, output neurons are less specific and may, for instance, be activated by a sequence in reverse order.

Keywords: dendritic computation, spiking neural networks, point neuron model

Procedia PDF Downloads 128
32040 Simulation and Analysis of Passive Parameters of Building in eQuest: A Case Study in Istanbul, Turkey

Authors: Mahdiyeh Zafaranchi

Abstract:

With rapid development of urbanization and improvement of living standards in the world, energy consumption and carbon emissions of the building sector are expected to increase in the near future; because of that, energy-saving issues have become more important among the engineers. Besides, the building sector is a major contributor to energy consumption and carbon emissions. The concept of efficient building appeared as a response to the need for reducing energy demand in this sector which has the main purpose of shifting from standard buildings to low-energy buildings. Although energy-saving should happen in all steps of a building during the life cycle (material production, construction, demolition), the main concept of efficient energy building is saving energy during the life expectancy of a building by using passive and active systems, and should not sacrifice comfort and quality to reach these goals. The main aim of this study is to investigate passive strategies (do not need energy consumption or use renewable energy) to achieve energy-efficient buildings. Energy retrofit measures were explored by eQuest software using a case study as a base model. The study investigates predictive accuracy for the major factors like thermal transmittance (U-value) of the material, windows, shading devices, thermal insulation, rate of the exposed envelope, window/wall ration, lighting system in the energy consumption of the building. The base model was located in Istanbul, Turkey. The impact of eight passive parameters on energy consumption had been indicated. After analyzing the base model by eQuest, a final scenario was suggested which had a good energy performance. The results showed a decrease in the U-values of materials, the rate of exposing buildings, and windows had a significant effect on energy consumption. Finally, savings in electric consumption of about 10.5%, and gas consumption by about 8.37% in the suggested model were achieved annually.

Keywords: efficient building, electric and gas consumption, eQuest, Passive parameters

Procedia PDF Downloads 108
32039 Dynamics of Hepatitis B Infection Prevention Practices among Pregnant Women Attending Antenatal Care in Central Uganda Using the Constructs of Information-Motivation-Behavioral Skills Model: A Case of Lubaga Hospital Kampala

Authors: Ismail Bamidele Afolabi, Abdulmujeeb Babatunde Aremu, Lawal Abdurraheem Maidoki, Nnodimele Onuigbo Atulomah

Abstract:

Background: Hepatitis B virus infection remains a significant global public health challenge with infectivity as well as the potential for transmission more than 50 to 100 times that of HIV. Annually, global HBV-related mortality is linked primarily to cirrhosis and liver carcinoma. The ever-increasing endemicity of HBV among children under-5-years, owing to vertical transmission and its lingering chronicity in developing countries, will hamper the global efforts concertedly endorsed towards eliminating viral hepatitis as a global public health threat by 2030. Objective: This study assessed information motivation behavioral skills model constructs as predictors of HBV infection prevention practices among consenting expectant mothers attending antenatal care in Central Uganda as a focal point of intervention towards breaking materno-foetal transmission of HBV. Methods: A cross-sectional study with a quantitative data collection approach based on the constructs of the IMB model was used to capture data on the study variables among 385 randomly selected pregnant women between September and October 2020. Data derived from the quantitative instrument were transformed into weighted aggregate scores using SPSS version 26. ANOVA and regression analysis were done to ascertain the study hypotheses with a significance level set as (p ≤ 0.05). Results: Relatively 60% of the respondents were aged between 18 and 28. Expectant mothers with secondary education (42.3%) were predominant. Furthermore, an average but inadequate knowledge (X ̅=5.97±6.61; B=0.57; p<.001), incorrect perception (X ̅=17.10±18.31; B=0.97; p=.014), and good behavioral skills (X ̅=12.39±13.37; B=0.56; p<.001) for adopting prevention practices all statistically predicted the unsatisfactory level of prevention practices (X ̅=15.03±16.20) among the study respondents as measured on rating scales of 12, 33, 21 and 30 respectively. Conclusion: Evidence from this study corroborates the imperativeness of IMB constructs in reducing the burden of HBV infection in developing countries. Therefore, the inadequate HBV knowledge and misperception among obstetric populations necessitate personalized health education during antenatal visits and subsequent health campaigns in order to inform better prevention practices and, in turn, reduce the lingering chronicity of HBV infection in developing countries.

Keywords: behavioral skills, HBV infection, knowledge, perception, pregnant women, prevention practices

Procedia PDF Downloads 86
32038 Attachment as a Predictor for Cognitive Rigidity

Authors: Barbara Gawda

Abstract:

Attachment model formed in childhood has an important impact on emotional development, personality, and social relationships. Attachment is also thought to have an impact on construction of affective-cognitive schemas and cognitive functioning. The aim of the current study was to verify whether there is an association between attachment and cognitive rigidity defined as dogmatism and intolerance of ambiguity. The analysis of 180 participants (persons of a similar age and education level, number of men and women was equal) was conducted. To test the attachment styles, the Revised Experiences in Close Relationships Inventory (ECR-R) was used. To examine cognitive rigidity, the Rokeach and Budner questionnaires were used. A multiple regression model was employed to examine whether attachment styles are predictors for dogmatism. The results confirmed that fearful-ambivalent attachment is the main predictor for dogmatism but not for intolerance of ambiguity.

Keywords: attachment styles, cognitive rigidity, dogmatism, intolerance of ambiguity

Procedia PDF Downloads 333
32037 Antihypertensive Activity of Alcoholic Extract of Citrus Paradise Juice in One Clip One Kidney Hypertension Model in Rats

Authors: Lokesh Bhatt, Jayesh Rathod

Abstract:

Hypertension is one of the most prevalent cardiovascular disorder. It is responsible for several other cardiovascular disorders. Although many drugs are available for the treatment of hypertension, still a large population has uncontrolled blood pressure. Thus there is an unmet need for new therapeutic approaches for the same. Fruit juice of Citrus paradise contains several flavonoids with vasodilatory activity. We hypothesized that alcoholic extract of Citrus paradise, which contains flavonoids, might attenuate hypertension. The objective of the present study was to evaluate the antihypertensive activity of alcoholic extract of Citrus paradise fruit juice in rats. Hypertension was induced using one clip one kidney model in rats. The renal artery was occluded for 4 h after removal of one kidney. Once stabilized, the ganglionic blockade was performed followed by removal of the arterial clip from the kidney. Removal of clip resulted in an increase in blood pressure which is due to release of renin from the kidney. Alcoholic extract of Citrus paradise fruit juice was then administered at 50 mg/kg and 100 mg/kg dose by intravenous injection. Blood pressure was monitored continuously. Alcoholic extract of Citrus paradise fruit juice reduced hypertension in dose-dependent manner. Antihypertensive activity was found to be associated with vasodilation. The results of the present study showed antihypertensive potential of alcoholic extract of Citrus paradise fruit juice.

Keywords: citrus paradise, alcoholic extract, one clip one kidney model, vasodilation

Procedia PDF Downloads 283
32036 Enhancement of Tribological Behavior for Diesel Engine Piston of Solid Skirt by an Optimal Choice of Interface Material

Authors: M. Amara, M. Tahar Abbes, A. Dokkiche, M. Benbrike

Abstract:

Shear stresses generate frictional forces thus lead to the reduction of engine performance due to the power losses. This friction can also cause damage to the piston material. Thus, the choice of an optimal material for the piston is necessary to improve the elastohydrodynamical contacts of the piston. In this study, to achieve this objective, an elastohydrodynamical lubrication model that satisfies the best tribological behavior of the piston with the optimum choice of material is developed. Several aluminum alloys composed of different components are studied in this simulation. An application is made on the piston 60 x 120 mm Diesel engine type F8L413 currently mounted on Deutz trucks TB230 by using different aluminum alloys where alloys based on aluminum-silicon have better tribological performance.

Keywords: EHD lubricated contacts, friction, properties of materials, tribological performance

Procedia PDF Downloads 264