Search results for: modeling methodology
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8740

Search results for: modeling methodology

7540 Transportation and Urban Land-Use System for the Sustainability of Cities, a Case Study of Muscat

Authors: Bader Eddin Al Asali, N. Srinivasa Reddy

Abstract:

Cities are dynamic in nature and are characterized by concentration of people, infrastructure, services and markets, which offer opportunities for production and consumption. Often growth and development in urban areas is not systematic, and is directed by number of factors like natural growth, land prices, housing availability, job locations-the central business district (CBD’s), transportation routes, distribution of resources, geographical boundaries, administrative policies, etc. One sided spatial and geographical development in cities leads to the unequal spatial distribution of population and jobs, resulting in high transportation activity. City development can be measured by the parameters such as urban size, urban form, urban shape, and urban structure. Urban Size is the city size and defined by the population of the city, and urban form is the location and size of the economic activity (CBD) over the geographical space. Urban shape is the geometrical shape of the city over which the distribution of population and economic activity occupied. And Urban Structure is the transport network within which the population and activity centers are connected by hierarchy of roads. Among the urban land-use systems transportation plays significant role and is one of the largest energy consuming sector. Transportation interaction among the land uses is measured in Passenger-Km and mean trip length, and is often used as a proxy for measurement of energy consumption in transportation sector. Among the trips generated in cities, work trips constitute more than 70 percent. Work trips are originated from the place of residence and destination to the place of employment. To understand the role of urban parameters on transportation interaction, theoretical cities of different size and urban specifications are generated through building block exercise using a specially developed interactive C++ programme and land use transportation modeling is carried. The land-use transportation modeling exercise helps in understanding the role of urban parameters and also to classify the cities for their urban form, structure, and shape. Muscat the capital city of Oman underwent rapid urbanization over the last four decades is taken as a case study for its classification. Also, a pilot survey is carried to capture urban travel characteristics. Analysis of land-use transportation modeling with field data classified Muscat as a linear city with polycentric CBD. Conclusions are drawn suggestion are given for policy making for the sustainability of Muscat City.

Keywords: land-use transportation, transportation modeling urban form, urban structure, urban rule parameters

Procedia PDF Downloads 257
7539 A Review of Different Studies on Hidden Markov Models for Multi-Temporal Satellite Images: Stationarity and Non-Stationarity Issues

Authors: Ali Ben Abbes, Imed Riadh Farah

Abstract:

Due to the considerable advances in Multi-Temporal Satellite Images (MTSI), remote sensing application became more accurate. Recently, many advances in modeling MTSI are developed using various models. The purpose of this article is to present an overview of studies using Hidden Markov Model (HMM). First of all, we provide a background of using HMM and their applications in this context. A comparison of the different works is discussed, and possible areas and challenges are highlighted. Secondly, we discussed the difference on vegetation monitoring as well as urban growth. Nevertheless, most research efforts have been used only stationary data. From another point of view, in this paper, we describe a new non-stationarity HMM, that is defined with a set of parts of the time series e.g. seasonal, trend and random. In addition, a new approach giving more accurate results and improve the applicability of the HMM in modeling a non-stationary data series. In order to assess the performance of the HMM, different experiments are carried out using Moderate Resolution Imaging Spectroradiometer (MODIS) NDVI time series of the northwestern region of Tunisia and Landsat time series of tres Cantos-Madrid in Spain.

Keywords: multi-temporal satellite image, HMM , nonstationarity, vegetation, urban

Procedia PDF Downloads 342
7538 Accelerated Evaluation of Structural Reliability under Tsunami Loading

Authors: Sai Hung Cheung, Zhe Shao

Abstract:

It is of our great interest to quantify the risk to structural dynamic systems due to earthquake-induced tsunamis in view of recent earthquake-induced tsunamis in Padang, 2004 and Tohoku, 2011 which brought huge losses of lives and properties. Despite continuous advancement in computational simulation of the tsunami and wave-structure interaction modeling, it still remains computationally challenging to evaluate the reliability of a structural dynamic system when uncertainties related to the system and its modeling are taken into account. The failure of the structure in a tsunami-wave-structural system is defined as any response quantities of the system exceeding specified thresholds during the time when the structure is subjected to dynamic wave impact due to earthquake-induced tsunamis. In this paper, an approach based on a novel integration of a recently proposed moving least squares response surface approach for stochastic sampling and the Subset Simulation algorithm is proposed. The effectiveness of the proposed approach is discussed by comparing its results with those obtained from the Subset Simulation algorithm without using the response surface approach.

Keywords: response surface, stochastic simulation, structural reliability tsunami, risk

Procedia PDF Downloads 659
7537 The Impact of the Use of Some Multiple Intelligence-Based Teaching Strategies on Developing Moral Intelligence and Inferential Jurisprudential Thinking among Secondary School Female Students in Saudi Arabia

Authors: Sameerah A. Al-Hariri Al-Zahrani

Abstract:

The current study aims at getting acquainted with the impact of the use of some multiple intelligence-based teaching strategies on developing moral intelligence and inferential jurisprudential thinking among secondary school female students. The study has endeavored to answer the following questions: What is the impact of the use of some multiple intelligence-based teaching strategies on developing inferential jurisprudential thinking and moral intelligence among first-year secondary school female students? In the frame of this main research question, the study seeks to answer the following sub-questions: (i) What are the inferential jurisprudential thinking skills among first-year secondary school female students? (ii) What are the components of moral intelligence among first year secondary school female students? (iii) What is the impact of the use of some multiple intelligence‐based teaching strategies (such as the strategies of analyzing values, modeling, Socratic discussion, collaborative learning, peer collaboration, collective stories, building emotional moments, role play, one-minute observation) on moral intelligence among first-year secondary school female students? (iv) What is the impact of the use of some multiple intelligence‐based teaching strategies (such as the strategies of analyzing values, modeling, Socratic discussion, collaborative learning, peer collaboration, collective stories, building emotional moments, role play, one-minute observation) on developing the capacity for inferential jurisprudential thinking of juristic rules among first-year secondary school female students? The study has used the descriptive-analytical methodology in surveying, analyzing, and reviewing the literature on previous studies in order to benefit from them in building the tools of the study and the materials of experimental treatment. The study has also used the experimental method to study the impact of the independent variable (multiple intelligence strategies) on the two dependent variables (moral intelligence and inferential jurisprudential thinking) in first-year secondary school female students’ learning. The sample of the study is made up of 70 female students that have been divided into two groups: an experimental group consisting of 35 students who have been taught through multiple intelligence strategies, and a control group consisting of the other 35 students who have been taught normally. The two tools of the study (inferential jurisprudential thinking test and moral intelligence scale) have been implemented on the two groups as a pre-test. The female researcher taught the experimental group and implemented the two tools of the study. After the experiment, which lasted eight weeks, was over, the study showed the following results: (i) The existence of significant statistical differences (0.05) between the mean average of the control group and that of the experimental group in the inferential jurisprudential thinking test (recognition of the evidence of jurisprudential rule, recognition of the motive for the jurisprudential rule, jurisprudential inferencing, analogical jurisprudence) in favor of the experimental group. (ii) The existence of significant statistical differences (0.05) between the mean average of the control group and that of the experimental group in the components of the moral intelligence scale (sympathy, conscience, moral wisdom, tolerance, justice, respect) in favor of the experimental group. The study has, thus, demonstrated the impact of the use of some multiple intelligence-based teaching strategies on developing moral intelligence and inferential jurisprudential thinking.

Keywords: moral intelligence, teaching, inferential jurisprudential thinking, secondary school

Procedia PDF Downloads 148
7536 Modeling and System Identification of a Variable Excited Linear Direct Drive

Authors: Heiko Weiß, Andreas Meister, Christoph Ament, Nils Dreifke

Abstract:

Linear actuators are deployed in a wide range of applications. This paper presents the modeling and system identification of a variable excited linear direct drive (LDD). The LDD is designed based on linear hybrid stepper technology exhibiting the characteristic tooth structure of mover and stator. A three-phase topology provides the thrust force caused by alternating strengthening and weakening of the flux of the legs. To achieve best possible synchronous operation, the phases are commutated sinusoidal. Despite the fact that these LDDs provide high dynamics and drive forces, noise emission limits their operation in calm workspaces. To overcome this drawback an additional excitation of the magnetic circuit is introduced to LDD using additional enabling coils instead of permanent magnets. The new degree of freedom can be used to reduce force variations and related noise by varying the excitation flux that is usually generated by permanent magnets. Hence, an identified simulation model is necessary to analyze the effects of this modification. Especially the force variations must be modeled well in order to reduce them sufficiently. The model can be divided into three parts: the current dynamics, the mechanics and the force functions. These subsystems are described with differential equations or nonlinear analytic functions, respectively. Ordinary nonlinear differential equations are derived and transformed into state space representation. Experiments have been carried out on a test rig to identify the system parameters of the complete model. Static and dynamic simulation based optimizations are utilized for identification. The results are verified in time and frequency domain. Finally, the identified model provides a basis for later design of control strategies to reduce existing force variations.

Keywords: force variations, linear direct drive, modeling and system identification, variable excitation flux

Procedia PDF Downloads 358
7535 Modeling of the Mechanism of Ion Channel Opening of the Visual Receptor's Rod on the Light and Allosteric Effect of Rhodopsin in the Phosphorylation Process

Authors: N. S. Vassilieva-Vashakmadze, R. A. Gakhokidze, I. M. Khachatryan

Abstract:

In the first part of the paper it is shown that both the depolarization of the cytoplasmic membrane of rods observed in invertebrates and hyperpolarization characteristic of vertebrates on the light may activate the functioning of ion (Na+) channels of cytoplasmic membrane of rods and thus provide the emergence of nerve impulse and its transfer to the neighboring neuron etc. In the second part, using the quantum mechanical program for modeling of the molecular processes, we got a clear picture demonstrating the effect of charged phosphate groups on the protein components of α-helical subunits of the visual rhodopsin receptor. The analysis shows that the phosphorylation of terminal amino acid of seventh α-helical subunits of the visual rhodopsin causes a redistribution of electron density on the atoms, i.e. polarization of subunits, also the changing the configuration of the nuclear subsystem, which corresponds to the deformation process in the molecule. Based on the use of models it can be concluded that this system has an internal relationship between polarization and deformation processes that indicates on the allosteric effect. The allosteric effect is based on quantum-mechanical principle of the self-consistency of the molecules.

Keywords: membrane potential, ion channels, visual rhodopsin, allosteric effect

Procedia PDF Downloads 260
7534 Modeling Approach for Evaluating Infiltration Rate of a Large-Scale Housing Stock

Authors: Azzam Alosaimi

Abstract:

Different countries attempt to reduce energy demands and Greenhouse Gas (GHG) emissions to mitigate global warming potential. They set different building codes to regulate excessive building’s energy losses. Energy losses occur due to pressure difference between the indoor and outdoor environments, and thus, heat transfers from one region to another. One major sources of energy loss is known as building airtightness. Building airtightness is the fundamental feature of the building envelope that directly impacts infiltration. Most of international building codes require minimum performance for new construction to ensure acceptable airtightness. The execution of airtightness required standards has become more challenging in recent years due to a lack of expertise and equipment, making it costly and time-consuming. Hence, researchers have developed predictive models to predict buildings infiltration rates to meet building codes and to reduce energy and cost. This research applies a theoretical modeling approach using Matlab software to predict mean infiltration rate distributions and total heat loss of Saudi Arabia’s housing stock.

Keywords: infiltration rate, energy demands, heating loss, cooling loss, carbon emissions

Procedia PDF Downloads 142
7533 Physics-Informed Convolutional Neural Networks for Reservoir Simulation

Authors: Jiangxia Han, Liang Xue, Keda Chen

Abstract:

Despite the significant progress over the last decades in reservoir simulation using numerical discretization, meshing is complex. Moreover, the high degree of freedom of the space-time flow field makes the solution process very time-consuming. Therefore, we present Physics-Informed Convolutional Neural Networks(PICNN) as a hybrid scientific theory and data method for reservoir modeling. Besides labeled data, the model is driven by the scientific theories of the underlying problem, such as governing equations, boundary conditions, and initial conditions. PICNN integrates governing equations and boundary conditions into the network architecture in the form of a customized convolution kernel. The loss function is composed of data matching, initial conditions, and other measurable prior knowledge. By customizing the convolution kernel and minimizing the loss function, the neural network parameters not only fit the data but also honor the governing equation. The PICNN provides a methodology to model and history-match flow and transport problems in porous media. Numerical results demonstrate that the proposed PICNN can provide an accurate physical solution from a limited dataset. We show how this method can be applied in the context of a forward simulation for continuous problems. Furthermore, several complex scenarios are tested, including the existence of data noise, different work schedules, and different good patterns.

Keywords: convolutional neural networks, deep learning, flow and transport in porous media, physics-informed neural networks, reservoir simulation

Procedia PDF Downloads 121
7532 Selection of Pichia kudriavzevii Strain for the Production of Single-Cell Protein from Cassava Processing Waste

Authors: Phakamas Rachamontree, Theerawut Phusantisampan, Natthakorn Woravutthikul, Peerapong Pornwongthong, Malinee Sriariyanun

Abstract:

A total of 115 yeast strains isolated from local cassava processing wastes were measured for crude protein content. Among these strains, the strain MSY-2 possessed the highest protein concentration (>3.5 mg protein/mL). By using molecular identification tools, it was identified to be a strain of Pichia kudriavzevii based on similarity of D1/D2 domain of 26S rDNA region. In this study, to optimize the protein production by MSY-2 strain, Response Surface Methodology (RSM) was applied. The tested parameters were the carbon content, nitrogen content, and incubation time. Here, the value of regression coefficient (R2) = 0.7194 could be explained by the model, which is high to support the significance of the model. Under the optimal condition, the protein content was produced up to 3.77 g per L of the culture and MSY-2 strain contain 66.8 g protein per 100 g of cell dry weight. These results revealed the plausibility of applying the novel strain of yeast in single-cell protein production.

Keywords: single cell protein, response surface methodology, yeast, cassava processing waste

Procedia PDF Downloads 386
7531 Probabilistic Modeling of Post-Liquefaction Ground Deformation

Authors: Javad Sadoghi Yazdi, Robb Eric S. Moss

Abstract:

This paper utilizes a probabilistic liquefaction triggering method for modeling post-liquefaction ground deformation. This cone penetration test CPT-based liquefaction triggering is employed to estimate the factor of safety against liquefaction (FSL) and compute the maximum cyclic shear strain (γmax). The study identifies a maximum PL value of 90% across various relative densities, which challenges the decrease from 90% to 70% as relative density decreases. It reveals that PL ranges from 5% to 50% for volumetric strain (εvol) less than 1%, while for εvol values between 1% and 3.2%, PL spans from 50% to 90%. The application of the CPT-based simplified liquefaction triggering procedures has been employed in previous researches to estimate liquefaction ground-failure indices, such as the Liquefaction Potential Index (LPI) and Liquefaction Severity Number (LSN). However, several studies have been conducted to highlight the variability in liquefaction probability calculations, suggesting a more accurate depiction of liquefaction likelihood. Consequently, the utilization of these simplified methods may not offer practical efficiency. This paper further investigates the efficacy of various established liquefaction vulnerability parameters, including LPI and LSN, in explaining the observed liquefaction-induced damage within residential zones of Christchurch, New Zealand using results from CPT database.

Keywords: cone penetration test (CPT), liquefaction, postliquefaction, ground failure

Procedia PDF Downloads 50
7530 Vibration Propagation in Structures Through Structural Intensity Analysis

Authors: Takhchi Jamal, Ouisse Morvan, Sadoulet-Reboul Emeline, Bouhaddi Noureddine, Gagliardini Laurent, Bornet Frederic, Lakrad Faouzi

Abstract:

Structural intensity is a technique that can be used to indicate both the magnitude and direction of power flow through a structure from the excitation source to the dissipation sink. However, current analysis is limited to the low frequency range. At medium and high frequencies, a rotational component appear in the field, masking the energy flow and make its understanding difficult or impossible. The objective of this work is to implement a methodology to filter out the rotational components of the structural intensity field in order to fully understand the energy flow in complex structures. The approach is based on the Helmholtz decomposition. It allows to decompose the structural intensity field into rotational, irrotational, and harmonic components. Only the irrotational component is needed to describe the net power flow from a source to a dissipative zone in the structure. The methodology has been applied on academic structures, and it allows a good analysis of the energy transfer paths.

Keywords: structural intensity, power flow, helmholt decomposition, irrotational intensity

Procedia PDF Downloads 163
7529 Adjusting Electricity Demand Data to Account for the Impact of Loadshedding in Forecasting Models

Authors: Migael van Zyl, Stefanie Visser, Awelani Phaswana

Abstract:

The electricity landscape in South Africa is characterized by frequent occurrences of loadshedding, a measure implemented by Eskom to manage electricity generation shortages by curtailing demand. Loadshedding, classified into stages ranging from 1 to 8 based on severity, involves the systematic rotation of power cuts across municipalities according to predefined schedules. However, this practice introduces distortions in recorded electricity demand, posing challenges to accurate forecasting essential for budgeting, network planning, and generation scheduling. Addressing this challenge requires the development of a methodology to quantify the impact of loadshedding and integrate it back into metered electricity demand data. Fortunately, comprehensive records of loadshedding impacts are maintained in a database, enabling the alignment of Loadshedding effects with hourly demand data. This adjustment ensures that forecasts accurately reflect true demand patterns, independent of loadshedding's influence, thereby enhancing the reliability of electricity supply management in South Africa. This paper presents a methodology for determining the hourly impact of load scheduling and subsequently adjusting historical demand data to account for it. Furthermore, two forecasting models are developed: one utilizing the original dataset and the other using the adjusted data. A comparative analysis is conducted to evaluate forecast accuracy improvements resulting from the adjustment process. By implementing this methodology, stakeholders can make more informed decisions regarding electricity infrastructure investments, resource allocation, and operational planning, contributing to the overall stability and efficiency of South Africa's electricity supply system.

Keywords: electricity demand forecasting, load shedding, demand side management, data science

Procedia PDF Downloads 44
7528 An Investigation on Material Removal Rate of EDM Process: A Response Surface Methodology Approach

Authors: Azhar Equbal, Anoop Kumar Sood, M. Asif Equbal, M. Israr Equbal

Abstract:

In the present work response surface methodology (RSM) based central composite design (CCD) is used for analyzing the electrical discharge machining (EDM) process. For experimentation, mild steel is selected as work piece and copper is used as electrode. Three machining parameters namely current (I), spark on time (Ton) and spark off time (Toff) are selected as the input variables. The output or response chosen is material removal rate (MRR) which is to be maximized. To reduce the number of runs face centered central composite design (FCCCD) was used. ANOVA was used to determine the significance of parameter and interactions. The suitability of model is tested using Anderson darling (AD) plot. The results conclude that different parameters considered i.e. current, pulse on and pulse off time; all have dominant effect on the MRR. At last, the optimized parameter setting for maximizing MRR is found through main effect plot analysis.

Keywords: EDM, electrode, MRR, RSM, ANOVA

Procedia PDF Downloads 289
7527 Spatial Variability of Environmental Parameters and Its Relationship with an Environmental Injustice on the Bike Paths of Santiago, Chile

Authors: Alicia Muñoz, Pedro Oyola, Cristian Henriquez

Abstract:

Pollution in Santiago de Chile has a spatial variability due to different factors, including meteorological parameters and emission sources. Socioenvironmental aspects are also significant for pollution in the canopy layer since it influences the type of edification, vegetal mass proportion and other environmental conditions. This study analyzes spatially urban pollution in Santiago, specifically, from the bike path perspective. Bike paths are located in high traffic zones, as consequence, users are constantly exposed to urban pollution. Measurements were made at the higher polluted hour, three days a week, including three transit regimes, on the most polluted month of the year. The environmental parameters are fine particulate matter (Model 8520, DustTrak Aerosol Monitor, TSI), temperature and relative humidity; it was also considerate urban parameters as sky view factor and vegetal mass. Identification of an environmental injustice will be achieved with a spatial modeling, including all urban factors and environmental mediations with an economic index of population.

Keywords: canopy layer, environmental injustice, spatial modeling, urban pollution

Procedia PDF Downloads 209
7526 Ontology-Driven Knowledge Discovery and Validation from Admission Databases: A Structural Causal Model Approach for Polytechnic Education in Nigeria

Authors: Bernard Igoche Igoche, Olumuyiwa Matthew, Peter Bednar, Alexander Gegov

Abstract:

This study presents an ontology-driven approach for knowledge discovery and validation from admission databases in Nigerian polytechnic institutions. The research aims to address the challenges of extracting meaningful insights from vast amounts of admission data and utilizing them for decision-making and process improvement. The proposed methodology combines the knowledge discovery in databases (KDD) process with a structural causal model (SCM) ontological framework. The admission database of Benue State Polytechnic Ugbokolo (Benpoly) is used as a case study. The KDD process is employed to mine and distill knowledge from the database, while the SCM ontology is designed to identify and validate the important features of the admission process. The SCM validation is performed using the conditional independence test (CIT) criteria, and an algorithm is developed to implement the validation process. The identified features are then used for machine learning (ML) modeling and prediction of admission status. The results demonstrate the adequacy of the SCM ontological framework in representing the admission process and the high predictive accuracies achieved by the ML models, with k-nearest neighbors (KNN) and support vector machine (SVM) achieving 92% accuracy. The study concludes that the proposed ontology-driven approach contributes to the advancement of educational data mining and provides a foundation for future research in this domain.

Keywords: admission databases, educational data mining, machine learning, ontology-driven knowledge discovery, polytechnic education, structural causal model

Procedia PDF Downloads 42
7525 Mapping of Geological Structures Using Aerial Photography

Authors: Ankit Sharma, Mudit Sachan, Anurag Prakash

Abstract:

Rapid growth in data acquisition technologies through drones, have led to advances and interests in collecting high-resolution images of geological fields. Being advantageous in capturing high volume of data in short flights, a number of challenges have to overcome for efficient analysis of this data, especially while data acquisition, image interpretation and processing. We introduce a method that allows effective mapping of geological fields using photogrammetric data of surfaces, drainage area, water bodies etc, which will be captured by airborne vehicles like UAVs, we are not taking satellite images because of problems in adequate resolution, time when it is captured may be 1 yr back, availability problem, difficult to capture exact image, then night vision etc. This method includes advanced automated image interpretation technology and human data interaction to model structures and. First Geological structures will be detected from the primary photographic dataset and the equivalent three dimensional structures would then be identified by digital elevation model. We can calculate dip and its direction by using the above information. The structural map will be generated by adopting a specified methodology starting from choosing the appropriate camera, camera’s mounting system, UAVs design ( based on the area and application), Challenge in air borne systems like Errors in image orientation, payload problem, mosaicing and geo referencing and registering of different images to applying DEM. The paper shows the potential of using our method for accurate and efficient modeling of geological structures, capture particularly from remote, of inaccessible and hazardous sites.

Keywords: digital elevation model, mapping, photogrammetric data analysis, geological structures

Procedia PDF Downloads 672
7524 Structural Evaluation of Airfield Pavement Using Finite Element Analysis Based Methodology

Authors: Richard Ji

Abstract:

Nondestructive deflection testing has been accepted widely as a cost-effective tool for evaluating the structural condition of airfield pavements. Backcalculation of pavement layer moduli can be used to characterize the pavement existing condition in order to compute the load bearing capacity of pavement. This paper presents an improved best-fit backcalculation methodology based on deflection predictions obtained using finite element method (FEM). The best-fit approach is based on minimizing the squared error between falling weight deflectometer (FWD) measured deflections and FEM predicted deflections. Then, concrete elastic modulus and modulus of subgrade reaction were back-calculated using Heavy Weight Deflectometer (HWD) deflections collected at the National Airport Pavement Testing Facility (NAPTF) test site. It is an alternative and more versatile method in considering concrete slab geometry and HWD testing locations compared to methods currently available.

Keywords: nondestructive testing, pavement moduli backcalculation, finite element method, concrete pavements

Procedia PDF Downloads 156
7523 Implementation of State-Space and Super-Element Techniques for the Modeling and Control of Smart Structures with Damping Characteristics

Authors: Nader Ghareeb, Rüdiger Schmidt

Abstract:

Minimizing the weight in flexible structures means reducing material and costs as well. However, these structures could become prone to vibrations. Attenuating these vibrations has become a pivotal engineering problem that shifted the focus of many research endeavors. One technique to do that is to design and implement an active control system. This system is mainly composed of a vibrating structure, a sensor to perceive the vibrations, an actuator to counteract the influence of disturbances, and finally a controller to generate the appropriate control signals. In this work, two different techniques are explored to create two different mathematical models of an active control system. The first model is a finite element model with a reduced number of nodes and it is called a super-element. The second model is in the form of state-space representation, i.e. a set of partial differential equations. The damping coefficients are calculated and incorporated into both models. The effectiveness of these models is demonstrated when the system is excited by its first natural frequency and an active control strategy is developed and implemented to attenuate the resulting vibrations. Results from both modeling techniques are presented and compared.

Keywords: damping coefficients, finite element analysis, super-element, state-space model

Procedia PDF Downloads 308
7522 Automating 2D CAD to 3D Model Generation Process: Wall pop-ups

Authors: Mohit Gupta, Chialing Wei, Thomas Czerniawski

Abstract:

In this paper, we have built a neural network that can detect walls on 2D sheets and subsequently create a 3D model in Revit using Dynamo. The training set includes 3500 labeled images, and the detection algorithm used is YOLO. Typically, engineers/designers make concentrated efforts to convert 2D cad drawings to 3D models. This costs a considerable amount of time and human effort. This paper makes a contribution in automating the task of 3D walls modeling. 1. Detecting Walls in 2D cad and generating 3D pop-ups in Revit. 2. Saving designer his/her modeling time in drafting elements like walls from 2D cad to 3D representation. An object detection algorithm YOLO is used for wall detection and localization. The neural network is trained over 3500 labeled images of size 256x256x3. Then, Dynamo is interfaced with the output of the neural network to pop-up 3D walls in Revit. The research uses modern technological tools like deep learning and artificial intelligence to automate the process of generating 3D walls without needing humans to manually model them. Thus, contributes to saving time, human effort, and money.

Keywords: neural networks, Yolo, 2D to 3D transformation, CAD object detection

Procedia PDF Downloads 131
7521 Bathymetric Change of Brahmaputra River and Its Influence on Flooding Scenario

Authors: Arup Kumar Sarma, Rohan Kar

Abstract:

The development of physical model of River like Brahmaputra, which finds its origin in the Chema Yundung glacier of Tibet and flows through India and Bangladesh, is always expensive and very much time consuming. With the advancement of computational technique, mathematical modeling has found wide application. MIKE 21C is one such commercial software, developed by Danish Hydraulic Institute (DHI), with the depth-averaged approach and a two-dimensional curvilinear finite-difference model, which is capable of modeling hydrodynamic and morphological processes with some limitations. The main purpose of this study are to generate bathymetry of the River Brahmaputra starting from “Sadia” at upstream to “Dhubri,” at downstream stretching a distance of approximately 695 km, for four different years: 1957, 1971, 1977, and 1981 over the grid generated in the MIKE 21C and to carry out the hydrodynamic simulation for these years to analyze the effect of bathymetry change on the surface water elevation. The study has established that bathymetric change can influence the flood level significantly in some of the river reaches and therefore the modification or updating of regular bathymetry is very much essential for the reliable flood routing in alluvial rivers.

Keywords: bathymetry, brahmaputra river, hydrodynamic model, surface water elevation

Procedia PDF Downloads 440
7520 Graphical User Interface for Presting Matlab Work for Reduction of Chromatic Disperion Using Digital Signal Processing for Optical Communication

Authors: Muhammad Faiz Liew Abdullah, Bhagwan Das, Nor Shahida, Abdul Fattah Chandio

Abstract:

This study presents the designed features of Graphical User Interface (GUI) for chromatic dispersion (CD) reduction using digital signal processing (DSP) techniques. GUI is specially designed for windows platform. The obtained simulation results from matlab are presented via this GUI. After importing results from matlab in GUI, It will present your work on any windows7 and onwards versions platforms without matlab software. First part of the GUI contains the research methodology block diagram and in the second part, output for each stage is shown in separate reserved area for the result display. Each stage of methodology has the captions to display the results. This GUI will be very helpful during presentations instead of making slides this GUI will present all your work easily in the absence of other software’s such as Matlab, Labview, MS PowerPoint. GUI is designed using C programming in MS Visio Studio.

Keywords: Matlab simulation results, C programming, MS VISIO studio, chromatic dispersion

Procedia PDF Downloads 447
7519 Disaggregating and Forecasting the Total Energy Consumption of a Building: A Case Study of a High Cooling Demand Facility

Authors: Juliana Barcelos Cordeiro, Khashayar Mahani, Farbod Farzan, Mohsen A. Jafari

Abstract:

Energy disaggregation has been focused by many energy companies since energy efficiency can be achieved when the breakdown of energy consumption is known. Companies have been investing in technologies to come up with software and/or hardware solutions that can provide this type of information to the consumer. On the other hand, not all people can afford to have these technologies. Therefore, in this paper, we present a methodology for breaking down the aggregate consumption and identifying the highdemanding end-uses profiles. These energy profiles will be used to build the forecast model for optimal control purpose. A facility with high cooling load is used as an illustrative case study to demonstrate the results of proposed methodology. We apply a high level energy disaggregation through a pattern recognition approach in order to extract the consumption profile of its rooftop packaged units (RTUs) and present a forecast model for the energy consumption.  

Keywords: energy consumption forecasting, energy efficiency, load disaggregation, pattern recognition approach

Procedia PDF Downloads 261
7518 In-Flight Aircraft Performance Model Enhancement Using Adaptive Lookup Tables

Authors: Georges Ghazi, Magali Gelhaye, Ruxandra Botez

Abstract:

Over the years, the Flight Management System (FMS) has experienced a continuous improvement of its many features, to the point of becoming the pilot’s primary interface for flight planning operation on the airplane. With the assistance of the FMS, the concept of distance and time has been completely revolutionized, providing the crew members with the determination of the optimized route (or flight plan) from the departure airport to the arrival airport. To accomplish this function, the FMS needs an accurate Aircraft Performance Model (APM) of the aircraft. In general, APMs that equipped most modern FMSs are established before the entry into service of an individual aircraft, and results from the combination of a set of ordinary differential equations and a set of performance databases. Unfortunately, an aircraft in service is constantly exposed to dynamic loads that degrade its flight characteristics. These degradations endow two main origins: airframe deterioration (control surfaces rigging, seals missing or damaged, etc.) and engine performance degradation (fuel consumption increase for a given thrust). Thus, after several years of service, the performance databases and the APM associated to a specific aircraft are no longer representative enough of the actual aircraft performance. It is important to monitor the trend of the performance deterioration and correct the uncertainties of the aircraft model in order to improve the accuracy the flight management system predictions. The basis of this research lies in the new ability to continuously update an Aircraft Performance Model (APM) during flight using an adaptive lookup table technique. This methodology was developed and applied to the well-known Cessna Citation X business aircraft. For the purpose of this study, a level D Research Aircraft Flight Simulator (RAFS) was used as a test aircraft. According to Federal Aviation Administration the level D is the highest certification level for the flight dynamics modeling. Basically, using data available in the Flight Crew Operating Manual (FCOM), a first APM describing the variation of the engine fan speed and aircraft fuel flow w.r.t flight conditions was derived. This model was next improved using the proposed methodology. To do that, several cruise flights were performed using the RAFS. An algorithm was developed to frequently sample the aircraft sensors measurements during the flight and compare the model prediction with the actual measurements. Based on these comparisons, a correction was performed on the actual APM in order to minimize the error between the predicted data and the measured data. In this way, as the aircraft flies, the APM will be continuously enhanced, making the FMS more and more precise and the prediction of trajectories more realistic and more reliable. The results obtained are very encouraging. Indeed, using the tables initialized with the FCOM data, only a few iterations were needed to reduce the fuel flow prediction error from an average relative error of 12% to 0.3%. Similarly, the FCOM prediction regarding the engine fan speed was reduced from a maximum error deviation of 5.0% to 0.2% after only ten flights.

Keywords: aircraft performance, cruise, trajectory optimization, adaptive lookup tables, Cessna Citation X

Procedia PDF Downloads 252
7517 Empowering Middle School Math Coordinators as Agents of Transformation: The Impact of the Mitar Program on Mathematical Literacy and Social-Emotional Learning Integration

Authors: Saleit Ron

Abstract:

The Mitar program was established to drive a shift in middle school mathematics education, emphasizing the connection of math to real-life situations, exploring mathematical modeling and literacy, and integrating social and emotional learning (SEL) components for enhanced excellence. The program envisions math coordinators as catalysts for change, equipping them to create educational materials, strengthen leadership skills, and develop SEL competencies within coordinator communities. These skills are then employed to lead transformative efforts within their respective schools. The program engaged 90 participants across six math coordinator communities during 2022-2023, involving 30-60 hours of annual learning. The process includes formative and summative evaluations through questionnaires and interviews, revealing participants' high contentment and successful integration of acquired skills into their schools. Reflections from participants highlighted the need for enhanced change leadership processes, often seeking more personalized mentoring to navigate challenges effectively.

Keywords: math coordinators, mathematical literacy, mathematical modeling, SEL competencies

Procedia PDF Downloads 41
7516 Exploring the Non-Verbalizable in Conservation Grazing: The Contradictions Illuminated by a ‘Go-Along’ Methodology

Authors: James Ormrod

Abstract:

This paper is concerned with volunteer livestock checking. Based on a pilot study consisting of ‘go-along’ interviews with livestock checkers, it argues that there are limitations to the insights that can be generated from approaches to ‘discourse analysis’ that would focus only on the verbalizable aspects of the practice. Volunteer livestock checking takes place across Europe as part of conservation projects aimed at maintaining particular habitats through the reintroduction of grazing animals. Volunteers are variously called ‘urban shepherds’, because these practices often take place on urban fringes, or ‘lookerers’, as their role is to make visual checks on the animals. Pilot research that took place on the South Downs (a chalk downland habitat on the South Coast of the UK) involved researchers accompanying volunteers as they checked on livestock. They were asked to give an account of what they were doing and then answer semi-structured interview questions. Participants drew on popular discourses on conservation and biodiversity, as framed by the local council who run the programme. They also framed their relationships to the animals in respect to the more formal limitations of their role as identified through the conservation programme. And yet these discourses, significant as they are, do not adequately explain why volunteers are drawn to, and emotionally invested in, lookering. The methodology employed allowed participants instead to gesture to features of the landscape and to recall memories, and for the researchers to see how volunteers interacted with the animals and the landscape in embodied and emotionally loaded ways. The paper argues that a psychosocial perspective that pays attention to the contradictions and tensions made visible through this methodology helps develop a fuller understanding of volunteer livestock checking as a social practice.

Keywords: conservation, human-animal relations, lookering, volunteering

Procedia PDF Downloads 120
7515 The Customization of 3D Last Form Design Based on Weighted Blending

Authors: Shih-Wen Hsiao, Chu-Hsuan Lee, Rong-Qi Chen

Abstract:

When it comes to last, it is regarded as the critical foundation of shoe design and development. Not only the last relates to the comfort of shoes wearing but also it aids the production of shoe styling and manufacturing. In order to enhance the efficiency and application of last development, a computer aided methodology for customized last form designs is proposed in this study. The reverse engineering is mainly applied to the process of scanning for the last form. Then the minimum energy is used for the revision of surface continuity, the surface of the last is reconstructed with the feature curves of the scanned last. When the surface of a last is reconstructed, based on the foundation of the proposed last form reconstruction module, the weighted arithmetic mean method is applied to the calculation on the shape morphing which differs from the grading for the control mesh of last, and the algorithm of subdivision is used to create the surface of last mesh, thus the feet-fitting 3D last form of different sizes is generated from its original form feature with functions remained. Finally, the practicability of the proposed methodology is verified through later case studies.

Keywords: 3D last design, customization, reverse engineering, weighted morphing, shape blending

Procedia PDF Downloads 329
7514 Optimization of Operational Water Quality Parameters in a Drinking Water Distribution System Using Response Surface Methodology

Authors: Sina Moradi, Christopher W. K. Chow, John Van Leeuwen, David Cook, Mary Drikas, Patrick Hayde, Rose Amal

Abstract:

Chloramine is commonly used as a disinfectant in drinking water distribution systems (DWDSs), particularly in Australia and the USA. Maintaining a chloramine residual throughout the DWDS is important in ensuring microbiologically safe water is supplied at the customer’s tap. In order to simulate how chloramine behaves when it moves through the distribution system, a water quality network model (WQNM) can be applied. In this work, the WQNM was based on mono-chloramine decomposition reactions, which enabled prediction of mono-chloramine residual at different locations through a DWDS in Australia, using the Bentley commercial hydraulic package (Water GEMS). The accuracy of WQNM predictions is influenced by a number of water quality parameters. Optimization of these parameters in order to obtain the closest results in comparison with actual measured data in a real DWDS would result in both cost reduction as well as reduction in consumption of valuable resources such as energy and materials. In this work, the optimum operating conditions of water quality parameters (i.e. temperature, pH, and initial mono-chloramine concentration) to maximize the accuracy of mono-chloramine residual predictions for two water supply scenarios in an entire network were determined using response surface methodology (RSM). To obtain feasible and economical water quality parameters for highest model predictability, Design Expert 8.0 software (Stat-Ease, Inc.) was applied to conduct the optimization of three independent water quality parameters. High and low levels of the water quality parameters were considered, inevitably, as explicit constraints, in order to avoid extrapolation. The independent variables were pH, temperature and initial mono-chloramine concentration. The lower and upper limits of each variable for two water supply scenarios were defined and the experimental levels for each variable were selected based on the actual conditions in studied DWDS. It was found that at pH of 7.75, temperature of 34.16 ºC, and initial mono-chloramine concentration of 3.89 (mg/L) during peak water supply patterns, root mean square error (RMSE) of WQNM for the whole network would be minimized to 0.189, and the optimum conditions for averaged water supply occurred at pH of 7.71, temperature of 18.12 ºC, and initial mono-chloramine concentration of 4.60 (mg/L). The proposed methodology to predict mono-chloramine residual can have a great potential for water treatment plant operators in accurately estimating the mono-chloramine residual through a water distribution network. Additional studies from other water distribution systems are warranted to confirm the applicability of the proposed methodology for other water samples.

Keywords: chloramine decay, modelling, response surface methodology, water quality parameters

Procedia PDF Downloads 210
7513 Implementation of Building Information Modeling in Turkish Government Sector Projects

Authors: Mohammad Lemar Zalmai, Mustafa Nabi Kocakaya, Cemil Akcay, Ekrem Manisali

Abstract:

In recent years, the Building Information Modeling (BIM) approach has been developed expeditiously. As people see the benefits of this approach, it has begun to be used widely in construction projects and some countries made it mandatory to get more benefits from it. To promote the implementation of BIM in construction projects, it will be helpful to get some relevant information from surveys and interviews. The purpose of this study is to research the current adoption and implementation of BIM in public projects in Turkey. This study specified the challenges of BIM implementation in Turkey and proposed some solutions to overcome them. In this context, the challenges for BIM implementation and the factors that affect the BIM usage are determined based on previous academic researches and expert opinions by conducting interviews and questionnaire surveys. Several methods are used to process information in order to obtain weights of different factors to make BIM widespread in Turkey. This study concluded interviews' and questionnaire surveys' outcomes and proposed some suggestions to promote the implementation of BIM in Turkey. We believe research findings will be a good reference for boosting BIM implementation in Turkey.

Keywords: building information modelling, BIM implementations, Turkish construction industry, Turkish government sector projects

Procedia PDF Downloads 123
7512 Insect Outbreaks, Harvesting and Wildfire in Forests: Mathematical Models for Coupling Disturbances

Authors: M. C. A. Leite, B. Chen-Charpentier, F. Agusto

Abstract:

A long-term goal of sustainable forest management is a relatively stable source of wood and a stable forest age-class structure has become the goal of many forest management practices. In the absence of disturbances, this forest management goal could easily be achieved. However, in the face of recurring insect outbreaks and other disruptive processes forest planning becomes more difficult, requiring knowledge of the effects on the forest of a wide variety of environmental factors (e.g., habitat heterogeneity, fire size and frequency, harvesting, insect outbreaks, and age distributions). The association between distinct forest disturbances and the potential effect on forest dynamics is a complex matter, particularly when evaluated over time and at large scale, and is not well understood. However, gaining knowledge in this area is crucial for a sustainable forest management. Mathematical modeling is a tool that can be used to broader the understanding in this area. In this talk we will introduce mathematical models formulation incorporating the effect of insect outbreaks either as a single disturbance in the forest population dynamics or coupled with other disturbances: either wildfire or harvesting. The results and ecological insights will be discussed.

Keywords: age-structured forest population, disturbances interaction, harvesting insects outbreak dynamics, mathematical modeling

Procedia PDF Downloads 510
7511 Benchmarking of Pentesting Tools

Authors: Esteban Alejandro Armas Vega, Ana Lucila Sandoval Orozco, Luis Javier García Villalba

Abstract:

The benchmarking of tools for dynamic analysis of vulnerabilities in web applications is something that is done periodically, because these tools from time to time update their knowledge base and search algorithms, in order to improve their accuracy. Unfortunately, the vast majority of these evaluations are made by software enthusiasts who publish their results on blogs or on non-academic websites and always with the same evaluation methodology. Similarly, academics who have carried out this type of analysis from a scientific approach, the majority, make their analysis within the same methodology as well the empirical authors. This paper is based on the interest of finding answers to questions that many users of this type of tools have been asking over the years, such as, to know if the tool truly test and evaluate every vulnerability that it ensures do, or if the tool, really, deliver a real report of all the vulnerabilities tested and exploited. This kind of questions have also motivated previous work but without real answers. The aim of this paper is to show results that truly answer, at least on the tested tools, all those unanswered questions. All the results have been obtained by changing the common model of benchmarking used for all those previous works.

Keywords: cybersecurity, IDS, security, web scanners, web vulnerabilities

Procedia PDF Downloads 305