Search results for: nonlinear static analysis
26884 Linearization of Y-Force Equation of Rigid Body Equation of Motion and Behavior of Fighter Aircraft under Imbalance Weight on Wings during Combat
Authors: Jawad Zakir, Syed Irtiza Ali Shah, Rana Shaharyar, Sidra Mahmood
Abstract:
Y-force equation comprises aerodynamic forces, drag and side force with side slip angle β and weight component along with the coupled roll (φ) and pitch angles (θ). This research deals with the linearization of Y-force equation using Small Disturbance theory assuming equilibrium flight conditions for different state variables of aircraft. By using assumptions of Small Disturbance theory in non-linear Y-force equation, finally reached at linearized lateral rigid body equation of motion; which says that in linearized Y-force equation, the lateral acceleration is dependent on the other different aerodynamic and propulsive forces like vertical tail, change in roll rate (Δp) from equilibrium, change in yaw rate (Δr) from equilibrium, change in lateral velocity due to side force, drag and side force components due to side slip, and the lateral equation from coupled rotating frame to decoupled rotating frame. This paper describes implementation of this lateral linearized equation for aircraft control systems. Another significant parameter considered on which y-force equation depends is ‘c’ which shows that any change bought in the weight of aircrafts wing will cause Δφ and cause lateral force i.e. Y_c. This simplification also leads to lateral static and dynamic stability. The linearization of equations is required because much of mathematics control system design for aircraft is based on linear equations. This technique is simple and eases the linearization of the rigid body equations of motion without using any high-speed computers.Keywords: Y-force linearization, small disturbance theory, side slip, aerodynamic force drag, lateral rigid body equation of motion
Procedia PDF Downloads 50226883 Process Monitoring Based on Parameterless Self-Organizing Map
Authors: Young Jae Choung, Seoung Bum Kim
Abstract:
Statistical Process Control (SPC) is a popular technique for process monitoring. A widely used tool in SPC is a control chart, which is used to detect the abnormal status of a process and maintain the controlled status of the process. Traditional control charts, such as Hotelling’s T2 control chart, are effective techniques to detect abnormal observations and monitor processes. However, many complicated manufacturing systems exhibit nonlinearity because of the different demands of the market. In this case, the unregulated use of a traditional linear modeling approach may not be effective. In reality, many industrial processes contain the nonlinear and time-varying properties because of the fluctuation of process raw materials, slowing shift of the set points, aging of the main process components, seasoning effects, and catalyst deactivation. The use of traditional SPC techniques with time-varying data will degrade the performance of the monitoring scheme. To address these issues, in the present study, we propose a parameterless self-organizing map (PLSOM)-based control chart. The PLSOM-based control chart not only can manage a situation where the distribution or parameter of the target observations changes, but also address the nonlinearity of modern manufacturing systems. The control limits of the proposed PLSOM chart are established by estimating the empirical level of significance on the percentile using a bootstrap method. Experimental results with simulated data and actual process data from a thin-film transistor-liquid crystal display process demonstrated the effectiveness and usefulness of the proposed chart.Keywords: control chart, parameter-less self-organizing map, self-organizing map, time-varying property
Procedia PDF Downloads 27926882 Co-Gasification Process for Green and Blue Hydrogen Production: Innovative Process Development, Economic Analysis, and Exergy Assessment
Authors: Yousaf Ayub
Abstract:
A co-gasification process, which involves the utilization of both biomass and plastic waste, has been developed to enable the production of blue and green hydrogen. To support this endeavor, an Aspen Plus simulation model has been meticulously created, and sustainability analysis is being conducted, focusing on economic viability, energy efficiency, advanced exergy considerations, and exergoeconomics evaluations. In terms of economic analysis, the process has demonstrated strong economic sustainability, as evidenced by an internal rate of return (IRR) of 8% at a process efficiency level of 70%. At present, the process has the potential to generate approximately 1100 kWh of electric power, with any excess electricity, beyond meeting the process requirements, capable of being harnessed for green hydrogen production via an alkaline electrolysis cell (AEC). This surplus electricity translates to a potential daily hydrogen production of around 200 kg. The exergy analysis of the model highlights that the gasifier component exhibits the lowest exergy efficiency, resulting in the highest energy losses, amounting to approximately 40%. Additionally, advanced exergy analysis findings pinpoint the gasifier as the primary source of exergy destruction, totaling around 9000 kW, with associated exergoeconomics costs amounting to 6500 $/h. Consequently, improving the gasifier's performance is a critical focal point for enhancing the overall sustainability of the process, encompassing energy, exergy, and economic considerations.Keywords: blue hydrogen, green hydrogen, co-gasification, waste valorization, exergy analysis
Procedia PDF Downloads 7226881 Laboratory Simulation of Subway Dynamic Stray Current Interference with Cathodically Protected Structures
Authors: Mohammad Derakhshani, Saeed Reza Allahkaram, Michael Isakani-Zakaria, Masoud Samadian, Hojat Sharifi Rasaey
Abstract:
Dynamic stray currents tend to change their magnitude and polarity with time at their source which will create anodic and cathodic spots on a nearby interfered structure. To date, one of the biggest known dynamic stray current sources are DC traction systems. Laboratory simulation is a suitable method to apply theoretical principles in order to identify effective parameters in dynamic stray current influenced corrosion. Simulation techniques can be utilized for various mitigation methods applied in a small scales for selection of the most efficient method with regards to field applications. In this research, laboratory simulation of potential fluctuations caused by dynamic stray current on a cathodically protected structure was investigated. A lab model capable of generating DC static and dynamic stray currents and simulating its effects on cathodically protected samples were developed based on stray current induced (contact-less) polarization technique. Stray current pick-up and discharge spots on an influenced structure were simulated by inducing fluctuations in the sample’s stationary potential. Two mitigation methods for dynamic stray current interference on buried structures namely application of sacrificial anodes as preferred discharge point for the stray current and potentially controlled cathodic protection was investigated. Results showed that the application of sacrificial anodes can be effective in reducing interference only in discharge spot. But cathodic protection through potential controlling is more suitable for mitigating dynamic stray current effects.Keywords: simulation, dynamic stray current, fluctuating potentials, sacrificial anode
Procedia PDF Downloads 30426880 Fractal Analysis of Polyacrylamide-Graphene Oxide Composite Gels
Authors: Gülşen Akın Evingür, Önder Pekcan
Abstract:
The fractal analysis is a bridge between the microstructure and macroscopic properties of gels. Fractal structure is usually provided to define the complexity of crosslinked molecules. The complexity in gel systems is described by the fractal dimension (Df). In this study, polyacrylamide- graphene oxide (GO) composite gels were prepared by free radical crosslinking copolymerization. The fractal analysis of polyacrylamide- graphene oxide (GO) composite gels were analyzed in various GO contents during gelation and were investigated by using Fluorescence Technique. The analysis was applied to estimate Df s of the composite gels. Fractal dimension of the polymer composite gels were estimated based on the power law exponent values using scaling models. In addition, here we aimed to present the geometrical distribution of GO during gelation. And we observed that as gelation proceeded GO plates first organized themselves into 3D percolation cluster with Df=2.52, then goes to diffusion limited clusters with Df =1.4 and then lines up to Von Koch curve with random interval with Df=1.14. Here, our goal is to try to interpret the low conductivity and/or broad forbidden gap of GO doped PAAm gels, by the distribution of GO in the final form of the produced gel.Keywords: composite gels, fluorescence, fractal, scaling
Procedia PDF Downloads 31026879 Reinforcement-Learning Based Handover Optimization for Cellular Unmanned Aerial Vehicles Connectivity
Authors: Mahmoud Almasri, Xavier Marjou, Fanny Parzysz
Abstract:
The demand for services provided by Unmanned Aerial Vehicles (UAVs) is increasing pervasively across several sectors including potential public safety, economic, and delivery services. As the number of applications using UAVs grows rapidly, more and more powerful, quality of service, and power efficient computing units are necessary. Recently, cellular technology draws more attention to connectivity that can ensure reliable and flexible communications services for UAVs. In cellular technology, flying with a high speed and altitude is subject to several key challenges, such as frequent handovers (HOs), high interference levels, connectivity coverage holes, etc. Additional HOs may lead to “ping-pong” between the UAVs and the serving cells resulting in a decrease of the quality of service and energy consumption. In order to optimize the number of HOs, we develop in this paper a Q-learning-based algorithm. While existing works focus on adjusting the number of HOs in a static network topology, we take into account the impact of cells deployment for three different simulation scenarios (Rural, Semi-rural and Urban areas). We also consider the impact of the decision distance, where the drone has the choice to make a switching decision on the number of HOs. Our results show that a Q-learning-based algorithm allows to significantly reduce the average number of HOs compared to a baseline case where the drone always selects the cell with the highest received signal. Moreover, we also propose which hyper-parameters have the largest impact on the number of HOs in the three tested environments, i.e. Rural, Semi-rural, or Urban.Keywords: drones connectivity, reinforcement learning, handovers optimization, decision distance
Procedia PDF Downloads 11726878 A Network Approach to Analyzing Financial Markets
Authors: Yusuf Seedat
Abstract:
The necessity to understand global financial markets has increased following the unfortunate spread of the recent financial crisis around the world. Financial markets are considered to be complex systems consisting of highly volatile move-ments whose indexes fluctuate without any clear pattern. Analytic methods of stock prices have been proposed in which financial markets are modeled using common network analysis tools and methods. It has been found that two key components of social network analysis are relevant to modeling financial markets, allowing us to forecast accurate predictions of stock prices within the financial market. Financial markets have a number of interacting components, leading to complex behavioral patterns. This paper describes a social network approach to analyzing financial markets as a viable approach to studying the way complex stock markets function. We also look at how social network analysis techniques and metrics are used to gauge an understanding of the evolution of financial markets as well as how community detection can be used to qualify and quantify in-fluence within a network.Keywords: network analysis, social networks, financial markets, stocks, nodes, edges, complex networks
Procedia PDF Downloads 19726877 A Sensor Placement Methodology for Chemical Plants
Authors: Omid Ataei Nia, Karim Salahshoor
Abstract:
In this paper, a new precise and reliable sensor network methodology is introduced for unit processes and operations using the Constriction Coefficient Particle Swarm Optimization (CPSO) method. CPSO is introduced as a new search engine for optimal sensor network design purposes. Furthermore, a Square Root Unscented Kalman Filter (SRUKF) algorithm is employed as a new data reconciliation technique to enhance the stability and accuracy of the filter. The proposed design procedure incorporates precision, cost, observability, reliability together with importance-of-variables (IVs) as a novel measure in Instrumentation Criteria (IC). To the best of our knowledge, no comprehensive approach has yet been proposed in the literature to take into account the importance of variables in the sensor network design procedure. In this paper, specific weight is assigned to each sensor, measuring a process variable in the sensor network to indicate the importance of that variable over the others to cater to the ultimate sensor network application requirements. A set of distinct scenarios has been conducted to evaluate the performance of the proposed methodology in a simulated Continuous Stirred Tank Reactor (CSTR) as a highly nonlinear process plant benchmark. The obtained results reveal the efficacy of the proposed method, leading to significant improvement in accuracy with respect to other alternative sensor network design approaches and securing the definite allocation of sensors to the most important process variables in sensor network design as a novel achievement.Keywords: constriction coefficient PSO, importance of variable, MRMSE, reliability, sensor network design, square root unscented Kalman filter
Procedia PDF Downloads 16426876 A Non-Linear Eddy Viscosity Model for Turbulent Natural Convection in Geophysical Flows
Authors: J. P. Panda, K. Sasmal, H. V. Warrior
Abstract:
Eddy viscosity models in turbulence modeling can be mainly classified as linear and nonlinear models. Linear formulations are simple and require less computational resources but have the disadvantage that they cannot predict actual flow pattern in complex geophysical flows where streamline curvature and swirling motion are predominant. A constitutive equation of Reynolds stress anisotropy is adopted for the formulation of eddy viscosity including all the possible higher order terms quadratic in the mean velocity gradients, and a simplified model is developed for actual oceanic flows where only the vertical velocity gradients are important. The new model is incorporated into the one dimensional General Ocean Turbulence Model (GOTM). Two realistic oceanic test cases (OWS Papa and FLEX' 76) have been investigated. The new model predictions match well with the observational data and are better in comparison to the predictions of the two equation k-epsilon model. The proposed model can be easily incorporated in the three dimensional Princeton Ocean Model (POM) to simulate a wide range of oceanic processes. Practically, this model can be implemented in the coastal regions where trasverse shear induces higher vorticity, and for prediction of flow in estuaries and lakes, where depth is comparatively less. The model predictions of marine turbulence and other related data (e.g. Sea surface temperature, Surface heat flux and vertical temperature profile) can be utilized in short term ocean and climate forecasting and warning systems.Keywords: Eddy viscosity, turbulence modeling, GOTM, CFD
Procedia PDF Downloads 20426875 Non-Linear Regression Modeling for Composite Distributions
Authors: Mostafa Aminzadeh, Min Deng
Abstract:
Modeling loss data is an important part of actuarial science. Actuaries use models to predict future losses and manage financial risk, which can be beneficial for marketing purposes. In the insurance industry, small claims happen frequently while large claims are rare. Traditional distributions such as Normal, Exponential, and inverse-Gaussian are not suitable for describing insurance data, which often show skewness and fat tails. Several authors have studied classical and Bayesian inference for parameters of composite distributions, such as Exponential-Pareto, Weibull-Pareto, and Inverse Gamma-Pareto. These models separate small to moderate losses from large losses using a threshold parameter. This research introduces a computational approach using a nonlinear regression model for loss data that relies on multiple predictors. Simulation studies were conducted to assess the accuracy of the proposed estimation method. The simulations confirmed that the proposed method provides precise estimates for regression parameters. It's important to note that this approach can be applied to datasets if goodness-of-fit tests confirm that the composite distribution under study fits the data well. To demonstrate the computations, a real data set from the insurance industry is analyzed. A Mathematica code uses the Fisher information algorithm as an iteration method to obtain the maximum likelihood estimation (MLE) of regression parameters.Keywords: maximum likelihood estimation, fisher scoring method, non-linear regression models, composite distributions
Procedia PDF Downloads 4026874 Research of the Three-Dimensional Visualization Geological Modeling of Mine Based on Surpac
Authors: Honggang Qu, Yong Xu, Rongmei Liu, Zhenji Gao, Bin Wang
Abstract:
Today's mining industry is advancing gradually toward digital and visual direction. The three-dimensional visualization geological modeling of mine is the digital characterization of mineral deposits and is one of the key technology of digital mining. Three-dimensional geological modeling is a technology that combines geological spatial information management, geological interpretation, geological spatial analysis and prediction, geostatistical analysis, entity content analysis and graphic visualization in a three-dimensional environment with computer technology and is used in geological analysis. In this paper, the three-dimensional geological modeling of an iron mine through the use of Surpac is constructed, and the weight difference of the estimation methods between the distance power inverse ratio method and ordinary kriging is studied, and the ore body volume and reserves are simulated and calculated by using these two methods. Compared with the actual mine reserves, its result is relatively accurate, so it provides scientific bases for mine resource assessment, reserve calculation, mining design and so on.Keywords: three-dimensional geological modeling, geological database, geostatistics, block model
Procedia PDF Downloads 8326873 Accessibility and Visibility through Space Syntax Analysis of the Linga Raj Temple in Odisha, India
Authors: S. Pramanik
Abstract:
Since the early ages, the Hindu temples have been interpreted through various Vedic philosophies. These temples are visited by pilgrims which demonstrate the rituals and religious belief of communities, reflecting a variety of actions and behaviors. Darsana— a direct seeing, is a part of the pilgrimage activity. During the process of Darsana, a devotee is prepared for entry in the temple to realize the cognizing Truth culminating in visualizing the idol of God, placed at the Garbhagriha (sanctum sanctorum). For this, the pilgrim must pass through a sequential arrangement of spaces. During the process of progress, the pilgrims visualize the spaces differently from various points of views. The viewpoints create a variety of spatial patterns in the minds of pilgrims coherent to the Hindu philosophies. The space organization and its order are perceived by various techniques of spatial analysis. A temple, as examples of Kalinga stylistic variations, has been chosen for the study. This paper intends to demonstrate some visual patterns generated during the process of Darsana (visibility) and its accessibility by Point Isovist Studies and Visibility Graph Analysis from the entrance (Simha Dwara) to The Sanctum sanctorum (Garbhagriha).Keywords: Hindu temple architecture, point isovist, space syntax analysis, visibility graph analysis
Procedia PDF Downloads 12426872 Low Voltage and High Field-Effect Mobility Thin Film Transistor Using Crystalline Polymer Nanocomposite as Gate Dielectric
Authors: Debabrata Bhadra, B. K. Chaudhuri
Abstract:
The operation of organic thin film transistors (OFETs) with low voltage is currently a prevailing issue. We have fabricated anthracene thin-film transistor (TFT) with an ultrathin layer (~450nm) of Poly-vinylidene fluoride (PVDF)/CuO nanocomposites as a gate insulator. We obtained a device with excellent electrical characteristics at low operating voltages (<1V). Different layers of the film were also prepared to achieve the best optimization of ideal gate insulator with various static dielectric constant (εr ). Capacitance density, leakage current at 1V gate voltage and electrical characteristics of OFETs with a single and multi layer films were investigated. This device was found to have highest field effect mobility of 2.27 cm2/Vs, a threshold voltage of 0.34V, an exceptionally low sub threshold slope of 380 mV/decade and an on/off ratio of 106. Such favorable combination of properties means that these OFETs can be utilized successfully as voltages below 1V. A very simple fabrication process has been used along with step wise poling process for enhancing the pyroelectric effects on the device performance. The output characteristic of OFET after poling were changed and exhibited linear current-voltage relationship showing the evidence of large polarization. The temperature dependent response of the device was also investigated. The stable performance of the OFET after poling operation makes it reliable in temperature sensor applications. Such High-ε CuO/PVDF gate dielectric appears to be highly promising candidates for organic non-volatile memory and sensor field-effect transistors (FETs).Keywords: organic field effect transistors, thin film transistor, gate dielectric, organic semiconductor
Procedia PDF Downloads 24626871 Tracing the Evolution of English and Urdu Languages: A Linguistic and Cultural Analysis
Authors: Aamna Zafar
Abstract:
Through linguistic and cultural analysis, this study seeks to trace the development of the English and Urdu languages. Along with examining how the vocabulary and syntax of English and Urdu have evolved over time and the linguistic trends that may be seen in these changes, this study will also look at the historical and cultural influences that have shaped the languages throughout time. The study will also look at how English and Urdu have changed over time, both in terms of language use and communication inside each other's cultures and globally. We'll research how these changes affect social relations and cultural identity, as well as how they might affect the future of these languages.Keywords: linguistic and cultural analysis, historical factors, cultural factors, vocabulary, syntax, significance
Procedia PDF Downloads 8126870 Effect of Aryl Imidazolium Ionic Liquids as Asphaltene Dispersants
Authors: Raghda Ahmed El-Nagar
Abstract:
Oil spills are one of the most serious environmental issues that have occurred during the production and transportation of petroleum crude oil. Chemical asphaltene dispersants are hazardous to the marine environment, so Ionic liquids (ILs) as asphaltene dispersants are a critical area of study. In this work, different aryl imidazolium ionic liquids were synthesized with high yield and elucidated via tools of analysis (Elemental analysis, FT-IR, and 1H-NMR). Thermogravimetric analysis confirmed that the prepared ILs posses high thermal stability. The critical micelle concentration (CMC), surface tension, and emulsification index were investigated. Evaluation of synthesized ILs as asphaltene dispersants were assessed at various concentrations, and data reveals high dispersion efficiency.Keywords: ionic liquids, oil spill, asphaltene dispersants, CMC, efficiency
Procedia PDF Downloads 19726869 Optimization of Element Type for FE Model and Verification of Analyses with Physical Tests
Authors: Mustafa Tufekci, Caner Guven
Abstract:
In Automotive Industry, sliding door systems that are also used as body closures, are safety members. Extreme product tests are realized to prevent failures in a design process, but these tests realized experimentally result in high costs. Finite element analysis is an effective tool used for the design process. These analyses are used before production of a prototype for validation of design according to customer requirement. In result of this, the substantial amount of time and cost is saved. Finite element model is created for geometries that are designed in 3D CAD programs. Different element types as bar, shell and solid, can be used for creating mesh model. The cheaper model can be created by the selection of element type, but combination of element type that was used in model, number and geometry of element and degrees of freedom affects the analysis result. Sliding door system is a good example which used these methods for this study. Structural analysis was realized for sliding door mechanism by using FE models. As well, physical tests that have same boundary conditions with FE models were realized. Comparison study for these element types, were done regarding test and analyses results then the optimum combination was achieved.Keywords: finite element analysis, sliding door mechanism, element type, structural analysis
Procedia PDF Downloads 33226868 Flashover Detection Algorithm Based on Mother Function
Authors: John A. Morales, Guillermo Guidi, B. M. Keune
Abstract:
Electric Power supply is a crucial topic for economic and social development. Power outages statistics show that discharges atmospherics are imperative phenomena to produce those outages. In this context, it is necessary to correctly detect when overhead line insulators are faulted. In this paper, an algorithm to detect if a lightning stroke generates or not permanent fault on insulator strings is proposed. On top of that, lightning stroke simulations developed by using the Alternative Transients Program, are used. Based on these insights, a novel approach is designed that depends on mother functions analysis corresponding to the given variance-covariance matrix. Signals registered at the insulator string are projected on corresponding axes by the means of Principal Component Analysis. By exploiting these new axes, it is possible to determine a flashover characteristic zone useful to a good insulation design. The proposed methodology for flashover detection extends the existing approaches for the analysis and study of lightning performance on transmission lines.Keywords: mother function, outages, lightning, sensitivity analysis
Procedia PDF Downloads 58926867 Application of Artificial Neural Network in Assessing Fill Slope Stability
Authors: An-Jui. Li, Kelvin Lim, Chien-Kuo Chiu, Benson Hsiung
Abstract:
This paper details the utilization of artificial intelligence (AI) in the field of slope stability whereby quick and convenient solutions can be obtained using the developed tool. The AI tool used in this study is the artificial neural network (ANN), while the slope stability analysis methods are the finite element limit analysis methods. The developed tool allows for the prompt prediction of the safety factors of fill slopes and their corresponding probability of failure (depending on the degree of variation of the soil parameters), which can give the practicing engineer a reasonable basis in their decision making. In fact, the successful use of the Extreme Learning Machine (ELM) algorithm shows that slope stability analysis is no longer confined to the conventional methods of modeling, which at times may be tedious and repetitive during the preliminary design stage where the focus is more on cost saving options rather than detailed design. Therefore, similar ANN-based tools can be further developed to assist engineers in this aspect.Keywords: landslide, limit analysis, artificial neural network, soil properties
Procedia PDF Downloads 21226866 Advancement of Computer Science Research in Nigeria: A Bibliometric Analysis of the Past Three Decades
Authors: Temidayo O. Omotehinwa, David O. Oyewola, Friday J. Agbo
Abstract:
This study aims to gather a proper perspective of the development landscape of Computer Science research in Nigeria. Therefore, a bibliometric analysis of 4,333 bibliographic records of Computer Science research in Nigeria in the last 31 years (1991-2021) was carried out. The bibliographic data were extracted from the Scopus database and analyzed using VOSviewer and the bibliometrix R package through the biblioshiny web interface. The findings of this study revealed that Computer Science research in Nigeria has a growth rate of 24.19%. The most developed and well-studied research areas in the Computer Science field in Nigeria are machine learning, data mining, and deep learning. The social structure analysis result revealed that there is a need for improved international collaborations. Sparsely established collaborations are largely influenced by geographic proximity. The funding analysis result showed that Computer Science research in Nigeria is under-funded. The findings of this study will be useful for researchers conducting Computer Science related research. Experts can gain insights into how to develop a strategic framework that will advance the field in a more impactful manner. Government agencies and policymakers can also utilize the outcome of this research to develop strategies for improved funding for Computer Science research.Keywords: bibliometric analysis, biblioshiny, computer science, Nigeria, science mapping
Procedia PDF Downloads 11726865 Experiment and Analytical Study on Fire Resistance Performance of Slot Type Concrete-Filled Tube
Authors: Bum Yean Cho, Heung-Youl Kim, Ki-Seok Kwon, Kang-Su Kim
Abstract:
In this study, a full-scale test and analysis (numerical analysis) of fire resistance performance of bare CFT column on which slot was used instead of existing welding method to connect the steel pipe on the concrete-filled tube were conducted. Welded CFT column is known to be vulnerable to high or low temperature because of low brittleness of welding part. As a result of a fire resistance performance test of slot CFT column after removing the welding part and fixing it by a slot which was folded into the tube, slot type CFT column indicated the improved fire resistance performance than welded CFT column by 28% or more. And as a result of conducting finite element analysis of slot type column using ABAQUS, analysis result proved the reliability of the test result in predicting the fire behavior and fire resistance hour.Keywords: CFT (concrete-filled tube) column, fire resistance performance, slot, weld
Procedia PDF Downloads 18826864 Decision Support Tool for Green Roofs Selection: A Multicriteria Analysis
Authors: I. Teotónio, C.O. Cruz, C.M. Silva, M. Manso
Abstract:
Diverse stakeholders show different concerns when choosing green roof systems. Also, green roof solutions vary in their cost and performance. Therefore, decision-makers continually face the difficult task of balancing benefits against green roofs costs. Decision analysis methods, as multicriteria analysis, can be used when the decision‑making process includes different perspectives, multiple objectives, and uncertainty. The present study adopts a multicriteria decision model to evaluate the installation of green roofs in buildings, determining the solution with the best trade-off between costs and benefits in agreement with the preferences of the users/investors. This methodology was applied to a real decision problem, assessing the preferences between different green roof systems in an existing building in Lisbon. This approach supports the decision-making process on green roofs and enables robust and informed decisions on urban planning while optimizing buildings retrofitting.Keywords: decision making, green roofs, investors preferences, multicriteria analysis, sustainable development
Procedia PDF Downloads 18726863 A Study of the Relationship between Time Management Behaviour and Job Satisfaction of Higher Education Institutes in India
Authors: Sania K. Rao, Feza T. Azmi
Abstract:
The purpose of the present study is to explore the relationship between time management behaviour and job satisfaction of academicians of higher education institutes in India. The analyses of this study were carried out with AMOS (version 20.0); and Confirmatory Factor Analysis (CFA) and Structural Equation Modelling (SEM) were conducted. The factor analysis and findings show that perceived control of time serves as the partial mediating factor to have a significant and positive influence on job satisfaction. Further, at the end, a number of suggestions to improve one’s time management behaviour were provided.Keywords: time management behaviour, job satisfaction, higher education, India, mediation analysis
Procedia PDF Downloads 39526862 Boundary Conditions for 2D Site Response Analysis in OpenSees
Authors: M. Eskandarighadi, C. R. McGann
Abstract:
It is observed from past experiences of earthquakes that local site conditions can significantly affect the strong ground motion characteristicssuch as frequency content, amplitude, and duration of seismic waves. The most common method for investigating site response is one-dimensional seismic site response analysis. The infinite horizontal length of the model and the homogeneous characteristic of the soil are crucial assumptions of this method. One boundary condition that can be used in the sides is tying the sides horizontally for vertical 1D wave propagation. However, 1D analysis cannot account for the 2D nature of wave propagation in the condition where the soil profile is not fully horizontal or has heterogeneity within layers. Therefore, 2D seismic site response analysis can be used to take all of these limitations into account for a better understanding of local site conditions. Different types of boundary conditions can be appliedin 2D site response models, such as tied boundary condition, massive columns, and free-field boundary condition. The tied boundary condition has been used in 1D analysis, which is useful for 1D wave propagation. Employing two massive columns at the sides is another approach for capturing the 2D nature of wave propagation. Free-field boundary condition can simulate the free-field motion that would exist far from the domain of interest. The goal for free-field boundary condition is to minimize the unwanted reflection from sides. This research focuses on the comparison between these methods with examples and discusses the details and limitations of each of these boundary conditions.Keywords: boundary condition, free-field, massive columns, opensees, site response analysis, wave propagation
Procedia PDF Downloads 19126861 Extracting the Coupled Dynamics in Thin-Walled Beams from Numerical Data Bases
Authors: Mohammad A. Bani-Khaled
Abstract:
In this work we use the Discrete Proper Orthogonal Decomposition transform to characterize the properties of coupled dynamics in thin-walled beams by exploiting numerical simulations obtained from finite element simulations. The outcomes of the will improve our understanding of the linear and nonlinear coupled behavior of thin-walled beams structures. Thin-walled beams have widespread usage in modern engineering application in both large scale structures (aeronautical structures), as well as in nano-structures (nano-tubes). Therefore, detailed knowledge in regard to the properties of coupled vibrations and buckling in these structures are of great interest in the research community. Due to the geometric complexity in the overall structure and in particular in the cross-sections it is necessary to involve computational mechanics to numerically simulate the dynamics. In using numerical computational techniques, it is not necessary to over simplify a model in order to solve the equations of motions. Computational dynamics methods produce databases of controlled resolution in time and space. These numerical databases contain information on the properties of the coupled dynamics. In order to extract the system dynamic properties and strength of coupling among the various fields of the motion, processing techniques are required. Time- Proper Orthogonal Decomposition transform is a powerful tool for processing databases for the dynamics. It will be used to study the coupled dynamics of thin-walled basic structures. These structures are ideal to form a basis for a systematic study of coupled dynamics in structures of complex geometry.Keywords: coupled dynamics, geometric complexity, proper orthogonal decomposition (POD), thin walled beams
Procedia PDF Downloads 42126860 A Lean Manufacturing Profile of Practices in the Metallurgical Industry: A Methodology for Multivariate Analysis
Authors: M. Jonathan D. Morales, R. Ramón Silva
Abstract:
The purpose of this project is to carry out an analysis and determine the profile of actual lean manufacturing processes in the Metropolitan Area of Bucaramanga. Through the analysis of qualitative and quantitative variables it was possible to establish how these manufacturers develop production practices that ensure their competitiveness and productivity in the market. In this study, a random sample of metallurgic and wrought iron companies was applied, following which a quantitative focus and analysis was used to formulate a qualitative methodology for measuring the level of lean manufacturing procedures in the industry. A qualitative evaluation was also carried out through a multivariate analysis using the Numerical Taxonomy System (NTSYS) program which should allow for the determination of Lean Manufacturing profiles. Through the results it was possible to observe how the companies in the sector are doing with respect to Lean Manufacturing Practices, as well as identify the level of management that these companies practice with respect to this topic. In addition, it was possible to ascertain that there is no one dominant profile in the sector when it comes to Lean Manufacturing. It was established that the companies in the metallurgic and wrought iron industry show low levels of Lean Manufacturing implementation. Each one carries out diverse actions that are insufficient to consolidate a sectoral strategy for developing a competitive advantage which enables them to tie together a production strategy.Keywords: production line management, metallurgic industry, lean manufacturing, productivity
Procedia PDF Downloads 46426859 On the Estimation of Crime Rate in the Southwest of Nigeria: Principal Component Analysis Approach
Authors: Kayode Balogun, Femi Ayoola
Abstract:
Crime is at alarming rate in this part of world and there are many factors that are contributing to this antisocietal behaviour both among the youths and old. In this work, principal component analysis (PCA) was used as a tool to reduce the dimensionality and to really know those variables that were crime prone in the study region. Data were collected on twenty-eight crime variables from National Bureau of Statistics (NBS) databank for a period of fifteen years, while retaining as much of the information as possible. We use PCA in this study to know the number of major variables and contributors to the crime in the Southwest Nigeria. The results of our analysis revealed that there were eight principal variables have been retained using the Scree plot and Loading plot which implies an eight-equation solution will be appropriate for the data. The eight components explained 93.81% of the total variation in the data set. We also found that the highest and commonly committed crimes in the Southwestern Nigeria were: Assault, Grievous Harm and Wounding, theft/stealing, burglary, house breaking, false pretence, unlawful arms possession and breach of public peace.Keywords: crime rates, data, Southwest Nigeria, principal component analysis, variables
Procedia PDF Downloads 45026858 Morphological Analysis of Manipuri Language: Wahei-Neinarol
Authors: Y. Bablu Singh, B. S. Purkayashtha, Chungkham Yashawanta Singh
Abstract:
Morphological analysis forms the basic foundation in NLP applications including syntax parsing Machine Translation (MT), Information Retrieval (IR) and automatic indexing in all languages. It is the field of the linguistics; it can provide valuable information for computer based linguistics task such as lemmatization and studies of internal structure of the words. Computational Morphology is the application of morphological rules in the field of computational linguistics, and it is the emerging area in AI, which studies the structure of words, which are formed by combining smaller units of linguistics information, called morphemes: the building blocks of words. Morphological analysis provides about semantic and syntactic role in a sentence. It analyzes the Manipuri word forms and produces several grammatical information associated with the words. The Morphological Analyzer for Manipuri has been tested on 3500 Manipuri words in Shakti Standard format (SSF) using Meitei Mayek as source; thereby an accuracy of 80% has been obtained on a manual check.Keywords: morphological analysis, machine translation, computational morphology, information retrieval, SSF
Procedia PDF Downloads 32726857 Agile Implementation of 'PULL' Principles in a Manufacturing Process Chain for Aerospace Composite Parts
Authors: Torsten Mielitz, Dietmar Schulz, York C. Roth
Abstract:
Market forecasts show a significant increase in the demand for aircraft within the next two decades and production rates will be adapted accordingly. Improvements and optimizations in the industrial system are becoming more important to cope with future challenges in manufacturing and assembly. Highest quality standards have to be met for aerospace parts, whereas cost effective production in industrial systems and methodologies are also a key driver. A look at other industries like e.g., automotive shows well established processes to streamline existing manufacturing systems. In this paper, the implementation of 'PULL' principles in an existing manufacturing process chain for a large scale composite part is presented. A nonlinear extrapolation based on 'Little's Law' showed a risk of a significant increase of parts needed in the process chain to meet future demand. A project has been set up to mitigate the risk whereas the methodology has been changed from a traditional milestone approach in the beginning towards an agile way of working in the end in order to facilitate immediate benefits in the shop-floor. Finally, delivery rates could be increased avoiding more semi-finished parts in the process chain (work in progress & inventory) by the successful implementation of the 'PULL' philosophy in the shop-floor between the work stations. Lessons learned during the running project as well as implementation and operations phases are discussed in order to share best practices.Keywords: aerospace composite part manufacturing, PULL principles, shop-floor implementation, lessons learned
Procedia PDF Downloads 17726856 Artificial Intelligence Assisted Sentiment Analysis of Hotel Reviews Using Topic Modeling
Authors: Sushma Ghogale
Abstract:
With a surge in user-generated content or feedback or reviews on the internet, it has become possible and important to know consumers' opinions about products and services. This data is important for both potential customers and businesses providing the services. Data from social media is attracting significant attention and has become the most prominent channel of expressing an unregulated opinion. Prospective customers look for reviews from experienced customers before deciding to buy a product or service. Several websites provide a platform for users to post their feedback for the provider and potential customers. However, the biggest challenge in analyzing such data is in extracting latent features and providing term-level analysis of the data. This paper proposes an approach to use topic modeling to classify the reviews into topics and conduct sentiment analysis to mine the opinions. This approach can analyse and classify latent topics mentioned by reviewers on business sites or review sites, or social media using topic modeling to identify the importance of each topic. It is followed by sentiment analysis to assess the satisfaction level of each topic. This approach provides a classification of hotel reviews using multiple machine learning techniques and comparing different classifiers to mine the opinions of user reviews through sentiment analysis. This experiment concludes that Multinomial Naïve Bayes classifier produces higher accuracy than other classifiers.Keywords: latent Dirichlet allocation, topic modeling, text classification, sentiment analysis
Procedia PDF Downloads 10426855 Text Analysis to Support Structuring and Modelling a Public Policy Problem-Outline of an Algorithm to Extract Inferences from Textual Data
Authors: Claudia Ehrentraut, Osama Ibrahim, Hercules Dalianis
Abstract:
Policy making situations are real-world problems that exhibit complexity in that they are composed of many interrelated problems and issues. To be effective, policies must holistically address the complexity of the situation rather than propose solutions to single problems. Formulating and understanding the situation and its complex dynamics, therefore, is a key to finding holistic solutions. Analysis of text based information on the policy problem, using Natural Language Processing (NLP) and Text analysis techniques, can support modelling of public policy problem situations in a more objective way based on domain experts knowledge and scientific evidence. The objective behind this study is to support modelling of public policy problem situations, using text analysis of verbal descriptions of the problem. We propose a formal methodology for analysis of qualitative data from multiple information sources on a policy problem to construct a causal diagram of the problem. The analysis process aims at identifying key variables, linking them by cause-effect relationships and mapping that structure into a graphical representation that is adequate for designing action alternatives, i.e., policy options. This study describes the outline of an algorithm used to automate the initial step of a larger methodological approach, which is so far done manually. In this initial step, inferences about key variables and their interrelationships are extracted from textual data to support a better problem structuring. A small prototype for this step is also presented.Keywords: public policy, problem structuring, qualitative analysis, natural language processing, algorithm, inference extraction
Procedia PDF Downloads 596