Search results for: dynamic modelling
3870 Strengthening by Assessment: A Case Study of Rail Bridges
Authors: Evangelos G. Ilias, Panagiotis G. Ilias, Vasileios T. Popotas
Abstract:
The United Kingdom has one of the oldest railway networks in the world dating back to 1825 when the world’s first passenger railway was opened. The network has some 40,000 bridges of various construction types using a wide range of materials including masonry, steel, cast iron, wrought iron, concrete and timber. It is commonly accepted that the successful operation of the network is vital for the economy of the United Kingdom, consequently the cost effective maintenance of the existing infrastructure is a high priority to maintain the operability of the network, prevent deterioration and to extend the life of the assets. Every bridge on the railway network is required to be assessed every eighteen years and a structured approach to assessments is adopted with three main types of progressively more detailed assessments used. These assessment types include Level 0 (standardized spreadsheet assessment tools), Level 1 (analytical hand calculations) and Level 2 (generally finite element analyses). There is a degree of conservatism in the first two types of assessment dictated to some extent by the relevant standards which can lead to some structures not achieving the required load rating. In these situations, a Level 2 Assessment is often carried out using finite element analysis to uncover ‘latent strength’ and improve the load rating. If successful, the more sophisticated analysis can save on costly strengthening or replacement works and avoid disruption to the operational railway. This paper presents the ‘strengthening by assessment’ achieved by Level 2 analyses. The use of more accurate analysis assumptions and the implementation of non-linear modelling and functions (material, geometric and support) to better understand buckling modes and the structural behaviour of historic construction details that are not specifically covered by assessment codes are outlined. Metallic bridges which are susceptible to loss of section size through corrosion have largest scope for improvement by the Level 2 Assessment methodology. Three case studies are presented, demonstrating the effectiveness of the sophisticated Level 2 Assessment methodology using finite element analysis against the conservative approaches employed for Level 0 and Level 1 Assessments. One rail overbridge and two rail underbridges that did not achieve the required load rating by means of a Level 1 Assessment due to the inadequate restraint provided by U-Frame action are examined and the increase in assessed capacity given by the Level 2 Assessment is outlined.Keywords: assessment, bridges, buckling, finite element analysis, non-linear modelling, strengthening
Procedia PDF Downloads 3093869 A Real-Time Moving Object Detection and Tracking Scheme and Its Implementation for Video Surveillance System
Authors: Mulugeta K. Tefera, Xiaolong Yang, Jian Liu
Abstract:
Detection and tracking of moving objects are very important in many application contexts such as detection and recognition of people, visual surveillance and automatic generation of video effect and so on. However, the task of detecting a real shape of an object in motion becomes tricky due to various challenges like dynamic scene changes, presence of shadow, and illumination variations due to light switch. For such systems, once the moving object is detected, tracking is also a crucial step for those applications that used in military defense, video surveillance, human computer interaction, and medical diagnostics as well as in commercial fields such as video games. In this paper, an object presents in dynamic background is detected using adaptive mixture of Gaussian based analysis of the video sequences. Then the detected moving object is tracked using the region based moving object tracking and inter-frame differential mechanisms to address the partial overlapping and occlusion problems. Firstly, the detection algorithm effectively detects and extracts the moving object target by enhancing and post processing morphological operations. Secondly, the extracted object uses region based moving object tracking and inter-frame difference to improve the tracking speed of real-time moving objects in different video frames. Finally, the plotting method was applied to detect the moving objects effectively and describes the object’s motion being tracked. The experiment has been performed on image sequences acquired both indoor and outdoor environments and one stationary and web camera has been used.Keywords: background modeling, Gaussian mixture model, inter-frame difference, object detection and tracking, video surveillance
Procedia PDF Downloads 4773868 A Study on Numerical Modelling of Rigid Pavement: Temperature and Thickness Effect
Authors: Amin Chegenizadeh, Mahdi Keramatikerman, Hamid Nikraz
Abstract:
Pavement engineering plays a significant role to develop cost effective and efficient highway and road networks. In general, pavement regarding structure is categorized in two core group namely flexible and rigid pavements. There are various benefits in application of rigid pavement. For instance, they have a longer life and lower maintenance costs in compare with the flexible pavement. In rigid pavement designs, temperature and thickness are two effective parameters that could widely affect the total cost of the project. In this study, a numerical modeling using Kenpave-Kenslab was performed to investigate the effect of these two important parameters in the rigid pavement.Keywords: rigid pavement, Kenpave, Kenslab, thickness, temperature
Procedia PDF Downloads 3733867 Ethicality of Algorithmic Pricing and Consumers’ Resistance
Authors: Zainab Atia, Hongwei He, Panagiotis Sarantopoulos
Abstract:
Over the past few years, firms have witnessed a massive increase in sophisticated algorithmic deployment, which has become quite pervasive in today’s modern society. With the wide availability of data for retailers, the ability to track consumers using algorithmic pricing has become an integral option in online platforms. As more companies are transforming their businesses and relying more on massive technological advancement, pricing algorithmic systems have brought attention and given rise to its wide adoption, with many accompanying benefits and challenges to be found within its usage. With the overall aim of increasing profits by organizations, algorithmic pricing is becoming a sound option by enabling suppliers to cut costs, allowing better services, improving efficiency and product availability, and enhancing overall consumer experiences. The adoption of algorithms in retail has been pioneered and widely used in literature across varied fields, including marketing, computer science, engineering, economics, and public policy. However, what is more, alarming today is the comprehensive understanding and focus of this technology and its associated ethical influence on consumers’ perceptions and behaviours. Indeed, due to algorithmic ethical concerns, consumers are found to be reluctant in some instances to share their personal data with retailers, which reduces their retention and leads to negative consumer outcomes in some instances. This, in its turn, raises the question of whether firms can still manifest the acceptance of such technologies by consumers while minimizing the ethical transgressions accompanied by their deployment. As recent modest research within the area of marketing and consumer behavior, the current research advances the literature on algorithmic pricing, pricing ethics, consumers’ perceptions, and price fairness literature. With its empirical focus, this paper aims to contribute to the literature by applying the distinction of the two common types of algorithmic pricing, dynamic and personalized, while measuring their relative effect on consumers’ behavioural outcomes. From a managerial perspective, this research offers significant implications that pertain to providing a better human-machine interactive environment (whether online or offline) to improve both businesses’ overall performance and consumers’ wellbeing. Therefore, by allowing more transparent pricing systems, businesses can harness their generated ethical strategies, which fosters consumers’ loyalty and extend their post-purchase behaviour. Thus, by defining the correct balance of pricing and right measures, whether using dynamic or personalized (or both), managers can hence approach consumers more ethically while taking their expectations and responses at a critical stance.Keywords: algorithmic pricing, dynamic pricing, personalized pricing, price ethicality
Procedia PDF Downloads 913866 Geographic Information System and Dynamic Segmentation of Very High Resolution Images for the Semi-Automatic Extraction of Sandy Accumulation
Authors: A. Bensaid, T. Mostephaoui, R. Nedjai
Abstract:
A considerable area of Algerian lands is threatened by the phenomenon of wind erosion. For a long time, wind erosion and its associated harmful effects on the natural environment have posed a serious threat, especially in the arid regions of the country. In recent years, as a result of increases in the irrational exploitation of natural resources (fodder) and extensive land clearing, wind erosion has particularly accentuated. The extent of degradation in the arid region of the Algerian Mecheria department generated a new situation characterized by the reduction of vegetation cover, the decrease of land productivity, as well as sand encroachment on urban development zones. In this study, we attempt to investigate the potential of remote sensing and geographic information systems for detecting the spatial dynamics of the ancient dune cords based on the numerical processing of LANDSAT images (5, 7, and 8) of three scenes 197/37, 198/36 and 198/37 for the year 2020. As a second step, we prospect the use of geospatial techniques to monitor the progression of sand dunes on developed (urban) lands as well as on the formation of sandy accumulations (dune, dunes fields, nebkha, barkhane, etc.). For this purpose, this study made use of the semi-automatic processing method for the dynamic segmentation of images with very high spatial resolution (SENTINEL-2 and Google Earth). This study was able to demonstrate that urban lands under current conditions are located in sand transit zones that are mobilized by the winds from the northwest and southwest directions.Keywords: land development, GIS, segmentation, remote sensing
Procedia PDF Downloads 1553865 Development of Noninvasive Method to Analyze Dynamic Changes of Matrix Stiffness and Elasticity Characteristics
Authors: Elena Petersen, Inna Kornienko, Svetlana Guryeva, Sergey Dobdin, Anatoly Skripal, Andrey Usanov, Dmitry Usanov
Abstract:
One of the most important unsolved problems in modern medicine is the increase of chronic diseases that lead to organ dysfunction or even complete loss of function. Current methods of treatment do not result in decreased mortality and disability statistics. Currently, the best treatment for many patients is still transplantation of organs and/or tissues. Therefore, finding a way of correct artificial matrix biofabrication in case of limited number of natural organs for transplantation is a critical task. One important problem that needs to be solved is development of a nondestructive and noninvasive method to analyze dynamic changes of mechanical characteristics of a matrix with minimal side effects on the growing cells. This research was focused on investigating the properties of matrix as a marker of graft condition. In this study, the collagen gel with human primary dermal fibroblasts in suspension (60, 120, 240*103 cells/mL) and collagen gel with cell spheroids were used as model objects. The stiffness and elasticity characteristics were evaluated by a semiconductor laser autodyne. The time and cell concentration dependency of the stiffness and elasticity were investigated. It was shown that these properties changed in a non-linear manner with respect to cell concentration. The maximum matrix stiffness was observed in the collagen gel with the cell concentration of 120*103 cells/mL. This study proved the opportunity to use the mechanical properties of matrix as a marker of graft condition, which can be measured by noninvasive semiconductor laser autodyne technique.Keywords: graft, matrix, noninvasive method, regenerative medicine, semiconductor laser autodyne
Procedia PDF Downloads 3443864 Predictions of Thermo-Hydrodynamic State for Single and Three Pads Gas Foil Bearings Operating at Steady-State Based on Multi-Physics Coupling Computer Aided Engineering Simulations
Authors: Tai Yuan Yu, Pei-Jen Wang
Abstract:
Oil-free turbomachinery is considered one of the critical technologies for future green power generation systems as rotor machinery systems. Oil-free technology allows clean, compact, and maintenance-free working, and gas foil bearings, abbreviated as GFBs, are important for the technology. Since the first applications in the auxiliary power units and air cycle machines in the 1970s, obvious improvement has been created to the computational models for dynamic rotor behavior. However, many technical issues are still poorly understood or remain unsolved, and some of those are thermal management and the pattern of how pressure will be distributed in bearing clearance. This paper presents a three-dimensional, abbreviated as 3D, fluid-structure interaction model of single pad foil bearings and three pad foil bearings to predict bearing working behavior that researchers could compare characteristics of those. The coupling analysis model involves dynamic working characteristics applied to all the gas film and mechanical structures. Therefore, the elastic deformation of foil structure and the hydrodynamic pressure of gas film can both be calculated by a finite element method program. As a result, the temperature distribution pattern could also be iteratively solved by coupling analysis. In conclusion, the working fluid state in a gas film of various pad forms of bearings working characteristic at constant rotational speed for both can be solved for comparisons with the experimental results.Keywords: fluid-structure interaction, multi-physics simulations, gas foil bearing, oil-free, transient thermo-hydrodynamic
Procedia PDF Downloads 1633863 Defects Estimation of Embedded Systems Components by a Bond Graph Approach
Authors: I. Gahlouz, A. Chellil
Abstract:
The paper concerns the estimation of system components faults by using an unknown inputs observer. To reach this goal, we used the Bond Graph approach to physical modelling. We showed that this graphical tool is allowing the representation of system components faults as unknown inputs within the state representation of the considered physical system. The study of the causal and structural features of the system (controllability, observability, finite structure, and infinite structure) based on the Bond Graph approach was hence fulfilled in order to design an unknown inputs observer which is used for the system component fault estimation.Keywords: estimation, bond graph, controllability, observability
Procedia PDF Downloads 4133862 Thermomechanical Damage Modeling of F114 Carbon Steel
Authors: A. El Amri, M. El Yakhloufi Haddou, A. Khamlichi
Abstract:
The numerical simulation based on the Finite Element Method (FEM) is widely used in academic institutes and in the industry. It is a useful tool to predict many phenomena present in the classical manufacturing forming processes such as fracture. But, the results of such numerical model depend strongly on the parameters of the constitutive behavior model. The influences of thermal and mechanical loads cause damage. The temperature and strain rate dependent materials’ properties and their modelling are discussed. A Johnson-Cook Model of damage has been selected for the numerical simulations. Virtual software called the ABAQUS 6.11 is used for finite element analysis. This model was introduced in order to give information concerning crack initiation during thermal and mechanical loads.Keywords: thermo-mechanical fatigue, failure, numerical simulation, fracture, damage
Procedia PDF Downloads 3933861 Assessment of Pakistan-China Economic Corridor: An Emerging Dynamic of 21st Century
Authors: Naad-E-Ali Sulehria
Abstract:
Pakistan and china have stepped in a new phase of strengthening fraternity as the dream of economic corridor once discerned by both countries is going to take a pragmatic shape. Pak-China economic corridor an under construction program is termed to be an emerging dynamic of 21st century that anticipates a nexus between Asian continent and Indian Ocean by extending its functions to adjoining East, South, Central and Western Asian regions. The $45.6 billion worth heavily invested megaprojects by China are meant to revive energy sector and building economic infrastructure in Pakistan. Evidently, these projects are a part of ‘southern extension’ of Silk Road economic belt which is going to draw out prominent incentives for both countries particularly bolstering China to acquire influential dominance over the regional trade and beyond. In pursuit to adhere, by these progressive plans both countries have began working on their respective assignments. This article discusses the economical development programs under China’s peripheral diplomacy regarding its region-specific-approach to accumulate trade of Persian Gulf and access the landlocked Central Asian states through Pakistan in a sublimate context to break US encirclement of Asia. Pakistan’s utmost preference to utilize its strategic channel as a trade hub to become an emerging economy and surpass its arch-rival India for strategic concerns is contemplated accordingly. The needs and feasibility of the economic gateway and the dividends it can provide in the contemporary scenario are examined carefully and analysis is drawn upon the future prospects of the Pakistan-China Economic corridor once completed.Keywords: pak-china economic corridor (PCEC), central asian republic states (CARs), new silk road economic belt, gawadar
Procedia PDF Downloads 3673860 A Finite Element Study of Laminitis in Horses
Authors: Naeim Akbari Shahkhosravi, Reza Kakavand, Helen M. S. Davies, Amin Komeili
Abstract:
Equine locomotion and performance are significantly affected by hoof health. One of the most critical diseases of the hoof is laminitis, which can lead to horse lameness in a severe condition. This disease exhibits the mechanical properties degradation of the laminar junction tissue within the hoof. Therefore, it is essential to investigate the biomechanics of the hoof, focusing specifically on excessive and cumulatively accumulated stresses within the laminar junction tissue. For this aim, the current study generated a novel equine hoof Finite Element (FE) model under dynamic physiological loading conditions and employing a hyperelastic material model. Associated tissues of the equine hoof were segmented from computed tomography scans of an equine forelimb, including the navicular bone, third phalanx, sole, frog, laminar junction, digital cushion, and medial- dorsal- lateral wall areas. The inner tissues were connected based on the hoof anatomy, and the hoof was under a dynamic loading over cyclic strides at the trot. The strain distribution on the hoof wall of the model was compared with the published in vivo strain measurements to validate the model. Then the validated model was used to study the development of laminitis. The ultimate stress tolerated by the laminar junction before rupture was considered as a stress threshold. The tissue damage was simulated through iterative reduction of the tissue’s mechanical properties in the presence of excessive maximum principal stresses. The findings of this investigation revealed how damage initiates from the medial and lateral sides of the tissue and propagates through the hoof dorsal area.Keywords: horse hoof, laminitis, finite element model, continuous damage
Procedia PDF Downloads 1823859 Numerical Assessment of Fire Characteristics with Bodies Engulfed in Hydrocarbon Pool Fire
Authors: Siva Kumar Bathina, Sudheer Siddapureddy
Abstract:
Fires accident becomes even worse when the hazardous equipment like reactors or radioactive waste packages are engulfed in fire. In this work, large-eddy numerical fire simulations are performed using fire dynamic simulator to predict the thermal behavior of such bodies engulfed in hydrocarbon pool fires. A radiatively dominated 0.3 m circular burner with n-heptane as the fuel is considered in this work. The fire numerical simulation results without anybody inside the fire are validated with the reported experimental data. The comparison is in good agreement for different flame properties like predicted mass burning rate, flame height, time-averaged center-line temperature, time-averaged center-line velocity, puffing frequency, the irradiance at the surroundings, and the radiative heat feedback to the pool surface. Cask of different sizes is simulated with SS304L material. The results are independent of the material of the cask simulated as the adiabatic surface temperature concept is employed in this study. It is observed that the mass burning rate increases with the blockage ratio (3% ≤ B ≤ 32%). However, the change in this increment is reduced at higher blockage ratios (B > 14%). This is because the radiative heat feedback to the fuel surface is not only from the flame but also from the cask volume. As B increases, the volume of the cask increases and thereby increases the radiative contribution to the fuel surface. The radiative heat feedback in the case of the cask engulfed in the fire is increased by 2.5% to 31% compared to the fire without cask.Keywords: adiabatic surface temperature, fire accidents, fire dynamic simulator, radiative heat feedback
Procedia PDF Downloads 1273858 Predictive Modelling of Curcuminoid Bioaccessibility as a Function of Food Formulation and Associated Properties
Authors: Kevin De Castro Cogle, Mirian Kubo, Maria Anastasiadi, Fady Mohareb, Claire Rossi
Abstract:
Background: The bioaccessibility of bioactive compounds is a critical determinant of the nutritional quality of various food products. Despite its importance, there is a limited number of comprehensive studies aimed at assessing how the composition of a food matrix influences the bioaccessibility of a compound of interest. This knowledge gap has prompted a growing need to investigate the intricate relationship between food matrix formulations and the bioaccessibility of bioactive compounds. One such class of bioactive compounds that has attracted considerable attention is curcuminoids. These naturally occurring phytochemicals, extracted from the roots of Curcuma longa, have gained popularity owing to their purported health benefits and also well known for their poor bioaccessibility Project aim: The primary objective of this research project is to systematically assess the influence of matrix composition on the bioaccessibility of curcuminoids. Additionally, this study aimed to develop a series of predictive models for bioaccessibility, providing valuable insights for optimising the formula for functional foods and provide more descriptive nutritional information to potential consumers. Methods: Food formulations enriched with curcuminoids were subjected to in vitro digestion simulation, and their bioaccessibility was characterized with chromatographic and spectrophotometric techniques. The resulting data served as the foundation for the development of predictive models capable of estimating bioaccessibility based on specific physicochemical properties of the food matrices. Results: One striking finding of this study was the strong correlation observed between the concentration of macronutrients within the food formulations and the bioaccessibility of curcuminoids. In fact, macronutrient content emerged as a very informative explanatory variable of bioaccessibility and was used, alongside other variables, as predictors in a Bayesian hierarchical model that predicted curcuminoid bioaccessibility accurately (optimisation performance of 0.97 R2) for the majority of cross-validated test formulations (LOOCV of 0.92 R2). These preliminary results open the door to further exploration, enabling researchers to investigate a broader spectrum of food matrix types and additional properties that may influence bioaccessibility. Conclusions: This research sheds light on the intricate interplay between food matrix composition and the bioaccessibility of curcuminoids. This study lays a foundation for future investigations, offering a promising avenue for advancing our understanding of bioactive compound bioaccessibility and its implications for the food industry and informed consumer choices.Keywords: bioactive bioaccessibility, food formulation, food matrix, machine learning, probabilistic modelling
Procedia PDF Downloads 683857 Regret-Regression for Multi-Armed Bandit Problem
Authors: Deyadeen Ali Alshibani
Abstract:
In the literature, the multi-armed bandit problem as a statistical decision model of an agent trying to optimize his decisions while improving his information at the same time. There are several different algorithms models and their applications on this problem. In this paper, we evaluate the Regret-regression through comparing with Q-learning method. A simulation on determination of optimal treatment regime is presented in detail.Keywords: optimal, bandit problem, optimization, dynamic programming
Procedia PDF Downloads 4533856 The Power of Words: The Use of Language in Ethan Frome
Authors: Ritu Sharma
Abstract:
In order to be objective, critics must examine the dynamic relationships between the author, the reader, the text, and the outside world. However, it is also crucial to recognize that because the language was created by God, meaning is ingrained in it. Meaning is located in and discovered through literature rather than being limited to the author, reader, text, or the outside world. The link between the author, the reader, and the text is crucial because literature unites an author and a reader through the use of language. Literature is a potent kind of communication, and Ethan Frome's audience is forever changed as a result of the book's language and the language its characters use. The narrative of Ethan Frome and his wife Zeena is presented in Ethan Frome. Ethan's story is told throughout the course of the book, revealed through the eyes of the narrator, an outsider passing through Starkfield, as well as through the insight that the narrator gains from the townspeople and his stay on the Frome farm. The story is set in the rural New England community of Starkfield, Massachusetts. The weather provides the ideal setting for Ethan and the narrator to get to know one another as the narrator gets preoccupied with unraveling the narrative that underlies Ethan's physical anomalies. In addition to telling a gripping tale and capturing human nature as it is, Ethan Frome uses its storyline to achieve something more significant. The book by Edith Wharton supports language. Zeena's deliberate and convincing language challenges relativity and meaninglessness. Ethan and Mattie's effort to effectively use words reflects the complexity of language, and their battle illustrates the influence that language may have if and when it is used. Ethan Frome defends the written word, the foundation upon which it is constructed, as a literary work. Communication is based on language, and as the characters respond to and get involved in disputes throughout the book, Zeena, Ethan, and Mattie, each reflects particular theories of communication that help define their uses of communication within the broader context of language.Keywords: dynamic relationships, potent, communication, complexity
Procedia PDF Downloads 913855 Proactive Competence Management for Employees: A Bottom-up Process Model for Developing Target Competence Profiles Based on the Employee's Tasks
Authors: Maximilian Cedzich, Ingo Dietz Von Bayer, Roland Jochem
Abstract:
In order for industrial companies to continue to succeed in dynamic, globalized markets, they must be able to train their employees in an agile manner and at short notice in line with the exogenous conditions that arise. For this purpose, it is indispensable to operate a proactive competence management system for employees that recognizes qualification needs timely in order to be able to address them promptly through qualification measures. However, there are hardly any approaches to be found in the literature that includes systematic, proactive competence management. In order to help close this gap, this publication presents a process model that systematically develops bottom-up, future-oriented target competence profiles based on the tasks of the employees. Concretely, in the first step, the tasks of the individual employees are examined for assumed future conditions. In other words, qualitative scenarios are considered for the individual tasks to determine how they are likely to change. In a second step, these scenario-based future tasks are translated into individual future-related target competencies of the employee using a matrix of generic task properties. The final step pursues the goal of validating the target competence profiles formed in this way within the framework of a management workshop. This process model provides industrial companies with a tool that they can use to determine the competencies required by their own employees in the future and compare them with the actual prevailing competencies. If gaps are identified between the target and the actual, these qualification requirements can be closed in the short term by means of qualification measures.Keywords: dynamic globalized markets, employee competence management, industrial companies, knowledge management
Procedia PDF Downloads 1893854 Experimental Research of Smoke Impact on the Performance of Cylindrical Eight Channel Cyclone
Authors: Pranas Baltrėnas, Dainius Paliulis
Abstract:
Cyclones are widely used for separating particles from gas in energy production objects. Efficiency of normal centrifugal air cleaning devices ranges from 85 to 90%, but weakness of many cyclones is low collection efficiency of particles less than 10 μm in diameter. Many factors have impact on cyclone efficiency – humidity, temperature, gas (air) composition, airflow velocity and etc. Many scientists evaluated only effect of origin and size of PM on cyclone efficiency. Effect of gas (air) composition and temperature on cyclone efficiency still demands contributions. Complex experimental research on efficiency of cylindrical eight-channel system with adjustable half-rings for removing fine dispersive particles (< 20 μm) was carried out. The impact of gaseous smoke components on removal of wood ashes was analyzed. Gaseous components, present in the smoke mixture, with the dynamic viscosity lower than that of same temperature air, decrease the d50 value, simultaneously increasing the overall particulate matter removal efficiency in the cyclone, i.e. this effect is attributed to CO2 and CO, while O2 and NO have the opposite effect. Air temperature influences the d50 value, an increase in air temperature yields an increase in d50 value, i.e. the overall particulate matter removal efficiency declines, the reason for this being an increasing dynamic air viscosity. At 120 °C temperature the d50 value is approximately 11.8 % higher than at air temperature of 20 °C. With an increase in smoke (gas) temperature from 20 °C to 50 °C, the aerodynamic resistance in a 1-tier eight-channel cylindrical cyclone drops from 1605 to 1380 Pa, from 1660 to 1420 Pa in a 2-tier eight-channel cylindrical cyclone, from 1715 to 1450 Pa in a 3-tier eight-channel cylindrical cyclone. The reason for a decline in aerodynamic resistance is the declining gas density. The aim of the paper is to analyze the impact of gaseous smoke components on the eight–channel cyclone with tangential inlet.Keywords: cyclone, adjustable half-rings, particulate matter, efficiency, gaseous compounds, smoke
Procedia PDF Downloads 2893853 Dialectics of Modern Law: Perspectives and Strategies of Resistance from the Margins
Authors: Nisar Alungal Chungath
Abstract:
“No human being is illegal" has become a dictum strongly upheld in the context of global immigration and migration, highlighting the ethical and moral dimensions of how societies and governments treat individuals and communities who have crossed political borders or are living in a country without legal authorization. It seeks to shift the focus from categorizing human beings as illegal immigrants to recognizing their inherent human rights and the complexities of their circumstances. As a complex social phenomenon, law has been a crucial instrument in shaping, regulating and governing human societies and vice versa. The law has now become a humongous political project of the modern majoritarian regimes to democratically illegitimize and illegalize the unpopular sections and minorities. Drawing from the theoretical frameworks of dialectics, the paper explores the philosophical underpinnings of the historical evolution and dynamic nature of modern law. The paper employs a phenomenological approach to analyze the dialectical relations between individuals, societies, and legal systems, aiming to shed light on the ethical and political implications of these interactions. By examining the historical essence of law, its relationship with social and cultural norms, and the role of power dynamics, this article argues for constantly maintaining the dialectics of law—the dynamic interplay between legal norms, social practices, cultural values, and historical contexts through a philosophical and phenomenological lens, in order to bridge the gap between universal principles and particular contexts. The paper will shed light to the dialectics of the law in the context of instances of the legal persecutions of the modern secular democracies such as Citizenship Amendment Act-2019, India.Keywords: phenomenology, dialectic, modern law, politics, resistance, margins
Procedia PDF Downloads 563852 Modelling of Pervaporation Separation of Butanol from Aqueous Solutions Using Polydimethylsiloxane Mixed Matrix Membranes
Authors: Arian Ebneyamini, Hoda Azimi, Jules Thibaults, F. Handan Tezel
Abstract:
In this study, a modification of Hennepe model for pervaporation separation of butanol from aqueous solutions using Polydimethylsiloxane (PDMS) mixed matrix membranes has been introduced and validated by experimental data. The model was compared to the original Hennepe model and few other models which are applicable for membrane gas separation processes such as Maxwell, Lewis Nielson and Pal. Theoretical modifications for non-ideal interface morphology have been offered to predict the permeability in case of interface void, interface rigidification and pore-blockage. The model was in a good agreement with experimental data.Keywords: butanol, PDMS, modeling, pervaporation, mixed matrix membranes
Procedia PDF Downloads 2213851 Currency Exchange Rate Forecasts Using Quantile Regression
Authors: Yuzhi Cai
Abstract:
In this paper, we discuss a Bayesian approach to quantile autoregressive (QAR) time series model estimation and forecasting. Together with a combining forecasts technique, we then predict USD to GBP currency exchange rates. Combined forecasts contain all the information captured by the fitted QAR models at different quantile levels and are therefore better than those obtained from individual models. Our results show that an unequally weighted combining method performs better than other forecasting methodology. We found that a median AR model can perform well in point forecasting when the predictive density functions are symmetric. However, in practice, using the median AR model alone may involve the loss of information about the data captured by other QAR models. We recommend that combined forecasts should be used whenever possible.Keywords: combining forecasts, MCMC, predictive density functions, quantile forecasting, quantile modelling
Procedia PDF Downloads 2563850 Finite Element Model to Investigate the Dynamic Behavior of Ring-Stiffened Conical Shell Fully and Partially Filled with Fluid
Authors: Mohammadamin Esmaeilzadehazimi, Morteza Shayan Arani, Mohammad Toorani, Aouni Lakis
Abstract:
This study uses a hybrid finite element method to predict the dynamic behavior of both fully and partially-filled truncated conical shells stiffened with ring stiffeners. The method combines classical shell theory and the finite element method, and employs displacement functions derived from exact solutions of Sanders' shell equilibrium equations for conical shells. The shell-fluid interface is analyzed by utilizing the velocity potential, Bernoulli's equation, and impermeability conditions to determine an explicit expression for fluid pressure. The equations of motion presented in this study apply to both conical and cylindrical shells. This study presents the first comparison of the method applied to ring-stiffened shells with other numerical and experimental findings. Vibration frequencies for conical shells with various boundary conditions and geometries in a vacuum and filled with water are compared with experimental and numerical investigations, achieving good agreement. The study thoroughly investigates the influence of geometric parameters, stiffener quantity, semi-vertex cone angle, level of water filled in the cone, and applied boundary conditions on the natural frequency of fluid-loaded ring-stiffened conical shells, and draws some useful conclusions. The primary advantage of the current method is its use of a minimal number of finite elements while achieving highly accurate results.Keywords: finite element method, fluid–structure interaction, conical shell, natural frequency, ring-stiffener
Procedia PDF Downloads 783849 On the Monitoring of Structures and Soils by Tromograph
Authors: Magarò Floriana, Zinno Raffaele
Abstract:
Since 2009, with the coming into force of the January 14, 2008 Ministerial Decree "New technical standards for construction", and the explanatory ministerial circular N°.617 of February 2, 2009, the question of seismic hazard and the design of seismic-resistant structures in Italy has acquired increasing importance. One of the most discussed aspects in recent Italian and international scientific literature concerns the dynamic interaction between land and structure, and the effects which dynamic coupling may have on individual buildings. In effect, from systems dynamics, it is well known that resonance can have catastrophic effects on a stimulated system, leading to a response that is not compatible with the previsions in the design phase. The method used in this study to estimate the frequency of oscillation of the structure is as follows: the analysis of HVSR (Horizontal to Vertical Spectral Ratio) relations. This allows for evaluation of very simple oscillation frequencies for land and structures. The tool used for data acquisition is an experimental digital tromograph. This is an engineered development of the experimental Languamply RE 4500 tromograph, equipped with an engineered amplification circuit and improved electronically using extremely small electronic components (size of each individual amplifier 16 x 26 mm). This tromograph is a modular system, completely "free" and "open", designed to interface Windows, Linux, OSX and Android with the outside world. It an amplifier designed to carry out microtremor measurements, yet which will also be useful for seismological and seismic measurements in general. The development of single amplifiers of small dimension allows for a very clean signal since being able to position it a few centimetres from the geophone eliminates cable “antenna” phenomena, which is a necessary characteristic in seeking to have signals which are clean at the very low voltages to be measured.Keywords: microtremor, HVSR, tromograph, structural engineering
Procedia PDF Downloads 4093848 C-eXpress: A Web-Based Analysis Platform for Comparative Functional Genomics and Proteomics in Human Cancer Cell Line, NCI-60 as an Example
Authors: Chi-Ching Lee, Po-Jung Huang, Kuo-Yang Huang, Petrus Tang
Abstract:
Background: Recent advances in high-throughput research technologies such as new-generation sequencing and multi-dimensional liquid chromatography makes it possible to dissect the complete transcriptome and proteome in a single run for the first time. However, it is almost impossible for many laboratories to handle and analysis these “BIG” data without the support from a bioinformatics team. We aimed to provide a web-based analysis platform for users with only limited knowledge on bio-computing to study the functional genomics and proteomics. Method: We use NCI-60 as an example dataset to demonstrate the power of the web-based analysis platform and data delivering system: C-eXpress takes a simple text file that contain the standard NCBI gene or protein ID and expression levels (rpkm or fold) as input file to generate a distribution map of gene/protein expression levels in a heatmap diagram organized by color gradients. The diagram is hyper-linked to a dynamic html table that allows the users to filter the datasets based on various gene features. A dynamic summary chart is generated automatically after each filtering process. Results: We implemented an integrated database that contain pre-defined annotations such as gene/protein properties (ID, name, length, MW, pI); pathways based on KEGG and GO biological process; subcellular localization based on GO cellular component; functional classification based on GO molecular function, kinase, peptidase and transporter. Multiple ways of sorting of column and rows is also provided for comparative analysis and visualization of multiple samples.Keywords: cancer, visualization, database, functional annotation
Procedia PDF Downloads 6193847 Improving Predictions of Coastal Benthic Invertebrate Occurrence and Density Using a Multi-Scalar Approach
Authors: Stephanie Watson, Fabrice Stephenson, Conrad Pilditch, Carolyn Lundquist
Abstract:
Spatial data detailing both the distribution and density of functionally important marine species are needed to inform management decisions. Species distribution models (SDMs) have proven helpful in this regard; however, models often focus only on species occurrences derived from spatially expansive datasets and lack the resolution and detail required to inform regional management decisions. Boosted regression trees (BRT) were used to produce high-resolution SDMs (250 m) at two spatial scales predicting probability of occurrence, abundance (count per sample unit), density (count per km2) and uncertainty for seven coastal seafloor taxa that vary in habitat usage and distribution to examine prediction differences and implications for coastal management. We investigated if small scale regionally focussed models (82,000 km2) can provide improved predictions compared to data-rich national scale models (4.2 million km2). We explored the variability in predictions across model type (occurrence vs abundance) and model scale to determine if specific taxa models or model types are more robust to geographical variability. National scale occurrence models correlated well with broad-scale environmental predictors, resulting in higher AUC (Area under the receiver operating curve) and deviance explained scores; however, they tended to overpredict in the coastal environment and lacked spatially differentiated detail for some taxa. Regional models had lower overall performance, but for some taxa, spatial predictions were more differentiated at a localised ecological scale. National density models were often spatially refined and highlighted areas of ecological relevance producing more useful outputs than regional-scale models. The utility of a two-scale approach aids the selection of the most optimal combination of models to create a spatially informative density model, as results contrasted for specific taxa between model type and scale. However, it is vital that robust predictions of occurrence and abundance are generated as inputs for the combined density model as areas that do not spatially align between models can be discarded. This study demonstrates the variability in SDM outputs created over different geographical scales and highlights implications and opportunities for managers utilising these tools for regional conservation, particularly in data-limited environments.Keywords: Benthic ecology, spatial modelling, multi-scalar modelling, marine conservation.
Procedia PDF Downloads 773846 Free Vibration Analysis of Timoshenko Beams at Higher Modes with Central Concentrated Mass Using Coupled Displacement Field Method
Authors: K. Meera Saheb, K. Krishna Bhaskar
Abstract:
Complex structures used in many fields of engineering are made up of simple structural elements like beams, plates etc. These structural elements, sometimes carry concentrated masses at discrete points, and when subjected to severe dynamic environment tend to vibrate with large amplitudes. The frequency amplitude relationship is very much essential in determining the response of these structural elements subjected to the dynamic loads. For Timoshenko beams, the effects of shear deformation and rotary inertia are to be considered to evaluate the fundamental linear and nonlinear frequencies. A commonly used method for solving vibration problem is energy method, or a finite element analogue of the same. In the present Coupled Displacement Field method the number of undetermined coefficients is reduced to half when compared to the famous Rayleigh Ritz method, which significantly simplifies the procedure to solve the vibration problem. This is accomplished by using a coupling equation derived from the static equilibrium of the shear flexible structural element. The prime objective of the present paper here is to study, in detail, the effect of a central concentrated mass on the large amplitude free vibrations of uniform shear flexible beams. Accurate closed form expressions for linear frequency parameter for uniform shear flexible beams with a central concentrated mass was developed and the results are presented in digital form.Keywords: coupled displacement field, coupling equation, large amplitude vibrations, moderately thick plates
Procedia PDF Downloads 2263845 Smart BIM Documents - the Development of the Ontology-Based Tool for Employer Information Requirements (OntEIR), and its Transformation into SmartEIR
Authors: Shadan Dwairi
Abstract:
Defining proper requirements is one of the key factors for a successful construction projects. Although there have been many attempts put forward in assist in identifying requirements, but still this area is under developed. In Buildings Information Modelling (BIM) projects. The Employer Information Requirements (EIR) is the fundamental requirements document and a necessary ingredient in achieving a successful BIM project. The provision on full and clear EIR is essential to achieving BIM Level-2. As Defined by PAS 1192-2, EIR is a “pre-tender document that sets out the information to be delivered and the standards and processes to be adopted by the supplier as part of the project delivery process”. It also notes that “EIR should be incorporated into tender documentation to enable suppliers to produce an initial BIM Execution Plan (BEP)”. The importance of effective definition of EIR lies in its contribution to a better productivity during the construction process in terms of cost and time, in addition to improving the quality of the built asset. Proper and clear information is a key aspect of the EIR, in terms of the information it contains and more importantly the information the client receives at the end of the project that will enable the effective management and operation of the asset, where typically about 60%-80% of the cost is spent. This paper reports on the research done in developing the Ontology-based tool for Employer Information Requirements (OntEIR). OntEIR has proven the ability to produce a full and complete set of EIRs, which ensures that the clients’ information needs for the final model delivered by BIM is clearly defined from the beginning of the process. It also reports on the work being done into transforming OntEIR into a smart tool for Defining Employer Information Requirements (smartEIR). smartEIR transforms the OntEIR tool into enabling it to develop custom EIR- tailored for the: Project Type, Project Requirements, and the Client Capabilities. The initial idea behind smartEIR is moving away from the notion “One EIR fits All”. smartEIR utilizes the links made in OntEIR and creating a 3D matrix that transforms it into a smart tool. The OntEIR tool is based on the OntEIR framework that utilizes both Ontology and the Decomposition of Goals to elicit and extract the complete set of requirements needed for a full and comprehensive EIR. A new ctaegorisation system for requirements is also introduced in the framework and tool, which facilitates the understanding and enhances the clarification of the requirements especially for novice clients. Findings of the evaluation of the tool that was done with experts in the industry, showed that the OntEIR tool contributes towards effective and efficient development of EIRs that provide a better understanding of the information requirements as requested by BIM, and support the production of a complete BIM Execution Plan (BEP) and a Master Information Delivery Plan (MIDP).Keywords: building information modelling, employer information requirements, ontology, web-based, tool
Procedia PDF Downloads 1273844 Simulation of Carbon Nanotubes/GaAs Hybrid PV Using AMPS-1D
Authors: Nima E. Gorji
Abstract:
The performance and characteristics of a hybrid heterojunction single-walled carbon nanotube and GaAs solar cell is modelled and numerically simulated using AMPS-1D device simulation tool. The device physics and performance parameters with different junction parameters are analysed. The results suggest that the open-circuit voltage changes very slightly by changing the work function, acceptor and donor density while the other electrical parameters reach to an optimum value. Increasing the concentration of a discrete defect density in the absorber layer decreases the electrical parameters. The current-voltage characteristics, quantum efficiency, band gap and thickness variation of the photovoltaic response will be quantitatively considered.Keywords: carbon nanotube, GaAs, hybrid solar cell, AMPS-1D modelling
Procedia PDF Downloads 3303843 High-Efficiency Comparator for Low-Power Application
Authors: M. Yousefi, N. Nasirzadeh
Abstract:
In this paper, dynamic comparator structure employing two methods for power consumption reduction with applications in low-power high-speed analog-to-digital converters have been presented. The proposed comparator has low consumption thanks to power reduction methods. They have the ability for offset adjustment. The comparator consumes 14.3 μW at 100 MHz which is equal to 11.8 fJ. The comparator has been designed and simulated in 180 nm CMOS. Layouts occupy 210 μm2.Keywords: efficiency, comparator, power, low
Procedia PDF Downloads 3583842 New Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator
Authors: Wedad Albalawi
Abstract:
The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques, and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then, dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is an arbitrary nonempty closed subset of the real numbers. Then, the dynamic inequalities on time scales have received a lot of attention in the literature and has become a major field in pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on Hardy and Coposon inequalities, using Steklov operator on time scale in double integrals to obtain special cases of time-scale inequalities of Hardy and Copson on high dimensions. The advantage of this study is that it uses the one-dimensional classical Hardy inequality to obtain higher dimensional on time scale versions that will be applied in the solution of the Cauchy problem for the wave equation. In addition, the obtained inequalities have various applications involving discontinuous domains such as bug populations, phytoremediation of metals, wound healing, maximization problems. The proof can be done by introducing restriction on the operator in several cases. The concepts in time scale version such as time scales calculus will be used that allows to unify and extend many problems from the theories of differential and of difference equations. In addition, using chain rule, and some properties of multiple integrals on time scales, some theorems of Fubini and the inequality of H¨older.Keywords: time scales, inequality of hardy, inequality of coposon, steklov operator
Procedia PDF Downloads 953841 Transformer Life Enhancement Using Dynamic Switching of Second Harmonic Feature in IEDs
Authors: K. N. Dinesh Babu, P. K. Gargava
Abstract:
Energization of a transformer results in sudden flow of current which is an effect of core magnetization. This current will be dominated by the presence of second harmonic, which in turn is used to segregate fault and inrush current, thus guaranteeing proper operation of the relay. This additional security in the relay sometimes obstructs or delays differential protection in a specific scenario, when the 2nd harmonic content was present during a genuine fault. This kind of scenario can result in isolation of the transformer by Buchholz and pressure release valve (PRV) protection, which is acted when fault creates more damage in transformer. Such delays involve a huge impact on the insulation failure, and chances of repairing or rectifying fault of problem at site become very dismal. Sometimes this delay can cause fire in the transformer, and this situation becomes havoc for a sub-station. Such occurrences have been observed in field also when differential relay operation was delayed by 10-15 ms by second harmonic blocking in some specific conditions. These incidences have led to the need for an alternative solution to eradicate such unwarranted delay in operation in future. Modern numerical relay, called as intelligent electronic device (IED), is embedded with advanced protection features which permit higher flexibility and better provisions for tuning of protection logic and settings. Such flexibility in transformer protection IEDs, enables incorporation of alternative methods such as dynamic switching of second harmonic feature for blocking the differential protection with additional security. The analysis and precautionary measures carried out in this case, have been simulated and discussed in this paper to ensure that similar solutions can be adopted to inhibit analogous issues in future.Keywords: differential protection, intelligent electronic device (IED), 2nd harmonic inhibit, inrush inhibit
Procedia PDF Downloads 300