Search results for: spatial principal component analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30182

Search results for: spatial principal component analysis

26282 Ceramide-PLGA Nanoparticle Formation to Apply to Atopic Dermatitis

Authors: Sang-Myung Jung, Gwang Heum Yoon, Hoo Chul Lee, Hwa Sung Shin

Abstract:

Ceramide, a component of stratum corneum at epidermis, helps to construct a rigid and dense skin barrier to prevent pathogens that cause atopic dermatitis. However, ceramide was too hydrophobic to be directly absorbed into stratum corneum and has risks of side effects by excessive treatment. To overcome the obstacles, ceramide was embedded into PLGA nanoparticles coated with chitosan. PLGA and chitosan have been known as biocompatible materials. PLGA was squeezed when faced with water and pumped ceramide out of PLGA nanoparticle. In addition, the chitosan coating layer helped initial adherence of nanoparticles to skin and regulate ceramide release until removed. This coating was degraded at weakly acid state like skin surface, finally ceramide release could be controlled. Finally, the nanoparticle was demonstrated to be non-cytotoxic and regenerate stratum corneum of atopic dermatitis model. Overall the nanoparticle is suggested as a novel and effective nanodrug to apply atopic dermatitis.

Keywords: nanoparticle, controlled release, atopic dermatitis, chitosan coating, ceramide

Procedia PDF Downloads 385
26281 Float Glass Manufacture Facility Design: Feasibility Study in Kuwait

Authors: Farah Al-Mutairi, Hadeer Al-Jeeraan, Lima Ali, Raya Al-Dabbous, Sarah Baroun

Abstract:

Lately, within the middle east, development has taken place in the construction area which increased the demand of a crucial component, where without it; stunning views from skyscrapers cannot been experienced, and natural light would not be able to be viewed from an indoor building. Glass has changed the path of living and building. Float glass is a type of glass that is flat and it is the type used in the construction and automobile sector. Facility design on the other hand is a study that improves the efficiency utilization of people, equipment, material and space. Kuwait's governmental future developing plan bears in mind the need of increase in industries to increase the growth domestic product(GDP) of the country. This project studies the feasibility of two designs of a float glass manufacture in Kuwait. The first Alternative, consists of one production line of capacity 500 tons of glass per day. The second alternative, consists of three production lines, each of capacity 500 tons of glass daily.

Keywords: float glass manufacture, Kuwait, feasibility float glass, facility design, float glass production

Procedia PDF Downloads 376
26280 Optimal Designof Brush Roll for Semiconductor Wafer Using CFD Analysis

Authors: Byeong-Sam Kim, Kyoungwoo Park

Abstract:

This research analyzes structure of flat panel display (FPD) such as LCD as quantitative through CFD analysis and modeling change to minimize the badness rate and rate of production decrease by damage of large scale plater at wafer heating chamber at semi-conductor manufacturing process. This glass panel and wafer device with atmospheric pressure or chemical vapor deposition equipment for transporting and transferring wafers, robot hands carry these longer and wider wafers can also be easily handled. As a contact handling system composed of several problems in increased potential for fracture or warping. A non-contact handling system is required to solve this problem. The panel and wafer warping makes it difficult to carry out conventional contact to analysis. We propose a new non-contact transportation system with combining air suction and blowout. The numerical analysis and experimental is, therefore, should be performed to obtain compared to results achieved with non-contact solutions. This wafer panel noncontact handler shows its strength in maintaining high cleanliness levels for semiconductor production processes.

Keywords: flat panel display, non contact transportation, heat treatment process, CFD analysis

Procedia PDF Downloads 402
26279 Spatio-Temporal Dynamic of Woody Vegetation Assessment Using Oblique Landscape Photographs

Authors: V. V. Fomin, A. P. Mikhailovich, E. M. Agapitov, V. E. Rogachev, E. A. Kostousova, E. S. Perekhodova

Abstract:

Ground-level landscape photos can be used as a source of objective data on woody vegetation and vegetation dynamics. We proposed a method for processing, analyzing, and presenting ground photographs, which has the following advantages: 1) researcher has to form holistic representation of the study area in form of a set of interlapping ground-level landscape photographs; 2) it is necessary to define or obtain characteristics of the landscape, objects, and phenomena present on the photographs; 3) it is necessary to create new or supplement existing textual descriptions and annotations for the ground-level landscape photographs; 4) single or multiple ground-level landscape photographs can be used to develop specialized geoinformation layers, schematic maps or thematic maps; 5) it is necessary to determine quantitative data that describes both images as a whole, and displayed objects and phenomena, using algorithms for automated image analysis. It is suggested to match each photo with a polygonal geoinformation layer, which is a sector consisting of areas corresponding with parts of the landscape visible in the photos. Calculation of visibility areas is performed in a geoinformation system within a sector using a digital model of a study area relief and visibility analysis functions. Superposition of the visibility sectors corresponding with various camera viewpoints allows matching landscape photos with each other to create a complete and wholesome representation of the space in question. It is suggested to user-defined data or phenomenons on the images with the following superposition over the visibility sector in the form of map symbols. The technology of geoinformation layers’ spatial superposition over the visibility sector creates opportunities for image geotagging using quantitative data obtained from raster or vector layers within the sector with the ability to generate annotations in natural language. The proposed method has proven itself well for relatively open and clearly visible areas with well-defined relief, for example, in mountainous areas in the treeline ecotone. When the polygonal layers of visibility sectors for a large number of different points of photography are topologically superimposed, a layer of visibility of sections of the entire study area is formed, which is displayed in the photographs. Also, as a result of this overlapping of sectors, areas that did not appear in the photo will be assessed as gaps. According to the results of this procedure, it becomes possible to obtain information about the photos that display a specific area and from which points of photography it is visible. This information may be obtained either as a query on the map or as a query for the attribute table of the layer. The method was tested using repeated photos taken from forty camera viewpoints located on Ray-Iz mountain massif (Polar Urals, Russia) from 1960 until 2023. It has been successfully used in combination with other ground-based and remote sensing methods of studying the climate-driven dynamics of woody vegetation in the Polar Urals. Acknowledgment: This research was collaboratively funded by the Russian Ministry for Science and Education project No. FEUG-2023-0002 (image representation) and Russian Science Foundation project No. 24-24-00235 (automated textual description).

Keywords: woody, vegetation, repeated, photographs

Procedia PDF Downloads 39
26278 A Social Network Analysis of the Palestinian Feminist Network Tal3at

Authors: Maath M. Musleh

Abstract:

This research aims to study recent trends in the Palestinian feminist movement through the case study of Tal3at. The study uses social network analysis as its primary method to analyze Twitter data. It attempts to interpret results through the lens of network theories and Parson’s AGIL paradigm. The study reveals major structural weaknesses in the Tal3at network. Our findings suggest that the movement will decline soon as sentiments of alienation amongst Palestinian women increases. These findings were validated by a couple of central actors in the network. This study contributes an SNA approach to the understanding of the understudied Palestinian feminism.

Keywords: feminism, Palestine, social network analysis, Tal3at

Procedia PDF Downloads 249
26277 Vibration Propagation in Body-in-White Structures Through Structural Intensity Analysis

Authors: Jamal Takhchi

Abstract:

The understanding of vibration propagation in complex structures such as automotive body in white remains a challenging issue in car design regarding NVH performances. The current analysis is limited to the low frequency range where modal concepts are dominant. Higher frequencies, between 200 and 1000 Hz, will become critical With the rise of electrification. EVs annoying sounds are mostly whines created by either Gears or e-motors between 300 Hz and 2 kHz. Structural intensity analysis was Experienced a few years ago on finite element models. The application was promising but limited by the fact that the propagating 3D intensity vector field is masked by a rotational Intensity field. This rotational field should be filtered using a differential operator. The expression of this operator in the framework of finite element modeling is not yet known. The aim of the proposed work is to implement this operator in the current dynamic solver (NASTRAN) of Stellantis and develop the Expected methodology for the mid-frequency structural analysis of electrified vehicles.

Keywords: structural intensity, NVH, body in white, irrotatational intensity

Procedia PDF Downloads 141
26276 Vertical Urban Design Guideline and Its Application to Measure Human Cognition and Emotions

Authors: Hee Sun (Sunny) Choi, Gerhard Bruyns, Wang Zhang, Sky Cheng, Saijal Sharma

Abstract:

This research addresses the need for a comprehensive framework that can guide the design and assessment of multi-level public spaces and public realms and their impact on the built environment. The study aims to understand and measure the neural mechanisms involved in this process. By doing so, it can lay the foundation for vertical and volumetric urbanism and ensure consistency and excellence in the field while also supporting scientific research methods for urban design with cognitive neuroscientists. To investigate these aspects, the paper focuses on the neighborhood scale in Hong Kong, specifically examining multi-level public spaces and quasi-public spaces within both commercial and residential complexes. The researchers use predictive Artificial Intelligence (AI) as a methodology to assess and comprehend the applicability of the urban design framework for vertical and volumetric urbanism. The findings aim to identify the factors that contribute to successful public spaces within a vertical living environment, thus introducing a new typology of public spaces.

Keywords: vertical urbanism, scientific research methods, spatial cognition, urban design guideline

Procedia PDF Downloads 58
26275 Response Analysis of a Steel Reinforced Concrete High-Rise Building during the 2011 Tohoku Earthquake

Authors: Naohiro Nakamura, Takuya Kinoshita, Hiroshi Fukuyama

Abstract:

The 2011 off The Pacific Coast of Tohoku Earthquake caused considerable damage to wide areas of eastern Japan. A large number of earthquake observation records were obtained at various places. To design more earthquake-resistant buildings and improve earthquake disaster prevention, it is necessary to utilize these data to analyze and evaluate the behavior of a building during an earthquake. This paper presents an earthquake response simulation analysis (hereafter a seismic response analysis) that was conducted using data recorded during the main earthquake (hereafter the main shock) as well as the earthquakes before and after it. The data were obtained at a high-rise steel-reinforced concrete (SRC) building in the bay area of Tokyo. We first give an overview of the building, along with the characteristics of the earthquake motion and the building during the main shock. The data indicate that there was a change in the natural period before and after the earthquake. Next, we present the results of our seismic response analysis. First, the analysis model and conditions are shown, and then, the analysis result is compared with the observational records. Using the analysis result, we then study the effect of soil-structure interaction on the response of the building. By identifying the characteristics of the building during the earthquake (i.e., the 1st natural period and the 1st damping ratio) by the Auto-Regressive eXogenous (ARX) model, we compare the analysis result with the observational records so as to evaluate the accuracy of the response analysis. In this study, a lumped-mass system SR model was used to conduct a seismic response analysis using observational data as input waves. The main results of this study are as follows: 1) The observational records of the 3/11 main shock put it between a level 1 and level 2 earthquake. The result of the ground response analysis showed that the maximum shear strain in the ground was about 0.1% and that the possibility of liquefaction occurring was low. 2) During the 3/11 main shock, the observed wave showed that the eigenperiod of the building became longer; this behavior could be generally reproduced in the response analysis. This prolonged eigenperiod was due to the nonlinearity of the superstructure, and the effect of the nonlinearity of the ground seems to have been small. 3) As for the 4/11 aftershock, a continuous analysis in which the subject seismic wave was input after the 3/11 main shock was input was conducted. The analyzed values generally corresponded well with the observed values. This means that the effect of the nonlinearity of the main shock was retained by the building. It is important to consider this when conducting the response evaluation. 4) The first period and the damping ratio during a vibration were evaluated by an ARX model. Our results show that the response analysis model in this study is generally good at estimating a change in the response of the building during a vibration.

Keywords: ARX model, response analysis, SRC building, the 2011 off the Pacific Coast of Tohoku Earthquake

Procedia PDF Downloads 153
26274 Structuring of Multilayer Aluminum Nickel by Lift-off Process Using Cheap Negative Resist

Authors: Muhammad Talal Asghar

Abstract:

The lift-off technique of the photoresist for metal patterning in integrated circuit (IC) packaging has been widely utilized in the field of microelectromechanical systems and semiconductor component manufacturing. The main advantage lies in cost-saving, reduction in complexity, and maturity of the process. The selection of photoresist depends upon many factors such as cost, the thickness of the resist, comfortable and valuable parameters extraction. In the present study, an extremely cheap dry film photoresist E8015 of thickness 38-micrometer is processed for the first time for edge profiling, according to the author's best knowledge. Successful extraction of the helpful parameter range for resist processing is performed. An undercut angle of 66 to 73 degrees is realized by parameter variation like exposure energy and development time. Finally, 10-micrometer thick metallic multilayer aluminum nickel is lifted off on the plain silicon wafer. Possible applications lie in controlled self-propagating reactions within structured metallic multilayer that may be utilized for IC packaging in the future.

Keywords: lift-off, IC packaging, photoresist, multilayer

Procedia PDF Downloads 199
26273 Rescaled Range Analysis of Seismic Time-Series: Example of the Recent Seismic Crisis of Alhoceima

Authors: Marina Benito-Parejo, Raul Perez-Lopez, Miguel Herraiz, Carolina Guardiola-Albert, Cesar Martinez

Abstract:

Persistency, long-term memory and randomness are intrinsic properties of time-series of earthquakes. The Rescaled Range Analysis (RS-Analysis) was introduced by Hurst in 1956 and modified by Mandelbrot and Wallis in 1964. This method represents a simple and elegant analysis which determines the range of variation of one natural property (the seismic energy released in this case) in a time interval. Despite the simplicity, there is complexity inherent in the property measured. The cumulative curve of the energy released in time is the well-known fractal geometry of a devil’s staircase. This geometry is used for determining the maximum and minimum value of the range, which is normalized by the standard deviation. The rescaled range obtained obeys a power-law with the time, and the exponent is the Hurst value. Depending on this value, time-series can be classified in long-term or short-term memory. Hence, an algorithm has been developed for compiling the RS-Analysis for time series of earthquakes by days. Completeness time distribution and locally stationarity of the time series are required. The interest of this analysis is their application for a complex seismic crisis where different earthquakes take place in clusters in a short period. Therefore, the Hurst exponent has been obtained for the seismic crisis of Alhoceima (Mediterranean Sea) of January-March, 2016, where at least five medium-sized earthquakes were triggered. According to the values obtained from the Hurst exponent for each cluster, a different mechanical origin can be detected, corroborated by the focal mechanisms calculated by the official institutions. Therefore, this type of analysis not only allows an approach to a greater understanding of a seismic series but also makes possible to discern different types of seismic origins.

Keywords: Alhoceima crisis, earthquake time series, Hurst exponent, rescaled range analysis

Procedia PDF Downloads 305
26272 Finding the Optimal Meeting Point Based on Travel Plans in Road Networks

Authors: Mohammad H. Ahmadi, Vahid Haghighatdoost

Abstract:

Given a set of source locations for a group of friends, and a set of trip plans for each group member as a sequence of Categories-of-Interests (COIs) (e.g., restaurant), and finally a specific COI as a common destination that all group members will gather together, in Meeting Point Based on Trip Plans (MPTPs) queries our goal is to find a Point-of-Interest (POI) from different COIs, such that the aggregate travel distance for the group is minimized. In this work, we considered two cases for aggregate function as Sum and Max. For solving this query, we propose an efficient pruning technique for shrinking the search space. Our approach contains three steps. In the first step, it prunes the search space around the source locations. In the second step, it prunes the search space around the centroid of source locations. Finally, we compute the intersection of all pruned areas as the final refined search space. We prove that the POIs beyond the refined area cannot be part of optimal answer set. The paper also covers an extensive performance study of the proposed technique.

Keywords: meeting point, trip plans, road networks, spatial databases

Procedia PDF Downloads 169
26271 Context-Aware Point-Of-Interests Recommender Systems Using Integrated Sentiment and Network Analysis

Authors: Ho Yeon Park, Kyoung-Jae Kim

Abstract:

Recently, user’s interests for location-based social network service increases according to the advances of social web and location-based technologies. It may be easy to recommend preferred items if we can use user’s preference, context and social network information simultaneously. In this study, we propose context-aware POI (point-of-interests) recommender systems using location-based network analysis and sentiment analysis which consider context, social network information and implicit user’s preference score. We propose a context-aware POI recommendation system consisting of three sub-modules and an integrated recommendation system of them. First, we will develop a recommendation module based on network analysis. This module combines social network analysis and cluster-indexing collaboration filtering. Next, this study develops a recommendation module using social singular value decomposition (SVD) and implicit SVD. In this research, we will develop a recommendation module that can recommend preference scores based on the frequency of POI visits of user in POI recommendation process by using social and implicit SVD which can reflect implicit feedback in collaborative filtering. We also develop a recommendation module using them that can estimate preference scores based on the recommendation. Finally, this study will propose a recommendation module using opinion mining and emotional analysis using data such as reviews of POIs extracted from location-based social networks. Finally, we will develop an integration algorithm that combines the results of the three recommendation modules proposed in this research. Experimental results show the usefulness of the proposed model in relation to the recommended performance.

Keywords: sentiment analysis, network analysis, recommender systems, point-of-interests, business analytics

Procedia PDF Downloads 233
26270 Evaluation of Engineering Cementitious Composites (ECC) with Different Percentage of Fibers

Authors: Bhaumik Merchant, Ajay Gelot

Abstract:

Concrete is good in compression but if any type of strain applied to it, it starts to fail. Where the steel is good tension, it can bear the deflection up to its elastic limits. This project is based on behavior of engineered cementitious composited (ECC) when it is replaced with the different amount of Polyvinyl Alcohol (PVA) Fibers. As for research, PVA fibers is used with cementitious up to 2% to evaluate the optimum amount of fiber on which we can find the maximum compressive, tensile and flexural strength. PVA is basically an adhesive which is used to formulate glue. Generally due to excessive loading, cracks develops which concludes to successive damage to the structural component. In research plasticizer is used to increase workability. With the help of optimum amount of PVA fibers, it can limit the crack widths up to 60µm to 100µm. Also can be used to reduce resources and funds for rehabilitation of structure. At the starting this fiber concrete can be double the cost as compare to conventional concrete but as it can amplify the duration of structure, it will be less costlier than the conventional concrete.

Keywords: compressive strength, engineered cementitious composites, flexural strength, polyvinyl alcohol fibers, rehabilitation of structures

Procedia PDF Downloads 273
26269 An Engineer-Oriented Life Cycle Assessment Tool for Building Carbon Footprint: The Building Carbon Footprint Evaluation System in Taiwan

Authors: Hsien-Te Lin

Abstract:

The purpose of this paper is to introduce the BCFES (building carbon footprint evaluation system), which is a LCA (life cycle assessment) tool developed by the Low Carbon Building Alliance (LCBA) in Taiwan. A qualified BCFES for the building industry should fulfill the function of evaluating carbon footprint throughout all stages in the life cycle of building projects, including the production, transportation and manufacturing of materials, construction, daily energy usage, renovation and demolition. However, many existing BCFESs are too complicated and not very designer-friendly, creating obstacles in the implementation of carbon reduction policies. One of the greatest obstacle is the misapplication of the carbon footprint inventory standards of PAS2050 or ISO14067, which are designed for mass-produced goods rather than building projects. When these product-oriented rules are applied to building projects, one must compute a tremendous amount of data for raw materials and the transportation of construction equipment throughout the construction period based on purchasing lists and construction logs. This verification method is very cumbersome by nature and unhelpful to the promotion of low carbon design. With a view to provide an engineer-oriented BCFE with pre-diagnosis functions, a component input/output (I/O) database system and a scenario simulation method for building energy are proposed herein. Most existing BCFESs base their calculations on a product-oriented carbon database for raw materials like cement, steel, glass, and wood. However, data on raw materials is meaningless for the purpose of encouraging carbon reduction design without a feedback mechanism, because an engineering project is not designed based on raw materials but rather on building components, such as flooring, walls, roofs, ceilings, roads or cabinets. The LCBA Database has been composited from existing carbon footprint databases for raw materials and architectural graphic standards. Project designers can now use the LCBA Database to conduct low carbon design in a much more simple and efficient way. Daily energy usage throughout a building's life cycle, including air conditioning, lighting, and electric equipment, is very difficult for the building designer to predict. A good BCFES should provide a simplified and designer-friendly method to overcome this obstacle in predicting energy consumption. In this paper, the author has developed a simplified tool, the dynamic Energy Use Intensity (EUI) method, to accurately predict energy usage with simple multiplications and additions using EUI data and the designed efficiency levels for the building envelope, AC, lighting and electrical equipment. Remarkably simple to use, it can help designers pre-diagnose hotspots in building carbon footprint and further enhance low carbon designs. The BCFES-LCBA offers the advantages of an engineer-friendly component I/O database, simplified energy prediction methods, pre-diagnosis of carbon hotspots and sensitivity to good low carbon designs, making it an increasingly popular carbon management tool in Taiwan. To date, about thirty projects have been awarded BCFES-LCBA certification and the assessment has become mandatory in some cities.

Keywords: building carbon footprint, life cycle assessment, energy use intensity, building energy

Procedia PDF Downloads 128
26268 Starch-Based Systems for the Nano-Delivery of Quercetin

Authors: Fernando G. Torres, Omar P. Troncoso

Abstract:

Quercetin is a naturally occurring polyphenol found in many vegetables, such as onion, with antioxidant properties. It is a dietary component with a documented role in reducing different human cancers. However, its low bioavailability, poor water solubility, and chemical instability limit its applications. Different nano-delivery systems such as nanoparticles, micelles, and nanohydrogels have been studied in order to improve the bioavailability of quercetin. Nanoparticles based on natural polymers such as starch have the advantage of being biocompatible, biodegradable, and non-toxic. In this study, quercetin was loaded into starch nanoparticles using a nanoprecipitation method. Different routes, using sodium tripolyphosphate and Tween® 80 as tensioactive agents, were tested in order to obtain an optimized starch-based nano-delivery system. The characterization of the nanoparticles loaded with quercetin was assessed by Fourier Transform Infrared Spectroscopy, Dynamic Light Scattering, Zeta potential, and Differential scanning calorimetry. UV-vis spectrophotometry was used to evaluate the loading efficiency and capacity of the samples. The results showed that starch-based systems could be successfully used for the nano-delivery of quercetin.

Keywords: starch nanoparticles, nanoprecipitation, quercetin, biomedical applications

Procedia PDF Downloads 122
26267 Sensitivity Analysis during the Optimization Process Using Genetic Algorithms

Authors: M. A. Rubio, A. Urquia

Abstract:

Genetic algorithms (GA) are applied to the solution of high-dimensional optimization problems. Additionally, sensitivity analysis (SA) is usually carried out to determine the effect on optimal solutions of changes in parameter values of the objective function. These two analyses (i.e., optimization and sensitivity analysis) are computationally intensive when applied to high-dimensional functions. The approach presented in this paper consists in performing the SA during the GA execution, by statistically analyzing the data obtained of running the GA. The advantage is that in this case SA does not involve making additional evaluations of the objective function and, consequently, this proposed approach requires less computational effort than conducting optimization and SA in two consecutive steps.

Keywords: optimization, sensitivity, genetic algorithms, model calibration

Procedia PDF Downloads 422
26266 Digital Image Steganography with Multilayer Security

Authors: Amar Partap Singh Pharwaha, Balkrishan Jindal

Abstract:

In this paper, a new method is developed for hiding image in a digital image with multilayer security. In the proposed method, the secret image is encrypted in the first instance using a flexible matrix based symmetric key to add first layer of security. Then another layer of security is added to the secret data by encrypting the ciphered data using Pythagorean Theorem method. The ciphered data bits (4 bits) produced after double encryption are then embedded within digital image in the spatial domain using Least Significant Bits (LSBs) substitution. To improve the image quality of the stego-image, an improved form of pixel adjustment process is proposed. To evaluate the effectiveness of the proposed method, image quality metrics including Peak Signal-to-Noise Ratio (PSNR), Mean Square Error (MSE), entropy, correlation, mean value and Universal Image Quality Index (UIQI) are measured. It has been found experimentally that the proposed method provides higher security as well as robustness. In fact, the results of this study are quite promising.

Keywords: Pythagorean theorem, pixel adjustment, ciphered data, image hiding, least significant bit, flexible matrix

Procedia PDF Downloads 324
26265 Human Error Analysis in the USA Marine Accidents Reports

Authors: J. Sánchez-Beaskoetxea

Abstract:

The analysis of accidents, such as marine accidents, is one of the most useful instruments to avoid future accidents. In the case of marine accidents, from a simple collision of a small boat in a port to the wreck of a gigantic tanker ship, the study of the causes of the accidents is the basis of a great part of the marine international legislation. Some countries have official institutions who investigate all the accidents in which a ship with their flag is involved. In the case of the USA, the National Transportation Safety Board (NTSB) is responsible for these researches. The NTSB, after a deep investigation into each accident, publishes a Marine Accident Report with the possible cause of the accident. This paper analyses all the Marine Accident Reports published by the NTBS and focuses its attention especially in the Human Errors that led to reported accidents. In this research, the different Human Errors made by crew members are cataloged in 10 different groups. After a complete analysis of all the reports, the statistical analysis on the Human Errors typology in marine accidents is presented in order to use it as a tool to avoid the same errors in the future.

Keywords: human error, marine accidents, ship crew, USA

Procedia PDF Downloads 403
26264 Human-Induced Vibration and Degree of Human Comfortability Analysis of Intersection Pedestrian Bridge

Authors: Yaowen Sheng, Jiuxian Liu

Abstract:

In order to analyze the pedestrian bridge dynamic characteristics and degree of comfortability, the finite element method and live load time history method is used to calculate the dynamic response of the bridge. The example bridge’s dynamic characteristics and degree of human comfortability need to be analyzed. The project background is a three-way intersection. The intersection has three side blocks. An intersection bridge is designed to help people cross the streets. The finite element model of the bridge is established by the Midas/Civil software, and the analysis of the model is done. The strength, stiffness, and stability checks are also completed. Apart from the static analysis of the bridge, the dynamic analysis of the bridge is also completed to avoid the problems resulted from vibrations. The results show that the pedestrian bridge has different dynamic characteristics compared to other normal bridges. The degree of human comfortability satisfies the requirements of Chinese and British specifications. The live load time history method can be used to calculate the dynamic response of the bridge.

Keywords: pedestrian bridge, steel box girder, human-induced vibration, finite element analysis, degree of human comfortability

Procedia PDF Downloads 144
26263 Automatic Change Detection for High-Resolution Satellite Images of Urban and Suburban Areas

Authors: Antigoni Panagiotopoulou, Lemonia Ragia

Abstract:

High-resolution satellite images can provide detailed information about change detection on the earth. In the present work, QuickBird images of spatial resolution 60 cm/pixel and WorldView images of resolution 30 cm/pixel are utilized to perform automatic change detection in urban and suburban areas of Crete, Greece. There is a relative time difference of 13 years among the satellite images. Multiindex scene representation is applied on the images to classify the scene into buildings, vegetation, water and ground. Then, automatic change detection is made possible by pixel-per-pixel comparison of the classified multi-temporal images. The vegetation index and the water index which have been developed in this study prove effective. Furthermore, the proposed change detection approach not only indicates whether changes have taken place or not but also provides specific information relative to the types of changes. Experimentations with other different scenes in the future could help optimize the proposed spectral indices as well as the entire change detection methodology.

Keywords: change detection, multiindex scene representation, spectral index, QuickBird, WorldView

Procedia PDF Downloads 122
26262 Axle Load Estimation of Moving Vehicles Using BWIM Technique

Authors: Changgil Lee, Seunghee Park

Abstract:

Although vehicle driving test for the development of BWIM system is necessary, but it needs much cost and time in addition application of various driving condition. Thus, we need the numerical-simulation method resolving the cost and time problems of vehicle driving test and the way of measuring response of bridge according to the various driving condition. Using the precision analysis model reflecting the dynamic characteristic is contributed to increase accuracy in numerical simulation. In this paper, we conduct a numerical simulation to apply precision analysis model, which reflects the dynamic characteristic of bridge using Bridge Weigh-in-Motion technique and suggest overload vehicle enforcement technology using precision analysis model.

Keywords: bridge weigh-in-motion(BWIM) system, precision analysis model, dynamic characteristic of bridge, numerical simulation

Procedia PDF Downloads 275
26261 A Novel Meta-Heuristic Algorithm Based on Cloud Theory for Redundancy Allocation Problem under Realistic Condition

Authors: H. Mousavi, M. Sharifi, H. Pourvaziri

Abstract:

Redundancy Allocation Problem (RAP) is a well-known mathematical problem for modeling series-parallel systems. It is a combinatorial optimization problem which focuses on determining an optimal assignment of components in a system design. In this paper, to be more practical, we have considered the problem of redundancy allocation of series system with interval valued reliability of components. Therefore, during the search process, the reliabilities of the components are considered as a stochastic variable with a lower and upper bounds. In order to optimize the problem, we proposed a simulated annealing based on cloud theory (CBSAA). Also, the Monte Carlo simulation (MCS) is embedded to the CBSAA to handle the random variable components’ reliability. This novel approach has been investigated by numerical examples and the experimental results have shown that the CBSAA combining MCS is an efficient tool to solve the RAP of systems with interval-valued component reliabilities.

Keywords: redundancy allocation problem, simulated annealing, cloud theory, monte carlo simulation

Procedia PDF Downloads 401
26260 Applicability of Linearized Model of Synchronous Generator for Power System Stability Analysis

Authors: J. Ritonja, B. Grcar

Abstract:

For the synchronous generator simulation and analysis and for the power system stabilizer design and synthesis a mathematical model of synchronous generator is needed. The model has to accurately describe dynamics of oscillations, while at the same time has to be transparent enough for an analysis and sufficiently simplified for design of control system. To study the oscillations of the synchronous generator against to the rest of the power system, the model of the synchronous machine connected to an infinite bus through a transmission line having resistance and inductance is needed. In this paper, the linearized reduced order dynamic model of the synchronous generator connected to the infinite bus is presented and analysed in details. This model accurately describes dynamics of the synchronous generator only in a small vicinity of an equilibrium state. With the digression from the selected equilibrium point the accuracy of this model is decreasing considerably. In this paper, the equations’ descriptions and the parameters’ determinations for the linearized reduced order mathematical model of the synchronous generator are explained and summarized and represent the useful origin for works in the areas of synchronous generators’ dynamic behaviour analysis and synchronous generator’s control systems design and synthesis. The main contribution of this paper represents the detailed analysis of the accuracy of the linearized reduced order dynamic model in the entire synchronous generator’s operating range. Borders of the areas where the linearized reduced order mathematical model represents accurate description of the synchronous generator’s dynamics are determined with the systemic numerical analysis. The thorough eigenvalue analysis of the linearized models in the entire operating range is performed. In the paper, the parameters of the linearized reduced order dynamic model of the laboratory salient poles synchronous generator were determined and used for the analysis. The theoretical conclusions were confirmed with the agreement of experimental and simulation results.

Keywords: eigenvalue analysis, mathematical model, power system stability, synchronous generator

Procedia PDF Downloads 231
26259 NextCovps: Design and Stress Analysis of Dome Composite Overwrapped Pressure Vessels using Geodesic Trajectory Approach

Authors: Ammar Maziz, Prateek Gupta, Thiago Vasconcellos Birro, Benoit Gely

Abstract:

Hydrogen as a sustainable fuel has the highest energy density per mass as compared to conventional non-renewable sources. As the world looks to move towards sustainability, especially in the sectors of aviation and automotive, it becomes important to address the issue of storage of hydrogen as compressed gas in high-pressure tanks. To improve the design for the efficient storage and transportation of Hydrogen, this paper presents the design and stress analysis of Dome Composite Overwrapped Pressure Vessels (COPVs) using the geodesic trajectory approach. The geodesic trajectory approach is used to optimize the dome design, resulting in a lightweight and efficient structure. Python scripting is employed to implement the mathematical modeling of the COPV, and after validating the model by comparison to the published paper, stress analysis is conducted using Abaqus commercial code. The results demonstrate the effectiveness of the geodesic trajectory approach in achieving a lightweight and structurally sound dome design, as well as the accuracy and reliability of the stress analysis using Abaqus commercial code. This study provides insights into the design and analysis of COPVs for aerospace applications, with the potential for further optimization and application in other industries.

Keywords: composite overwrapped pressure vessels, carbon fiber, geodesic trajectory approach, dome design, stress analysis, plugin python

Procedia PDF Downloads 75
26258 Failure Inference and Optimization for Step Stress Model Based on Bivariate Wiener Model

Authors: Soudabeh Shemehsavar

Abstract:

In this paper, we consider the situation under a life test, in which the failure time of the test units are not related deterministically to an observable stochastic time varying covariate. In such a case, the joint distribution of failure time and a marker value would be useful for modeling the step stress life test. The problem of accelerating such an experiment is considered as the main aim of this paper. We present a step stress accelerated model based on a bivariate Wiener process with one component as the latent (unobservable) degradation process, which determines the failure times and the other as a marker process, the degradation values of which are recorded at times of failure. Parametric inference based on the proposed model is discussed and the optimization procedure for obtaining the optimal time for changing the stress level is presented. The optimization criterion is to minimize the approximate variance of the maximum likelihood estimator of a percentile of the products’ lifetime distribution.

Keywords: bivariate normal, Fisher information matrix, inverse Gaussian distribution, Wiener process

Procedia PDF Downloads 309
26257 Exploring the Role of Phosphorylation on the β-lactamase Activity of OXA24/40

Authors: Dharshika Rajalingam, Jeffery W. Peng

Abstract:

Acinetobacter baumannii is a challenging threat to global health, recognized as a multidrug-resistant pathogen. -lactamase is one of the principal resistant mechanisms developed by A. baumannii to survive against -lactam antibiotics. OXA24/40 is one of the types of -lactamases, a well-documented carbapenem hydrolyzing class D -lactamases (CHDL). It was revealed that OXA24/40 showed resistivity against doripenem, one of the carbapenems, by two different mechanisms as hydrolysis and -lactonization. Furthermore, it undergoes genetic mutations to broaden the -lactamase activity to survive against antibiotic environments. One of the crucial characterizations of prokaryotes to develop adaptation is post-translational modification (PTM), mainly phosphorylation. However, the PTM of OXA24/40 is an unknown feature, and the impact of PTM on antibiotic resistivity is yet to be explored. We approached these hypotheses using NMR and MS techniques and found that the OXA24/40 could be phosphorylated in vitro. The Ser81 at the active STFK motif of OXA24/40 of catalytic pocket was identified as the site of phosphorylation using 1D 31P NMR experiment, whereas S81 is required to form an acyl-enzyme complex between enzyme and -lactam antibiotics. The activity of completely phosphorylated OXA24/40 wild type against doripenem revealed that the phosphorylation of active Ser inactivates the -lactamases activity of OXA24/40. The 1D 1H CPMG NMR-based activity assay of phosphorylated OXA24/40 against doripenem confirmed that both deactivating mechanisms are inhibited by phosphorylation. Carbamylated Lysine at the active STFK motif is one of the critical features of CHDL required for the acylation and deacylation reactions of the enzyme. The 1D 13C NMR experiment confirmed that the K84 of phosphorylated OXA24/40 is de-carbamylated. Phosphorylation of OXA24/40 affects both active S81 and carbamylated K84 of OXA24 that are required for the resistivity of -lactamase. So, phosphorylation could be one of the reasons for the genetic mutation of OXA24/40 for the development of antibiotic resistivity. Further research can lead to an understanding of the effect of phosphorylation on the clinical mutants of the OXA24-like -lactamase family on the broadening of -lactamase activity.

Keywords: OXA24/40, phosphorylation, clinical mutants, resistivity

Procedia PDF Downloads 64
26256 A Geographical Framework for Studying the Territorial Sustainability Based on Land Use Change

Authors: Miguel Ramirez, Ivan Lizarazo

Abstract:

The emergence of various interpretations of sustainability, including weak and strong paradigms, can be traced back to the definition of sustainable development provided in the 1987 Brundtland report and the subsequent evolution of the sustainability concept. However, there has been limited scholarly attention given to clarifying the concept of sustainability within the theoretical and conceptual framework of geography. The discipline has predominantly been focused on understanding the diverse conceptions of sustainability within its epistemological boundaries, resulting in tensions between sustainability paradigms and their associated dimensions, including the incorporation of political perspectives, with particular emphasis on environmental geography's epistemology. In response to this gap, a conceptual framework for sustainability is proposed, effectively integrating spatial and territorial concepts. This framework aims to enhance geography's role in contributing to sustainability by utilizing the land system theory, which is based on the dynamics of land use change. Such an integrated conceptual framework enables incorporating methodological tools such as remote sensing, encompassing various earth observations and fusion methods, and supervised classification techniques. Additionally, it looks for better integration of socioecological information, thereby capturing essential population-related features.

Keywords: geography, sustainability, land change science, territorial sustainability

Procedia PDF Downloads 58
26255 Discriminant Analysis as a Function of Predictive Learning to Select Evolutionary Algorithms in Intelligent Transportation System

Authors: Jorge A. Ruiz-Vanoye, Ocotlán Díaz-Parra, Alejandro Fuentes-Penna, Daniel Vélez-Díaz, Edith Olaco García

Abstract:

In this paper, we present the use of the discriminant analysis to select evolutionary algorithms that better solve instances of the vehicle routing problem with time windows. We use indicators as independent variables to obtain the classification criteria, and the best algorithm from the generic genetic algorithm (GA), random search (RS), steady-state genetic algorithm (SSGA), and sexual genetic algorithm (SXGA) as the dependent variable for the classification. The discriminant classification was trained with classic instances of the vehicle routing problem with time windows obtained from the Solomon benchmark. We obtained a classification of the discriminant analysis of 66.7%.

Keywords: Intelligent Transportation Systems, data-mining techniques, evolutionary algorithms, discriminant analysis, machine learning

Procedia PDF Downloads 451
26254 Evaluation of NoSQL in the Energy Marketplace with GraphQL Optimization

Authors: Michael Howard

Abstract:

The growing popularity of electric vehicles in the United States requires an ever-expanding infrastructure of commercial DC fast charging stations. The U.S. Department of Energy estimates 33,355 publicly available DC fast charging stations as of September 2023. In 2017, 115,370 gasoline stations were operating in the United States, much more ubiquitous than DC fast chargers. Range anxiety is an important impediment to the adoption of electric vehicles and is even more relevant in underserved regions in the country. The peer-to-peer energy marketplace helps fill the demand by allowing private home and small business owners to rent their 240 Volt, level-2 charging facilities. The existing, publicly accessible outlets are wrapped with a Cloud-connected microcontroller managing security and charging sessions. These microcontrollers act as Edge devices communicating with a Cloud message broker, while both buyer and seller users interact with the framework via a web-based user interface. The database storage used by the marketplace framework is a key component in both the cost of development and the performance that contributes to the user experience. A traditional storage solution is the SQL database. The architecture and query language have been in existence since the 1970s and are well understood and documented. The Structured Query Language supported by the query engine provides fine granularity with user query conditions. However, difficulty in scaling across multiple nodes and cost of its server-based compute have resulted in a trend in the last 20 years towards other NoSQL, serverless approaches. In this study, we evaluate the NoSQL vs. SQL solutions through a comparison of Google Cloud Firestore and Cloud SQL MySQL offerings. The comparison pits Google's serverless, document-model, non-relational, NoSQL against the server-base, table-model, relational, SQL service. The evaluation is based on query latency, flexibility/scalability, and cost criteria. Through benchmarking and analysis of the architecture, we determine whether Firestore can support the energy marketplace storage needs and if the introduction of a GraphQL middleware layer can overcome its deficiencies.

Keywords: non-relational, relational, MySQL, mitigate, Firestore, SQL, NoSQL, serverless, database, GraphQL

Procedia PDF Downloads 38
26253 Managing Construction and Demolition Wastes - A Case Study of Multi Triagem, Lda

Authors: Cláudia Moço, Maria Santos, Carlos Arsénio, Débora Mendes, Miguel Oliveira. José Paulo Da Silva

Abstract:

Construction industry generates large amounts of waste all over the world. About 450 million tons of construction and demolition wastes (C&DW) are produced annually in the European Union. C&DW are highly heterogeneous materials in size and composition, which imposes strong difficulties on their management. Directive n.º 2008/98/CE, of the European Parliament and of the Council of 6 November establishes that 70 % of the C&DW have to be recycled by 2020. To evaluate possible applications of these materials, a detailed physical, chemical and environmental characterization is necessary. Multi Triagem, Lda. is a company located in Algarve (Portugal) and was supported by the European Regional Development Fund (grant QREN 30307 Multivalor) to quantify and characterize the received C&DW, in order to evaluate their possible applications. This evaluation, performed in collaboration with the University of Algarve, involves a physical, chemical and environmental detailed characterization of the received C&DW. In this work we report on the amounts, trial procedures and properties of the C&DW received over a period of fifteen month. In this period the company received C&DW coming from 393 different origins. The total amount was 32.458 tons, mostly mixtures containing concrete, masonry/mortar and soil/rock. Most of C&DW came from demodulation constructions and diggings. The organic/inert component, namely metal, glass, wood and plastics, were screened first and account for about 3 % of the received materials. The remaining materials were screened and grouped according to their origin and contents, the latter evaluated by visual inspection. Twenty five samples were prepared and submitted to a detailed physical, chemical and environmental analysis. The C&DW aggregates show lower quality properties than natural aggregates for concrete preparation and unbound layers of road pavements. However, chemical analyzes indicated that most samples are environmentally safe. A continuous monitoring of the presence of heavy metals and organic compounds is needed in order to perform a proper screening of the C&DW. C&DW aggregates provide a good alternative to natural aggregates.

Keywords: construction and demolition wastes, waste classification, waste composition, waste screening

Procedia PDF Downloads 335