Search results for: sound propagation models
5466 A Comparative Study on the Dimensional Error of 3D CAD Model and SLS RP Model for Reconstruction of Cranial Defect
Authors: L. Siva Rama Krishna, Sriram Venkatesh, M. Sastish Kumar, M. Uma Maheswara Chary
Abstract:
Rapid Prototyping (RP) is a technology that produces models and prototype parts from 3D CAD model data, CT/MRI scan data, and model data created from 3D object digitizing systems. There are several RP process like Stereolithography (SLA), Solid Ground Curing (SGC), Selective Laser Sintering (SLS), Fused Deposition Modelling (FDM), 3D Printing (3DP) among them SLS and FDM RP processes are used to fabricate pattern of custom cranial implant. RP technology is useful in engineering and biomedical application. This is helpful in engineering for product design, tooling and manufacture etc. RP biomedical applications are design and development of medical devices, instruments, prosthetics and implantation; it is also helpful in planning complex surgical operation. The traditional approach limits the full appreciation of various bony structure movements and therefore the custom implants produced are difficult to measure the anatomy of parts and analyse the changes in facial appearances accurately. Cranioplasty surgery is a surgical correction of a defect in cranial bone by implanting a metal or plastic replacement to restore the missing part. This paper aims to do a comparative study on the dimensional error of CAD and SLS RP Models for reconstruction of cranial defect by comparing the virtual CAD with the physical RP model of a cranial defect.Keywords: rapid prototyping, selective laser sintering, cranial defect, dimensional error
Procedia PDF Downloads 3255465 Effect of Wind Braces to Earthquake Resistance of Steel Structures
Authors: H. Gokdemir
Abstract:
All structures are subject to vertical and lateral loads. Under these loads, structures make deformations and deformation values of structural elements mustn't exceed their capacity for structural stability. Especially, lateral loads cause critical deformations because of their random directions and magnitudes. Wind load is one of the lateral loads which can act in any direction and any magnitude. Although wind has nearly no effect on reinforced concrete structures, it must be considered for steel structures, roof systems and slender structures like minarets. Therefore, every structure must be able to resist wind loads acting parallel and perpendicular to any side. One of the effective methods for resisting lateral loads is assembling cross steel elements between columns which are called as wind bracing. These cross elements increases lateral rigidity of a structure and prevent exceeding of deformation capacity of the structural system. So, this means cross elements are also effective in resisting earthquake loads too. In this paper; Effects of wind bracing to earthquake resistance of structures are studied. Structure models (with and without wind bracing) are generated and these models are solved under both earthquake and wind loads with different seismic zone parameters. It is concluded by the calculations that; in low-seismic risk zones, wind bracing can easily resist earthquake loads and no additional reinforcement for earthquake loads is necessary. Similarly; in high-seismic risk zones, earthquake cross elements resist wind loads too.Keywords: wind bracings, earthquake, steel structures, vertical and lateral loads
Procedia PDF Downloads 4705464 Estimating Anthropometric Dimensions for Saudi Males Using Artificial Neural Networks
Authors: Waleed Basuliman
Abstract:
Anthropometric dimensions are considered one of the important factors when designing human-machine systems. In this study, the estimation of anthropometric dimensions has been improved by using Artificial Neural Network (ANN) model that is able to predict the anthropometric measurements of Saudi males in Riyadh City. A total of 1427 Saudi males aged 6 to 60 years participated in measuring 20 anthropometric dimensions. These anthropometric measurements are considered important for designing the work and life applications in Saudi Arabia. The data were collected during eight months from different locations in Riyadh City. Five of these dimensions were used as predictors variables (inputs) of the model, and the remaining 15 dimensions were set to be the measured variables (Model’s outcomes). The hidden layers varied during the structuring stage, and the best performance was achieved with the network structure 6-25-15. The results showed that the developed Neural Network model was able to estimate the body dimensions of Saudi male population in Riyadh City. The network's mean absolute percentage error (MAPE) and the root mean squared error (RMSE) were found to be 0.0348 and 3.225, respectively. These results were found less, and then better, than the errors found in the literature. Finally, the accuracy of the developed neural network was evaluated by comparing the predicted outcomes with regression model. The ANN model showed higher coefficient of determination (R2) between the predicted and actual dimensions than the regression model.Keywords: artificial neural network, anthropometric measurements, back-propagation
Procedia PDF Downloads 4875463 Assessing Level of Pregnancy Rate and Milk Yield in Indian Murrah Buffaloes
Authors: V. Jamuna, A. K. Chakravarty, C. S. Patil, Vijay Kumar, M. A. Mir, Rakesh Kumar
Abstract:
Intense selection of buffaloes for milk production at organized herds of the country without giving due attention to fertility traits viz. pregnancy rate has lead to deterioration in their performances. Aim of study is to develop an optimum model for predicting pregnancy rate and to assess the level of pregnancy rate with respect to milk production Murrah buffaloes. Data pertaining to 1224 lactation records of Murrah buffaloes spread over a period 21 years were analyzed and it was observed that pregnancy rate depicted negative phenotypic association with lactation milk yield (-0.08 ± 0.04). For developing optimum model for pregnancy rate in Murrah buffaloes seven simple and multiple regression models were developed. Among the seven models, model II having only Service period as an independent reproduction variable, was found to be the best prediction model, based on the four statistical criterions (high coefficient of determination (R 2), low mean sum of squares due to error (MSSe), conceptual predictive (CP) value, and Bayesian information criterion (BIC). For standardizing the level of fertility with milk production, pregnancy rate was classified into seven classes with the increment of 10% in all parities, life time and their corresponding average pregnancy rate in relation to the average lactation milk yield (MY).It was observed that to achieve around 2000 kg MY which can be considered optimum for Indian Murrah buffaloes, level of pregnancy rate should be in between 30-50%.Keywords: life time, pregnancy rate, production, service period, standardization
Procedia PDF Downloads 6355462 Mobility Management for Pedestrian Accident Predictability and Mitigation Strategies Using Multiple
Authors: Oscar Norman Nekesa, Yoshitaka Kajita
Abstract:
Tom Mboya Street is a vital urban corridor within the spectrum of Nairobi city, it experiences high volumes of pedestrian and vehicular traffic. Despite past intervention measures to lessen this catastrophe, rates have remained high. This highlights significant safety concerns that need urgent attention. This study investigates the correlation and pedestrian accident predictability with significant independent variables using multiple linear regression to model to develop effective mobility management strategies for accident mitigation. The methodology involves collecting and analyzing data on pedestrian accidents and various related independent variables. Data sources include the National Transport and Safety Authority (NTSA), Kenya National Bureau of Statistics, and Nairobi City County records, covering five years. This study aims to investigate that traffic volumes (pedestrian and vehicle), Vehicular speed, human factors, illegal parking, policy issues, urban-land use, built environment, traffic signals conditions, inadequate lighting, and insufficient traffic control measures significantly have predictability with the rate of pedestrian accidents. Explanatory variables related to road design and geometry are significant in predictor models for the Tom Mboya Road link but less influential in junction along the 5 km stretch road models. The most impactful variable across all models was vehicular traffic flow. The study recommends infrastructural improvements, enhanced enforcement, and public awareness campaigns to reduce accidents and improve urban mobility. These insights can inform policy-making and urban planning to enhance pedestrian safety along the dense packed Tom Mboya Street and similar urban settings. The findings will inform evidence-based interventions to enhance pedestrian safety and improve urban mobility.Keywords: multiple linear regression, urban mobility, traffic management, Nairobi, Tom Mboya street, infrastructure conditions., pedestrian safety, correlation and prediction
Procedia PDF Downloads 255461 E-government Status and Impact on Development in the Arab Region
Authors: Sukaina Al-Nasrawi, Maysoun Ibrahim
Abstract:
Information and communication technologies (ICT) have affected recent public administration and governance. Electronic Government (e-government) services were developed to simplify government procedures and improve interaction with citizens on one hand and to create new governance models to empower citizens and involve them in the decision-making process while increasing transparency on another hand. It is worth noting that efficient governance models enable sustainable development at the social and economic levels. Currently, the status of e-government national strategies and implementation programs vary from one country to another. This variance in the development levels of e-government initiatives and applications noted the digital divide between countries of the same region, thereby highlighting the difficulty to reach regional integration. Many Arab countries realized the need for a well-articulated e-government strategy and launched national e-government initiatives. In selected Arab countries, the focus of e-government initiatives and programs shifted from the provision of services to advanced concepts such as open data initiatives. This paper aims at over viewing the e-government achievements of Arab countries and areas for enhancement, and share best practices in the area.of the best e-government programmes from the Arab region the world. It will also shed the light on the impact of the information society in general and e-government, in specific, on the social and economic development in the Arab region.Keywords: Information and Communication Technologies (ICT), services, e-government, development, Arab region, digital divide, citizens
Procedia PDF Downloads 2915460 Comparison of Hydrogen and Electrification Perspectives in Decarbonizing the Transport Sector
Authors: Matteo Nicoli, Gianvito Colucci, Valeria Di Cosmo, Daniele Lerede, Laura Savoldi
Abstract:
The transport sector is currently responsible for approximately 1/3 of greenhouse gas emissions in Europe. In the wider context of achieving carbon neutrality of the global energy system, different alternatives are available to decarbonizethe transport sector. In particular, while electricity is already the most consumed energy commodity in rail transport, battery electric vehicles are one of the zero-emissions options on the market for road transportation. On the other hand, hydrogen-based fuel cell vehicles are available for road and non-road vehicles. The European Commission is strongly pushing toward the integration of hydrogen in the energy systems of European countries and its widespread adoption as an energy vector to achieve the Green Deal targets. Furthermore, the Italian government is defining hydrogen-related objectives with the publication of a dedicated Hydrogen Strategy. The adoption of energy system optimization models to study the possible penetration of alternative zero-emitting transport technologies gives the opportunity to perform an overall analysis of the effects that the development of innovative technologies has on the entire energy system and on the supply-side, devoted to the production of energy carriers such as hydrogen and electricity. Using an open-source modeling framework such as TEMOA, this work aims to compare the role of hydrogen and electric vehicles in the decarbonization of the transport sector. The analysis investigates the advantages and disadvantages of adopting the two options, from the economic point of view (costs associated with the two options) and the environmental one (looking at the emissions reduction perspectives). Moreover, an analysis on the profitability of the investments in hydrogen and electric vehicles will be performed. The study investigates the evolution of energy consumption and greenhouse gas emissions in different transportation modes (road, rail, navigation, and aviation) by detailed analysis of the full range of vehicles included in the techno-economic database used in the TEMOA model instance adopted for this work. The transparency of the analysis is guaranteed by the accessibility of the TEMOA models, based on an open-access source code and databases.Keywords: battery electric vehicles, decarbonization, energy system optimization models, fuel cell vehicles, hydrogen, open-source modeling, TEMOA, transport
Procedia PDF Downloads 1125459 Time Series Modelling for Forecasting Wheat Production and Consumption of South Africa in Time of War
Authors: Yiseyon Hosu, Joseph Akande
Abstract:
Wheat is one of the most important staple food grains of human for centuries and is largely consumed in South Africa. It has a special place in the South African economy because of its significance in food security, trade, and industry. This paper modelled and forecast the production and consumption of wheat in South Africa in the time covid-19 and the ongoing Russia-Ukraine war by using annual time series data from 1940–2021 based on the ARIMA models. Both the averaging forecast and selected models forecast indicate that there is the possibility of an increase with respect to production. The minimum and maximum growth in production is projected to be between 3million and 10 million tons, respectively. However, the model also forecast a possibility of depression with respect to consumption in South Africa. Although Covid-19 and the war between Ukraine and Russia, two major producers and exporters of global wheat, are having an effect on the volatility of the prices currently, the wheat production in South African is expected to increase and meat the consumption demand and provided an opportunity for increase export with respect to domestic consumption. The forecasting of production and consumption behaviours of major crops play an important role towards food and nutrition security, these findings can assist policymakers and will provide them with insights into the production and pricing policy of wheat in South Africa.Keywords: ARIMA, food security, price volatility, staple food, South Africa
Procedia PDF Downloads 1025458 Problems in Computational Phylogenetics: The Germano-Italo-Celtic Clade
Authors: Laura Mclean
Abstract:
A recurring point of interest in computational phylogenetic analysis of Indo-European family trees is the inference of a Germano-Italo-Celtic clade in some versions of the trees produced. The presence of this clade in the models is intriguing as there is little evidence for innovations shared among Germanic, Italic, and Celtic, the evidence generally used in the traditional method to construct a subgroup. One source of this unexpected outcome could be the input to the models. The datasets in the various models used so far, for the most part, take as their basis the Swadesh list, a list compiled by Morris Swadesh and then revised several times, containing up to 207 words that he believed were resistant to change among languages. The judgments made by Swadesh for this list, however, were subjective and based on his intuition rather than rigorous analysis. Some scholars used the Swadesh 200 list as the basis for their Indo-European dataset and made cognacy judgements for each of the words on the list. Another dataset is largely based on the Swadesh 207 list as well although the authors include additional lexical and non-lexical data, and they implement ‘split coding’ to deal with cases of polymorphic characters. A different team of scholars uses a different dataset, IECoR, which combines several different lists, one of which is the Swadesh 200 list. In fact, the Swadesh list is used in some form in every study surveyed and each dataset has three words that, when they are coded as cognates, seemingly contribute to the inference of a Germano-Italo-Celtic clade which could happen due to these clades sharing three words among only themselves. These three words are ‘fish’, ‘flower’, and ‘man’ (in the case of ‘man’, one dataset includes Lithuanian in the cognacy coding and removes the word ‘man’ from the screened data). This collection of cognates shared among Germanic, Italic, and Celtic that were deemed important enough to be included on the Swadesh list, without the ability to account for possible reasons for shared cognates that are not shared innovations, gives an impression of affinity between the Germanic, Celtic, and Italic branches without adequate methodological support. However, by changing how cognacy is defined (ie. root cognates, borrowings vs inherited cognates etc.), we will be able to identify whether these three cognates are significant enough to infer a clade for Germanic, Celtic, and Italic. This paper examines the question of what definition of cognacy should be used for phylogenetic datasets by examining the Germano-Italo-Celtic clade as a case study and offers insights into the reconstruction of a Germano-Italo-Celtic clade.Keywords: historical, computational, Italo-Celtic, Germanic
Procedia PDF Downloads 505457 Modeling Aeration of Sharp Crested Weirs by Using Support Vector Machines
Authors: Arun Goel
Abstract:
The present paper attempts to investigate the prediction of air entrainment rate and aeration efficiency of a free over-fall jets issuing from a triangular sharp crested weir by using regression based modelling. The empirical equations, support vector machine (polynomial and radial basis function) models and the linear regression techniques were applied on the triangular sharp crested weirs relating the air entrainment rate and the aeration efficiency to the input parameters namely drop height, discharge, and vertex angle. It was observed that there exists a good agreement between the measured values and the values obtained using empirical equations, support vector machine (Polynomial and rbf) models, and the linear regression techniques. The test results demonstrated that the SVM based (Poly & rbf) model also provided acceptable prediction of the measured values with reasonable accuracy along with empirical equations and linear regression techniques in modelling the air entrainment rate and the aeration efficiency of a free over-fall jets issuing from triangular sharp crested weir. Further sensitivity analysis has also been performed to study the impact of input parameter on the output in terms of air entrainment rate and aeration efficiency.Keywords: air entrainment rate, dissolved oxygen, weir, SVM, regression
Procedia PDF Downloads 4365456 The Road Ahead: Merging Human Cyber Security Expertise with Generative AI
Authors: Brennan Lodge
Abstract:
Amidst a complex regulatory landscape, Retrieval Augmented Generation (RAG) emerges as a transformative tool for Governance Risk and Compliance (GRC) officers. This paper details the application of RAG in synthesizing Large Language Models (LLMs) with external knowledge bases, offering GRC professionals an advanced means to adapt to rapid changes in compliance requirements. While the development for standalone LLM’s (Large Language Models) is exciting, such models do have their downsides. LLM’s cannot easily expand or revise their memory, and they can’t straightforwardly provide insight into their predictions, and may produce “hallucinations.” Leveraging a pre-trained seq2seq transformer and a dense vector index of domain-specific data, this approach integrates real-time data retrieval into the generative process, enabling gap analysis and the dynamic generation of compliance and risk management content. We delve into the mechanics of RAG, focusing on its dual structure that pairs parametric knowledge contained within the transformer model with non-parametric data extracted from an updatable corpus. This hybrid model enhances decision-making through context-rich insights, drawing from the most current and relevant information, thereby enabling GRC officers to maintain a proactive compliance stance. Our methodology aligns with the latest advances in neural network fine-tuning, providing a granular, token-level application of retrieved information to inform and generate compliance narratives. By employing RAG, we exhibit a scalable solution that can adapt to novel regulatory challenges and cybersecurity threats, offering GRC officers a robust, predictive tool that augments their expertise. The granular application of RAG’s dual structure not only improves compliance and risk management protocols but also informs the development of compliance narratives with pinpoint accuracy. It underscores AI’s emerging role in strategic risk mitigation and proactive policy formation, positioning GRC officers to anticipate and navigate the complexities of regulatory evolution confidently.Keywords: cybersecurity, gen AI, retrieval augmented generation, cybersecurity defense strategies
Procedia PDF Downloads 955455 Defect Identification in Partial Discharge Patterns of Gas Insulated Switchgear and Straight Cable Joint
Authors: Chien-Kuo Chang, Yu-Hsiang Lin, Yi-Yun Tang, Min-Chiu Wu
Abstract:
With the trend of technological advancement, the harm caused by power outages is substantial, mostly due to problems in the power grid. This highlights the necessity for further improvement in the reliability of the power system. In the power system, gas-insulated switches (GIS) and power cables play a crucial role. Long-term operation under high voltage can cause insulation materials in the equipment to crack, potentially leading to partial discharges. If these partial discharges (PD) can be analyzed, preventative maintenance and replacement of equipment can be carried out, there by improving the reliability of the power grid. This research will diagnose defects by identifying three different defects in GIS and three different defects in straight cable joints, for a total of six types of defects. The partial discharge data measured will be converted through phase analysis diagrams and pulse sequence analysis. Discharge features will be extracted using convolutional image processing, and three different deep learning models, CNN, ResNet18, and MobileNet, will be used for training and evaluation. Class Activation Mapping will be utilized to interpret the black-box problem of deep learning models, with each model achieving an accuracy rate of over 95%. Lastly, the overall model performance will be enhanced through an ensemble learning voting method.Keywords: partial discharge, gas-insulated switches, straight cable joint, defect identification, deep learning, ensemble learning
Procedia PDF Downloads 785454 A Posteriori Trading-Inspired Model-Free Time Series Segmentation
Authors: Plessen Mogens Graf
Abstract:
Within the context of multivariate time series segmentation, this paper proposes a method inspired by a posteriori optimal trading. After a normalization step, time series are treated channelwise as surrogate stock prices that can be traded optimally a posteriori in a virtual portfolio holding either stock or cash. Linear transaction costs are interpreted as hyperparameters for noise filtering. Trading signals, as well as trading signals obtained on the reversed time series, are used for unsupervised channelwise labeling before a consensus over all channels is reached that determines the final segmentation time instants. The method is model-free such that no model prescriptions for segments are made. Benefits of proposed approach include simplicity, computational efficiency, and adaptability to a wide range of different shapes of time series. Performance is demonstrated on synthetic and real-world data, including a large-scale dataset comprising a multivariate time series of dimension 1000 and length 2709. Proposed method is compared to a popular model-based bottom-up approach fitting piecewise affine models and to a recent model-based top-down approach fitting Gaussian models and found to be consistently faster while producing more intuitive results in the sense of segmenting time series at peaks and valleys.Keywords: time series segmentation, model-free, trading-inspired, multivariate data
Procedia PDF Downloads 1365453 The Role of Sustainable Development in the Design and Planning of Smart Cities Using GIS Techniques: Models of Arab Cities
Authors: Ahmed M. Jihad
Abstract:
The paper presents the concept of sustainable development, and the role of geographic techniques in the design, planning and presentation of maps of smart cities with geographical vision, and the identification of programs and tools, and models of maps of Arab cities, is the problem of research in how to apply, process and experience these programs? What is the role of geographic techniques in planning and mapping the optimal place for these cities? The paper proposes an addition to the designs of Iraqi cities, as it can be developed in the future to serve as a model for interactive smart cities by developing its services. The importance of this paper stems from the concept of sustainable development dynamic which has become a method of development imposed by the present era in rapid development to achieve social balance and specialized programs in draw paper argues that ensuring sustainable development is achieved through the use of information technology. The paper will follow the theoretical presentation of the importance of the concept of development, design tools and programs. The paper follows the method of analysis of modern systems (System Analysis Approach) through the latest programs will provide results can be said that the new Iraqi cities can be developed with smart technologies, like some of the Arab and European cities that were newly created through the introduction of international investment, and therefore Plans can be made to select the best programs in manufacturing and producing maps and smart cities in the future.Keywords: geographic techniques, planning the cities, smart cities, sustainable development
Procedia PDF Downloads 2105452 Experiments to Study the Vapor Bubble Dynamics in Nucleate Pool Boiling
Authors: Parul Goel, Jyeshtharaj B. Joshi, Arun K. Nayak
Abstract:
Nucleate boiling is characterized by the nucleation, growth and departure of the tiny individual vapor bubbles that originate in the cavities or imperfections present in the heating surface. It finds a wide range of applications, e.g. in heat exchangers or steam generators, core cooling in power reactors or rockets, cooling of electronic circuits, owing to its highly efficient transfer of large amount of heat flux over small temperature differences. Hence, it is important to be able to predict the rate of heat transfer and the safety limit heat flux (critical heat flux, heat flux higher than this can lead to damage of the heating surface) applicable for any given system. A large number of experimental and analytical works exist in the literature, and are based on the idea that the knowledge of the bubble dynamics on the microscopic scale can lead to the understanding of the full picture of the boiling heat transfer. However, the existing data in the literature are scattered over various sets of conditions and often in disagreement with each other. The correlations obtained from such data are also limited to the range of conditions they were established for and no single correlation is applicable over a wide range of parameters. More recently, a number of researchers have been trying to remove empiricism in the heat transfer models to arrive at more phenomenological models using extensive numerical simulations; these models require state-of-the-art experimental data for a wide range of conditions, first for input and later, for their validation. With this idea in mind, experiments with sub-cooled and saturated demineralized water have been carried out under atmospheric pressure to study the bubble dynamics- growth rate, departure size and frequencies for nucleate pool boiling. A number of heating elements have been used to study the dependence of vapor bubble dynamics on the heater surface finish and heater geometry along with the experimental conditions like the degree of sub-cooling, super heat and the heat flux. An attempt has been made to compare the data obtained with the existing data and the correlations in the literature to generate an exhaustive database for the pool boiling conditions.Keywords: experiment, boiling, bubbles, bubble dynamics, pool boiling
Procedia PDF Downloads 3025451 A Discourse on the Rhythmic Pattern Employed in Yoruba Sakara Music of Nigeria
Authors: Oludare Olupemi Ezekiel
Abstract:
This research examines the rhythmic structure of Sakara music by tracing its roots and analyzing the various rhythmic patterns of this neo-traditional genre, as well as the contributions of the major exponents and contemporary practitioners, using these as a model for understanding and establishing African rhythms. Biography of the major exponents and contemporary practitioners, interviews and participant observational methods were used to elicit information. Samples of the genre which were chosen at random were transcribed, notated and analyzed for academic use and documentation. The research affirmed that rhythms such as the Hemiola, Cross-rhythm, Clave or Bell rhythm, Percussive, Speech and Melodic rhythm and other relevant rhythmic theories were prevalent and applicable to Sakara music, while making important contributions to musical scholarship through its analysis of the music. The analysis and discussions carried out in the research pointed towards a conclusion that the Yoruba musicians are guided by some preconceptions and sound musical considerations in making their rhythmic patterns, used as compositional techniques and not mere incidental occurrence. These rhythmic patterns, with its consequential socio-cultural connotations, enhance musical values and national identity in Nigeria. The study concludes by recommending that musicologists need to carry out more research into this and other neo-traditional genres in order to advance the globalisation of African music.Keywords: compositional techniques, globalisation, identity, neo-traditional, rhythmic theory, Sakara music
Procedia PDF Downloads 4425450 Modeling Slow Crack Growth under Thermal and Chemical Effects for Fitness Predictions of High-Density Polyethylene Material
Authors: Luis Marquez, Ge Zhu, Vikas Srivastava
Abstract:
High-density polyethylene (HDPE) is one of the most commonly used thermoplastic polymer materials for water and gas pipelines. Slow crack growth failure is a well-known phenomenon in high-density polyethylene material and causes brittle failure well below the yield point with no obvious sign. The failure of transportation pipelines can cause catastrophic environmental and economic consequences. Using the non-destructive testing method to predict slow crack growth failure behavior is the primary preventative measurement employed by the pipeline industry but is often costly and time-consuming. Phenomenological slow crack growth models are useful to predict the slow crack growth behavior in the polymer material due to their ability to evaluate slow crack growth under different temperature and loading conditions. We developed a quantitative method to assess the slow crack growth behavior in the high-density polyethylene pipeline material under different thermal conditions based on existing physics-based phenomenological models. We are also working on developing an experimental protocol and quantitative model that can address slow crack growth behavior under different chemical exposure conditions to improve the safety, reliability, and resilience of HDPE-based pipeline infrastructure.Keywords: mechanics of materials, physics-based modeling, civil engineering, fracture mechanics
Procedia PDF Downloads 2055449 Improvements in Double Q-Learning for Anomalous Radiation Source Searching
Authors: Bo-Bin Xiaoa, Chia-Yi Liua
Abstract:
In the task of searching for anomalous radiation sources, personnel holding radiation detectors to search for radiation sources may be exposed to unnecessary radiation risk, and automated search using machines becomes a required project. The research uses various sophisticated algorithms, which are double Q learning, dueling network, and NoisyNet, of deep reinforcement learning to search for radiation sources. The simulation environment, which is a 10*10 grid and one shielding wall setting in it, improves the development of the AI model by training 1 million episodes. In each episode of training, the radiation source position, the radiation source intensity, agent position, shielding wall position, and shielding wall length are all set randomly. The three algorithms are applied to run AI model training in four environments where the training shielding wall is a full-shielding wall, a lead wall, a concrete wall, and a lead wall or a concrete wall appearing randomly. The 12 best performance AI models are selected by observing the reward value during the training period and are evaluated by comparing these AI models with the gradient search algorithm. The results show that the performance of the AI model, no matter which one algorithm, is far better than the gradient search algorithm. In addition, the simulation environment becomes more complex, the AI model which applied Double DQN combined Dueling and NosiyNet algorithm performs better.Keywords: double Q learning, dueling network, NoisyNet, source searching
Procedia PDF Downloads 1125448 Studying the Dynamical Response of Nano-Microelectromechanical Devices for Nanomechanical Testing of Nanostructures
Authors: Mohammad Reza Zamani Kouhpanji
Abstract:
Characterizing the fatigue and fracture properties of nanostructures is one of the most challenging tasks in nanoscience and nanotechnology due to lack of a MEMS/NEMS device for generating uniform cyclic loadings at high frequencies. Here, the dynamic response of a recently proposed MEMS/NEMS device under different inputs signals is completely investigated. This MEMS/NEMS device is designed and modeled based on the electromagnetic force induced between paired parallel wires carrying electrical currents, known as Ampere’s Force Law (AFL). Since this MEMS/NEMS device only uses two paired wires for actuation part and sensing part, it represents highly sensitive and linear response for nanostructures with any stiffness and shapes (single or arrays of nanowires, nanotubes, nanosheets or nanowalls). In addition to studying the maximum gains at different resonance frequencies of the MEMS/NEMS device, its dynamical responses are investigated for different inputs and nanostructure properties to demonstrate the capability, usability, and reliability of the device for wide range of nanostructures. This MEMS/NEMS device can be readily integrated into SEM/TEM instruments to provide real time study of the fatigue and fracture properties of nanostructures as well as their softening or hardening behaviors, and initiation and/or propagation of nanocracks in them.Keywords: MEMS/NEMS devices, paired wire actuators and sensors, dynamical response, fatigue and fracture characterization, Ampere’s force law
Procedia PDF Downloads 3995447 Prediction of the Lateral Bearing Capacity of Short Piles in Clayey Soils Using Imperialist Competitive Algorithm-Based Artificial Neural Networks
Authors: Reza Dinarvand, Mahdi Sadeghian, Somaye Sadeghian
Abstract:
Prediction of the ultimate bearing capacity of piles (Qu) is one of the basic issues in geotechnical engineering. So far, several methods have been used to estimate Qu, including the recently developed artificial intelligence methods. In recent years, optimization algorithms have been used to minimize artificial network errors, such as colony algorithms, genetic algorithms, imperialist competitive algorithms, and so on. In the present research, artificial neural networks based on colonial competition algorithm (ANN-ICA) were used, and their results were compared with other methods. The results of laboratory tests of short piles in clayey soils with parameters such as pile diameter, pile buried length, eccentricity of load and undrained shear resistance of soil were used for modeling and evaluation. The results showed that ICA-based artificial neural networks predicted lateral bearing capacity of short piles with a correlation coefficient of 0.9865 for training data and 0.975 for test data. Furthermore, the results of the model indicated the superiority of ICA-based artificial neural networks compared to back-propagation artificial neural networks as well as the Broms and Hansen methods.Keywords: artificial neural network, clayey soil, imperialist competition algorithm, lateral bearing capacity, short pile
Procedia PDF Downloads 1525446 Energy Service Companies as a Facilitator for Implementation of Energy-Environment Conventions
Authors: Bahareh Arghand
Abstract:
The establishment of rules and regulations for more effective energy-environment interactions are essential to achieving sustainable development. Sustainable development requires mechanisms that can promote compliance in energy-environment conventions. There are many binding agreements and non-binding instruments at regional and international levels on energy and the environment. These conventions try to decrease conflicts of interest between energy, environment and economic by legal principles and practical mechanisms. The major core of conventions is their implementations because the poor implementation and enforcement power affect their success. In this regard, the main goal of this study is proposing the effective implementation mechanisms. Energy service companies' (ESCOs) activities can improve energy efficiency and decrease the environmental degradations. Therefore, it can be proposed and assessed the merit mechanism of ESCO performance as a facilitator to implement energy-environment conventions. An assessment of ESCO performance, including its potentials, problems, and limitations, as a facilitator for effective implementation of the energy-environment convention, is included. This study is oriented towards effective development and application of laws and the function of ESCOs as appropriate economic instruments and facilitator for implementation of energy-environment conventions. The resulting system of close cooperation between the energy-environment conventions and ESCOs is geared toward advancing environmental protection and economic factors by the transfer of environmentally-sound technologies that meet sustainable development objectives.Keywords: energy-environment conventions, energy service company, facilitator mechanism, sustainable development
Procedia PDF Downloads 1825445 Analysis of Overall Thermo-Elastic Properties of Random Particulate Nanocomposites with Various Interphase Models
Authors: Lidiia Nazarenko, Henryk Stolarski, Holm Altenbach
Abstract:
In the paper, a (hierarchical) approach to analysis of thermo-elastic properties of random composites with interphases is outlined and illustrated. It is based on the statistical homogenization method – the method of conditional moments – combined with recently introduced notion of the energy-equivalent inhomogeneity which, in this paper, is extended to include thermal effects. After exposition of the general principles, the approach is applied in the investigation of the effective thermo-elastic properties of a material with randomly distributed nanoparticles. The basic idea of equivalent inhomogeneity is to replace the inhomogeneity and the surrounding it interphase by a single equivalent inhomogeneity of constant stiffness tensor and coefficient of thermal expansion, combining thermal and elastic properties of both. The equivalent inhomogeneity is then perfectly bonded to the matrix which allows to analyze composites with interphases using techniques devised for problems without interphases. From the mechanical viewpoint, definition of the equivalent inhomogeneity is based on Hill’s energy equivalence principle, applied to the problem consisting only of the original inhomogeneity and its interphase. It is more general than the definitions proposed in the past in that, conceptually and practically, it allows to consider inhomogeneities of various shapes and various models of interphases. This is illustrated considering spherical particles with two models of interphases, Gurtin-Murdoch material surface model and spring layer model. The resulting equivalent inhomogeneities are subsequently used to determine effective thermo-elastic properties of randomly distributed particulate composites. The effective stiffness tensor and coefficient of thermal extension of the material with so defined equivalent inhomogeneities are determined by the method of conditional moments. Closed-form expressions for the effective thermo-elastic parameters of a composite consisting of a matrix and randomly distributed spherical inhomogeneities are derived for the bulk and the shear moduli as well as for the coefficient of thermal expansion. Dependence of the effective parameters on the interphase properties is included in the resulting expressions, exhibiting analytically the nature of the size-effects in nanomaterials. As a numerical example, the epoxy matrix with randomly distributed spherical glass particles is investigated. The dependence of the effective bulk and shear moduli, as well as of the effective thermal expansion coefficient on the particle volume fraction (for different radii of nanoparticles) and on the radius of nanoparticle (for fixed volume fraction of nanoparticles) for different interphase models are compared to and discussed in the context of other theoretical predictions. Possible applications of the proposed approach to short-fiber composites with various types of interphases are discussed.Keywords: effective properties, energy equivalence, Gurtin-Murdoch surface model, interphase, random composites, spherical equivalent inhomogeneity, spring layer model
Procedia PDF Downloads 1855444 Quasi-Photon Monte Carlo on Radiative Heat Transfer: An Importance Sampling and Learning Approach
Authors: Utkarsh A. Mishra, Ankit Bansal
Abstract:
At high temperature, radiative heat transfer is the dominant mode of heat transfer. It is governed by various phenomena such as photon emission, absorption, and scattering. The solution of the governing integrodifferential equation of radiative transfer is a complex process, more when the effect of participating medium and wavelength properties are taken into consideration. Although a generic formulation of such radiative transport problem can be modeled for a wide variety of problems with non-gray, non-diffusive surfaces, there is always a trade-off between simplicity and accuracy of the problem. Recently, solutions of complicated mathematical problems with statistical methods based on randomization of naturally occurring phenomena have gained significant importance. Photon bundles with discrete energy can be replicated with random numbers describing the emission, absorption, and scattering processes. Photon Monte Carlo (PMC) is a simple, yet powerful technique, to solve radiative transfer problems in complicated geometries with arbitrary participating medium. The method, on the one hand, increases the accuracy of estimation, and on the other hand, increases the computational cost. The participating media -generally a gas, such as CO₂, CO, and H₂O- present complex emission and absorption spectra. To model the emission/absorption accurately with random numbers requires a weighted sampling as different sections of the spectrum carries different importance. Importance sampling (IS) was implemented to sample random photon of arbitrary wavelength, and the sampled data provided unbiased training of MC estimators for better results. A better replacement to uniform random numbers is using deterministic, quasi-random sequences. Halton, Sobol, and Faure Low-Discrepancy Sequences are used in this study. They possess better space-filling performance than the uniform random number generator and gives rise to a low variance, stable Quasi-Monte Carlo (QMC) estimators with faster convergence. An optimal supervised learning scheme was further considered to reduce the computation costs of the PMC simulation. A one-dimensional plane-parallel slab problem with participating media was formulated. The history of some randomly sampled photon bundles is recorded to train an Artificial Neural Network (ANN), back-propagation model. The flux was calculated using the standard quasi PMC and was considered to be the training target. Results obtained with the proposed model for the one-dimensional problem are compared with the exact analytical and PMC model with the Line by Line (LBL) spectral model. The approximate variance obtained was around 3.14%. Results were analyzed with respect to time and the total flux in both cases. A significant reduction in variance as well a faster rate of convergence was observed in the case of the QMC method over the standard PMC method. However, the results obtained with the ANN method resulted in greater variance (around 25-28%) as compared to the other cases. There is a great scope of machine learning models to help in further reduction of computation cost once trained successfully. Multiple ways of selecting the input data as well as various architectures will be tried such that the concerned environment can be fully addressed to the ANN model. Better results can be achieved in this unexplored domain.Keywords: radiative heat transfer, Monte Carlo Method, pseudo-random numbers, low discrepancy sequences, artificial neural networks
Procedia PDF Downloads 2235443 Global Health, Humanitarian Medical Aid, and the Ethics of Rationing
Authors: N. W. Paul, S. Michl
Abstract:
In our globalized world we need to appreciate the fact that questions of health and justice need to be addressed on a global scale, too. The way in which diverse governmental and non-governmental initiatives are trying to answer the need for humanitarian medical aid has long since been a visible result of globalized responsibility. While the intention of humanitarian medical aids seems to be evident, the allocation of resources has become more and more an ethical and societal challenge. With a rising number and growing dimension of humanitarian catastrophes around the globe the search for ethically justifiable ways to decide who might benefit from limited resources has become a pressing question. Rooted in theories of justice (Rawls) and concepts of social welfare (Sen) we developed and implemented a model for an ethically sound distribution of a limited annual budget for humanitarian care in one of the largest medical universities of Germany. Based on our long lasting experience with civil casualties of war (Afghanistan) and civil war (Libya) as well as with under- and uninsured and/or stateless patients we are now facing the on-going refugee crisis as our most recent challenge in terms of global health and justice. Against this background, the paper strives to a) explain key issues of humanitarian medical aid in the 21st century, b) explore the problem of rationing from an ethical point of view, c) suggest a tool for the rational allocation of scarce resources in humanitarian medical aid, d) present actual cases of humanitarian care that have been managed with our toolbox, and e) discuss the international applicability of our model beyond local contexts.Keywords: humanitarian care, medical ethics, allocation, rationing
Procedia PDF Downloads 3975442 Automation of AAA Game Development using AI and Procedural Generation
Authors: Paul Toprac, Branden Heng, Harsheni Siddharthan, Allison Tseng, Sarah Abraham, Etienne Vouga
Abstract:
The goal of this project was to evaluate and document the capabilities and limitations of AI tools for empowering small teams to create high budget, high profile (AAA) 3D games typically developed by large studios. Two teams of novice game developers attempted to create two different games using AI and Unreal Engine 5.3. First, the teams evaluated 60 AI art, design, sound, and programming tools by considering their capability, ease of use, cost, and license restrictions. Then, the teams used a shortlist of 13 AI tools for game development. During this process, the following tools were found to be the most productive: (1) ChatGPT 4.0 for both game and narrative concepting and documentation; (2) Dall-E 3 and OpenArt for concept art; (3) Beatoven for music drafting; (4) Epic PCG for level design; and (5) ChatGPT 4.0 and Github Copilot for generating simple code and to complement human-made tutorials as an additional learning resource. While current generative AI may appear impressive at first glance, the assets they produce fall short of AAA industry standards. Generative AI tools are helpful when brainstorming ideas such as concept art and basic storylines, but they still cannot replace human input or creativity at this time. Regarding programming, AI can only effectively generate simple code and act as an additional learning resource. Thus, generative AI tools are at best tools to enhance developer productivity rather than as a system to replace developers.Keywords: AAA games, AI, automation tools, game development
Procedia PDF Downloads 235441 Effects of Aircraft Wing Configuration on Aerodynamic Efficiency
Authors: Aderet Pantierer, Shmuel Pantierer, Atif Saeed, Amir Elzawawy
Abstract:
In recent years, air travel has seen volatile growth. Due to this growth, the maximization of efficiency and space utilization has been a major issue for aircraft manufacturers. Elongation of the wingspan of aircraft has resulted in increased lift; and, thereby, efficiency. However, increasing the wingspan of aircraft has been detrimental to the manufacturing process and has led to airport congestion and required airport reconfiguration to accommodate the extended wingspans of aircraft. This project outlines differing wing configurations of a commercial aircraft and the effects on the aerodynamic loads produced. Multiple wing configurations are analyzed using Finite Element Models. These models are then validated by testing one wing configuration in a wind tunnel under laminar flow and turbulent flow conditions. The wing configurations to be tested include high and low wing aircraft, as well as various combinations of the two, including a unique model hereon referred to as an infinity wing. The infinity wing configuration consists of both a high and low wing, with the two wings connected by a vertical airfoil. This project seeks to determine if a wing configuration consisting of multiple airfoils produces more lift than the standard wing configurations and is able to provide a solution to manufacturing limitations as well as airport congestion. If the analysis confirms the hypothesis, a trade study will be performed to determine if and when an arrangement of multiple wings would be cost-effective.Keywords: aerodynamics, aircraft design, aircraft efficiency, wing configuration, wing design
Procedia PDF Downloads 2635440 Polarization Effects in Cosmic-Ray Acceleration by Cyclotron Auto-Resonance
Authors: Yousef I. Salamin
Abstract:
Theoretical investigations, analytical as well as numerical, have shown that electrons can be accelerated to GeV energies by the process of cyclotron auto-resonance acceleration (CARA). In CARA, the particle would be injected along the lines of a uniform magnetic field aligned parallel to the direction of propagation of a plane-wave radiation field. Unfortunately, an accelerator based on CARA would be prohibitively too long and too expensive to build and maintain. However, the process stands a better chance of success near the polar cap of a compact object (such as a neutron star, a black hole or a magnetar) or in an environment created in the wake of a binary neutron-star or blackhole merger. Dynamics of the nuclides ₁H¹, ₂He⁴, ₂₆Fe⁵⁶, and ₂₈Ni⁶², in such astrophysical conditions, have been investigated by single-particle calculations and many-particle simulations. The investigations show that these nuclides can reach ZeV energies (1 ZeV = 10²¹ eV) due to interaction with super-intense radiation of wavelengths = 1 and 10 m and = 50 pm and magnetic fields of strengths at the mega- and giga-tesla levels. Examples employing radiation intensities in the range 10³²-10⁴² W/m² have been used. Employing a two-parameter model for representing the radiation field, CARA is analytically generalized to include any state of polarization, and the basic working equations are derived rigorously and in closed analytic form.Keywords: compact objects, cosmic-ray acceleration, cyclotron auto-resonance, polarization effects, zevatron
Procedia PDF Downloads 1235439 Hydrological Characterization of a Watershed for Streamflow Prediction
Authors: Oseni Taiwo Amoo, Bloodless Dzwairo
Abstract:
In this paper, we extend the versatility and usefulness of GIS as a methodology for any river basin hydrologic characteristics analysis (HCA). The Gurara River basin located in North-Central Nigeria is presented in this study. It is an on-going research using spatial Digital Elevation Model (DEM) and Arc-Hydro tools to take inventory of the basin characteristics in order to predict water abstraction quantification on streamflow regime. One of the main concerns of hydrological modelling is the quantification of runoff from rainstorm events. In practice, the soil conservation service curve (SCS) method and the Conventional procedure called rational technique are still generally used these traditional hydrological lumped models convert statistical properties of rainfall in river basin to observed runoff and hydrograph. However, the models give little or no information about spatially dispersed information on rainfall and basin physical characteristics. Therefore, this paper synthesizes morphometric parameters in generating runoff. The expected results of the basin characteristics such as size, area, shape, slope of the watershed and stream distribution network analysis could be useful in estimating streamflow discharge. Water resources managers and irrigation farmers could utilize the tool for determining net return from available scarce water resources, where past data records are sparse for the aspect of land and climate.Keywords: hydrological characteristic, stream flow, runoff discharge, land and climate
Procedia PDF Downloads 3415438 The Impact of Artificial Intelligence in the Development of Textile and Fashion Industry
Authors: Basem Kamal Abasakhiroun Farag
Abstract:
Fashion, like many other areas of design, has undergone numerous developments over the centuries. The aim of the article is to recognize and evaluate the importance of advanced technologies in fashion design and to examine how they are transforming the role of contemporary fashion designers by transforming the creative process. It also discusses how contemporary culture is involved in such developments and how it influences fashion design in terms of conceptualization and production. The methodology used is based on examining various examples of the use of technology in fashion design and drawing parallels between what was feasible then and what is feasible today. Comparison of case studies, examples of existing fashion designs and experiences with craft methods; We therefore observe patterns that help us predict the direction of future developments in this area. Discussing the technological elements in fashion design helps us understand the driving force behind the trend. The research presented in the article shows that there is a trend towards significantly increasing interest and progress in the field of fashion technology, leading to the emergence of hybrid artisanal methods. In summary, as fashion technologies advance, their role in clothing production is becoming increasingly important, extending far beyond the humble sewing machine.Keywords: fashion, identity, such, textiles ambient intelligence, proximity sensors, shape memory materials, sound sensing garments, wearable technology bio textiles, fashion trends, nano textiles, new materials, smart textiles, techno textiles fashion design, functional aesthetics, 3D printing.
Procedia PDF Downloads 675437 Evaluation of Modern Natural Language Processing Techniques via Measuring a Company's Public Perception
Authors: Burak Oksuzoglu, Savas Yildirim, Ferhat Kutlu
Abstract:
Opinion mining (OM) is one of the natural language processing (NLP) problems to determine the polarity of opinions, mostly represented on a positive-neutral-negative axis. The data for OM is usually collected from various social media platforms. In an era where social media has considerable control over companies’ futures, it’s worth understanding social media and taking actions accordingly. OM comes to the fore here as the scale of the discussion about companies increases, and it becomes unfeasible to gauge opinion on individual levels. Thus, the companies opt to automize this process by applying machine learning (ML) approaches to their data. For the last two decades, OM or sentiment analysis (SA) has been mainly performed by applying ML classification algorithms such as support vector machines (SVM) and Naïve Bayes to a bag of n-gram representations of textual data. With the advent of deep learning and its apparent success in NLP, traditional methods have become obsolete. Transfer learning paradigm that has been commonly used in computer vision (CV) problems started to shape NLP approaches and language models (LM) lately. This gave a sudden rise to the usage of the pretrained language model (PTM), which contains language representations that are obtained by training it on the large datasets using self-supervised learning objectives. The PTMs are further fine-tuned by a specialized downstream task dataset to produce efficient models for various NLP tasks such as OM, NER (Named-Entity Recognition), Question Answering (QA), and so forth. In this study, the traditional and modern NLP approaches have been evaluated for OM by using a sizable corpus belonging to a large private company containing about 76,000 comments in Turkish: SVM with a bag of n-grams, and two chosen pre-trained models, multilingual universal sentence encoder (MUSE) and bidirectional encoder representations from transformers (BERT). The MUSE model is a multilingual model that supports 16 languages, including Turkish, and it is based on convolutional neural networks. The BERT is a monolingual model in our case and transformers-based neural networks. It uses a masked language model and next sentence prediction tasks that allow the bidirectional training of the transformers. During the training phase of the architecture, pre-processing operations such as morphological parsing, stemming, and spelling correction was not used since the experiments showed that their contribution to the model performance was found insignificant even though Turkish is a highly agglutinative and inflective language. The results show that usage of deep learning methods with pre-trained models and fine-tuning achieve about 11% improvement over SVM for OM. The BERT model achieved around 94% prediction accuracy while the MUSE model achieved around 88% and SVM did around 83%. The MUSE multilingual model shows better results than SVM, but it still performs worse than the monolingual BERT model.Keywords: BERT, MUSE, opinion mining, pretrained language model, SVM, Turkish
Procedia PDF Downloads 146